sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
listlengths
1
1.84k
created_at
stringlengths
25
25
arxiv
listlengths
0
201
languages
listlengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
listlengths
0
722
processed_texts
listlengths
1
723
tokens_length
listlengths
1
723
input_texts
listlengths
1
61
embeddings
listlengths
768
768
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={INFORMS Journal on Data Science}, year={2022} } ```
{}
text-classification
avichr/hebEMO_disgust
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={INFORMS Journal on Data Science}, year={2022} } ```
{}
text-classification
avichr/hebEMO_fear
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2021). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. arXiv preprint arXiv:2102.01909. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={arXiv preprint arXiv:2102.01909}, year={2021} } ```
{}
text-classification
avichr/hebEMO_joy
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2021). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. arXiv preprint arXiv:2102.01909.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2021). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. arXiv preprint arXiv:2102.01909." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2021). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. arXiv preprint arXiv:2102.01909." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={INFORMS Journal on Data Science}, year={2022} } ```
{}
text-classification
avichr/hebEMO_sadness
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={INFORMS Journal on Data Science}, year={2022} } ```
{}
text-classification
avichr/hebEMO_surprise
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# HebEMO - Emotion Recognition Model for Modern Hebrew <img align="right" src="https://github.com/avichaychriqui/HeBERT/blob/main/data/heBERT_logo.png?raw=true" width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. ## Emotion UGC Data Description Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and [eight emotions](https://en.wikipedia.org/wiki/Robert_Plutchik#Plutchik's_wheel_of_emotions): anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. | | anger | disgust | expectation | fear | happy | sadness | surprise | trust | sentiment | |------:|------:|--------:|------------:|-----:|------:|--------:|---------:|------:|-----------| | **ratio** | 0.78 | 0.83 | 0.58 | 0.45 | 0.12 | 0.59 | 0.17 | 0.11 | 0.25 | ## Performance ### Emotion Recognition | emotion | f1-score | precision | recall | |-------------|----------|-----------|----------| | anger | 0.96 | 0.99 | 0.93 | | disgust | 0.97 | 0.98 | 0.96 | |anticipation | 0.82 | 0.80 | 0.87 | | fear | 0.79 | 0.88 | 0.72 | | joy | 0.90 | 0.97 | 0.84 | | sadness | 0.90 | 0.86 | 0.94 | | surprise | 0.40 | 0.44 | 0.37 | | trust | 0.83 | 0.86 | 0.80 | *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis | | precision | recall | f1-score | |--------------|-----------|--------|----------| | neutral | 0.83 | 0.56 | 0.67 | | positive | 0.96 | 0.92 | 0.94 | | negative | 0.97 | 0.99 | 0.98 | | accuracy | | | 0.97 | | macro avg | 0.92 | 0.82 | 0.86 | | weighted avg | 0.96 | 0.97 | 0.96 | *Sentiment (polarity) analysis model is also available on AWS! for more information visit [AWS' git](https://github.com/aws-samples/aws-lambda-docker-serverless-inference/tree/main/hebert-sentiment-analysis-inference-docker-lambda)* ## How to use ### Emotion Recognition Model An online model can be found at [huggingface spaces](https://huggingface.co/spaces/avichr/HebEMO_demo) or as [colab notebook](https://colab.research.google.com/drive/1Jw3gOWjwVMcZslu-ttXoNeD17lms1-ff?usp=sharing) ``` # !pip install pyplutchik==0.0.7 # !pip install transformers==4.14.1 !git clone https://github.com/avichaychriqui/HeBERT.git from HeBERT.src.HebEMO import * HebEMO_model = HebEMO() HebEMO_model.hebemo(input_path = 'data/text_example.txt') # return analyzed pandas.DataFrame hebEMO_df = HebEMO_model.hebemo(text='ื”ื—ื™ื™ื ื™ืคื™ื ื•ืžืื•ืฉืจื™ื', plot=True) ``` <img src="https://github.com/avichaychriqui/HeBERT/blob/main/data/hebEMO1.png?raw=true" width="300" height="300" /> ### For sentiment classification model (polarity ONLY): from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ## Contact us [Avichay Chriqui](mailto:[email protected]) <br> [Inbal yahav](mailto:[email protected]) <br> The Coller Semitic Languages AI Lab <br> Thank you, ืชื•ื“ื”, ุดูƒุฑุง <br> ## If you used this model please cite us as : Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming. ``` @article{chriqui2021hebert, title={HeBERT \& HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition}, author={Chriqui, Avihay and Yahav, Inbal}, journal={INFORMS Journal on Data Science}, year={2022} } ```
{}
text-classification
avichr/hebEMO_trust
[ "transformers", "pytorch", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
HebEMO - Emotion Recognition Model for Modern Hebrew ==================================================== <img align="right" src="URL width="250"> HebEMO is a tool that detects polarity and extracts emotions from modern Hebrew User-Generated Content (UGC), which was trained on a unique Covid-19 related dataset that we collected and annotated. HebEMO yielded a high performance of weighted average F1-score = 0.96 for polarity classification. Emotion detection reached an F1-score of 0.78-0.97, with the exception of *surprise*, which the model failed to capture (F1 = 0.41). These results are better than the best-reported performance, even when compared to the English language. Emotion UGC Data Description ---------------------------- Our UGC data includes comments posted on news articles collected from 3 major Israeli news sites, between January 2020 to August 2020. The total size of the data is ~150 MB, including over 7 million words and 350K sentences. ~2000 sentences were annotated by crowd members (3-10 annotators per sentence) for overall sentiment (polarity) and eight emotions: anger, disgust, anticipation , fear, joy, sadness, surprise and trust. The percentage of sentences in which each emotion appeared is found in the table below. Performance ----------- ### Emotion Recognition *The above metrics is for positive class (meaning, the emotion is reflected in the text).* ### Sentiment (Polarity) Analysis *Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git* How to use ---------- ### Emotion Recognition Model An online model can be found at huggingface spaces or as colab notebook <img src="URL width="300" height="300" /> ### For sentiment classification model (polarity ONLY): ``` from transformers import AutoTokenizer, AutoModel, pipeline tokenizer = AutoTokenizer.from_pretrained("avichr/heBERT_sentiment_analysis") #same as 'avichr/heBERT' tokenizer model = AutoModel.from_pretrained("avichr/heBERT_sentiment_analysis") # how to use? sentiment_analysis = pipeline( "sentiment-analysis", model="avichr/heBERT_sentiment_analysis", tokenizer="avichr/heBERT_sentiment_analysis", return_all_scores = True ) sentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื') >>> [[{'label': 'neutral', 'score': 0.9978172183036804}, >>> {'label': 'positive', 'score': 0.0014792329166084528}, >>> {'label': 'negative', 'score': 0.0007035882445052266}]] sentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื') >>> [[{'label': 'neutral', 'score': 0.00047328314394690096}, >>> {'label': 'possitive', 'score': 0.9994067549705505}, >>> {'label': 'negetive', 'score': 0.00011996887042187154}]] sentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื') >>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, >>> {'label': 'possitive', 'score': 8.876807987689972e-05}, >>> {'label': 'negetive', 'score': 0.9998190999031067}]] ``` Contact us ---------- Avichay Chriqui Inbal yahav The Coller Semitic Languages AI Lab Thank you, ืชื•ื“ื”, ุดูƒุฑุง If you used this model please cite us as : ------------------------------------------ Chriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming.
[ "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n", "### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*", "### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------", "### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />", "### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")", "# how to use?\nsentiment_analysis = pipeline(\n \"sentiment-analysis\",\n model=\"avichr/heBERT_sentiment_analysis\",\n tokenizer=\"avichr/heBERT_sentiment_analysis\",\n return_all_scores = True\n)\n\nsentiment_analysis('ืื ื™ ืžืชืœื‘ื˜ ืžื” ืœืื›ื•ืœ ืœืืจื•ื—ืช ืฆื”ืจื™ื™ื')\t\n>>> [[{'label': 'neutral', 'score': 0.9978172183036804},\n>>> {'label': 'positive', 'score': 0.0014792329166084528},\n>>> {'label': 'negative', 'score': 0.0007035882445052266}]]\n\nsentiment_analysis('ืงืคื” ื–ื” ื˜ืขื™ื')\n>>> [[{'label': 'neutral', 'score': 0.00047328314394690096},\n>>> {'label': 'possitive', 'score': 0.9994067549705505},\n>>> {'label': 'negetive', 'score': 0.00011996887042187154}]]\n\nsentiment_analysis('ืื ื™ ืœื ืื•ื”ื‘ ืืช ื”ืขื•ืœื')\n>>> [[{'label': 'neutral', 'score': 9.214012970915064e-05}, \n>>> {'label': 'possitive', 'score': 8.876807987689972e-05}, \n>>> {'label': 'negetive', 'score': 0.9998190999031067}]]\n\n```\n\nContact us\n----------\n\n\nAvichay Chriqui \n\nInbal yahav \n\nThe Coller Semitic Languages AI Lab \n\nThank you, ืชื•ื“ื”, ุดูƒุฑุง \n\n\n\nIf you used this model please cite us as :\n------------------------------------------\n\n\nChriqui, A., & Yahav, I. (2022). HeBERT & HebEMO: a Hebrew BERT Model and a Tool for Polarity Analysis and Emotion Recognition. INFORMS Journal on Data Science, forthcoming." ]
[ 36, 30, 43, 42, 101, 454 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Emotion Recognition\n\n\n\n*The above metrics is for positive class (meaning, the emotion is reflected in the text).*### Sentiment (Polarity) Analysis\n\n\n\n*Sentiment (polarity) analysis model is also available on AWS! for more information visit AWS' git*\n\n\nHow to use\n----------### Emotion Recognition Model\n\n\nAn online model can be found at huggingface spaces or as colab notebook\n\n\n<img src=\"URL width=\"300\" height=\"300\" />### For sentiment classification model (polarity ONLY):\n\n\n\n```\nfrom transformers import AutoTokenizer, AutoModel, pipeline\n\ntokenizer = AutoTokenizer.from_pretrained(\"avichr/heBERT_sentiment_analysis\") #same as 'avichr/heBERT' tokenizer\nmodel = AutoModel.from_pretrained(\"avichr/heBERT_sentiment_analysis\")" ]
[ -0.05714952573180199, 0.11110629886388779, -0.00238785264082253, 0.030166203156113625, 0.10004875808954239, 0.003578263334929943, 0.13375815749168396, 0.07131693512201309, 0.11786022782325745, 0.06436245143413544, 0.03026989847421646, 0.01848757639527321, 0.06968981772661209, 0.005771971307694912, -0.04110298305749893, -0.3015744984149933, -0.037111446261405945, 0.038363151252269745, 0.19287624955177307, 0.10721689462661743, 0.0613904632627964, -0.06863821297883987, 0.12464040517807007, 0.08453105390071869, -0.10529478639364243, 0.016964741051197052, -0.025218775495886803, -0.0626642182469368, 0.071998730301857, 0.014943886548280716, 0.01185645256191492, 0.011508060619235039, 0.0592644177377224, -0.2508513927459717, 0.015520672313869, -0.012714093551039696, 0.027932919561862946, 0.09504154324531555, 0.11821948736906052, -0.1255655139684677, 0.17186178267002106, -0.05372054502367973, 0.06846633553504944, 0.03917394205927849, -0.04998712241649628, -0.14551544189453125, -0.048287324607372284, 0.04706742241978645, 0.03753489628434181, 0.05281132459640503, -0.045090626925230026, 0.21014092862606049, -0.07574846595525742, 0.060586199164390564, 0.20933230221271515, -0.12187093496322632, -0.018357345834374428, -0.02350684255361557, -0.05073586851358414, -0.002743202494457364, -0.1504358947277069, 0.033678386360406876, 0.02144741266965866, 0.02178831584751606, 0.06206134334206581, -0.025116538628935814, 0.04233046621084213, -0.0010931285796687007, -0.0852745994925499, 0.005963102448731661, 0.1210358738899231, 0.08719753473997116, -0.031067829579114914, -0.09670931845903397, -0.020166272297501564, -0.0705728679895401, -0.029231471940875053, -0.003114991122856736, 0.029646791517734528, -0.021665174514055252, -0.045085787773132324, 0.03772739693522453, -0.12392906844615936, 0.04919584095478058, -0.05435098707675934, 0.22750426828861237, -0.07043153047561646, 0.03246469795703888, 0.056103311479091644, 0.020898159593343735, -0.0989692211151123, -0.11091220378875732, -0.02067069336771965, -0.06519303470849991, 0.009859486483037472, -0.040591564029455185, -0.07718770951032639, -0.1263759881258011, -0.04162575304508209, 0.023624904453754425, 0.015619133599102497, 0.03410203382372856, 0.023408962413668633, -0.03583158552646637, 0.15798155963420868, 0.31907522678375244, -0.041169192641973495, 0.008635057136416435, -0.008613990619778633, 0.016303490847349167, 0.0389983132481575, -0.019488299265503883, -0.06948169320821762, 0.032966889441013336, 0.06531037390232086, -0.05796249210834503, -0.16131705045700073, 0.11433148384094238, -0.14055339992046356, -0.03655851632356644, -0.029600918292999268, -0.06053897365927696, 0.0653640404343605, 0.0642823800444603, -0.022201981395483017, 0.2159423679113388, 0.03275194391608238, -0.011150185950100422, -0.004169536754488945, 0.09930907934904099, -0.02696658857166767, 0.039299581199884415, -0.030739840120077133, -0.0962187647819519, 0.03825491666793823, 0.007158012595027685, 0.06093386560678482, -0.13912837207317352, -0.08478449285030365, -0.015799932181835175, 0.03411271795630455, -0.0707305371761322, 0.07326676696538925, -0.054602112621068954, 0.058576617389917374, 0.02969813160598278, 0.03265715017914772, -0.10108142346143723, -0.017265543341636658, -0.03607228398323059, 0.04076690971851349, 0.1539379060268402, 0.04533466324210167, 0.0230806116014719, -0.20417210459709167, -0.03615845739841461, -0.1490575224161148, 0.10328312963247299, -0.12945491075515747, 0.15024657547473907, -0.05657362937927246, -0.04148262366652489, 0.09579368680715561, -0.009971863590180874, 0.012130624614655972, 0.18902716040611267, -0.16440202295780182, -0.1270899623632431, 0.06938919425010681, -0.08679357171058655, -0.019908828660845757, 0.15028038620948792, -0.034142956137657166, 0.07786925882101059, 0.09523604810237885, 0.21709021925926208, -0.017264416441321373, -0.11219816654920578, -0.037877555936574936, 0.09571424126625061, 0.00066460354719311, 0.13471102714538574, -0.00034640979720279574, 0.06600665301084518, -0.07069630175828934, 0.06097520887851715, 0.02401112951338291, 0.05582548677921295, -0.04916468635201454, -0.04400520399212837, -0.004304068628698587, -0.011916105635464191, 0.033191993832588196, 0.028248287737369537, 0.03283471614122391, -0.06851252913475037, -0.07446593791246414, -0.09948024898767471, 0.03688353672623634, -0.060179587453603745, 0.03121056593954563, -0.09216148406267166, 0.07842104136943817, -0.009014283306896687, 0.012966523878276348, -0.15300238132476807, 0.11455217003822327, 0.02898498810827732, -0.031007423996925354, 0.04860084876418114, 0.003194626187905669, 0.058442048728466034, 0.0035918070934712887, 0.03371811658143997, -0.022975008934736252, 0.07020119577646255, 0.04990507289767265, -0.04215923696756363, -0.19277051091194153, 0.005603697616606951, -0.07316403836011887, 0.14732982218265533, -0.04899526759982109, -0.002847484312951565, 0.12890982627868652, 0.061536405235528946, 0.03104916773736477, 0.0002760096685960889, 0.02417522482573986, -0.012345975264906883, -0.04362119361758232, -0.03069707378745079, 0.025879116728901863, 0.018307628110051155, -0.13139156997203827, 0.08631858974695206, -0.05442635715007782, -0.02765212580561638, 0.14489158987998962, -0.06336673349142075, -0.1034504622220993, -0.08824540674686432, -0.008060231804847717, 0.06989827752113342, 0.007297232747077942, 0.03362454101443291, 0.19085641205310822, 0.07332015782594681, 0.06556550413370132, -0.00957244262099266, 0.030255312100052834, 0.025312067940831184, -0.13725502789020538, -0.041280362755060196, 0.09148289263248444, -0.10350339114665985, -0.13187247514724731, 0.07460962980985641, 0.15447017550468445, -0.03210891783237457, 0.1007431223988533, 0.02485879510641098, -0.03865909203886986, -0.05131882429122925, -0.12463521957397461, -0.04318670555949211, 0.03614585101604462, -0.11165683716535568, -0.06405649334192276, 0.03404178470373154, -0.10538577288389206, -0.08078263700008392, -0.09100288152694702, 0.01572766713798046, 0.10150313377380371, 0.062242161482572556, 0.00006127321830717847, 0.015162918716669083, 0.0011884834384545684, 0.06380827724933624, 0.02307678945362568, 0.0001299414288951084, -0.04117754474282265, -0.008884834125638008, -0.10097693651914597, 0.14696277678012848, -0.11139912903308868, -0.2494504451751709, -0.060597844421863556, -0.08913914114236832, -0.0721609815955162, 0.048158369958400726, 0.016501890495419502, -0.22097724676132202, -0.12226349860429764, -0.09853415191173553, 0.07074825465679169, 0.11512831598520279, -0.009843975305557251, 0.036244362592697144, -0.07569064944982529, -0.011033756658434868, -0.08385533839464188, -0.03435356169939041, -0.07314208894968033, 0.04282936453819275, 0.10121342539787292, -0.01960579678416252, 0.06711738556623459, 0.124812550842762, -0.020396782085299492, 0.03894612565636635, -0.008883945643901825, 0.1831986904144287, -0.04651079326868057, -0.04051648825407028, 0.021148860454559326, -0.03832155466079712, 0.07234557718038559, 0.09328436851501465, 0.07717116177082062, -0.06812035292387009, 0.013046244159340858, 0.017308952286839485, -0.030889468267560005, -0.10171978920698166, -0.08448684215545654, 0.01806516759097576, 0.060829732567071915, 0.041249919682741165, -0.02308112382888794, 0.017580939456820488, 0.08620613813400269, 0.027402451261878014, -0.04862137883901596, -0.08799982070922852, 0.1348295509815216, 0.15897208452224731, -0.04597923532128334, 0.006312469951808453, -0.002093250397592783, -0.07156861573457718, 0.1391761600971222, -0.02606562152504921, 0.03905526548624039, -0.006347245536744595, 0.13306619226932526, -0.06507845222949982, 0.10127142816781998, 0.05916126072406769, 0.11035788804292679, -0.056689176708459854, -0.04411059617996216, -0.05994047597050667, -0.05087381973862648, -0.11436105519533157, 0.11330374330282211, -0.0066363574005663395, 0.05523999035358429, -0.05022811517119408, -0.051965054124593735, 0.1077115535736084, 0.22793975472450256, 0.0591561533510685, -0.25323981046676636, -0.0686786025762558, 0.009947765618562698, -0.007084956858307123, -0.0025991182774305344, -0.01752382703125477, -0.03765197843313217, -0.07935363054275513, 0.12352833151817322, -0.044781215488910675, 0.04465048760175705, -0.13828526437282562, 0.08113045990467072, -0.0637948140501976, 0.0654272511601448, -0.028730688616633415, 0.052809033542871475, -0.11454065144062042, 0.17555560171604156, 0.04587596282362938, -0.057071104645729065, -0.08517764508724213, -0.016935739666223526, 0.10902922600507736, 0.11880172789096832, 0.07715781778097153, 0.01460183970630169, 0.09450516104698181, -0.18441644310951233, -0.05594347044825554, 0.020573992282152176, 0.05104583501815796, -0.05053224787116051, 0.047349609434604645, 0.000524677918292582, -0.033798713237047195, -0.009068838320672512, 0.036170314997434616, -0.08415994793176651, -0.1429595649242401, 0.012065151706337929, 0.0094917519018054, 0.010294497013092041, 0.03689603880047798, -0.12225450575351715, -0.05125316604971886, 0.16116246581077576, 0.1477653831243515, -0.028124691918492317, -0.1124531552195549, 0.09252732992172241, -0.024481387808918953, -0.05152057111263275, -0.08866789191961288, -0.039120689034461975, 0.16852989792823792, -0.019151227548718452, -0.12278773635625839, 0.11279238015413284, -0.05016443133354187, -0.05057869479060173, -0.04117446392774582, 0.035293202847242355, 0.061951201409101486, 0.02523101679980755, 0.038087598979473114, -0.02076493203639984, -0.07539035379886627, -0.10231488943099976, 0.08803560584783554, 0.13238798081874847, -0.11072815954685211, 0.03893778845667839, 0.039849523454904556, -0.10440170019865036, -0.09982553124427795, 0.03690626472234726, 0.10076136887073517, 0.14097817242145538, -0.03923410922288895, 0.039919182658195496, 0.1715613752603531, -0.024582477286458015, -0.18060359358787537, 0.04535749554634094, 0.037998098880052567, 0.01002370472997427, 0.1293017417192459, -0.09022380411624908, 0.08750459551811218, -0.0768595039844513, 0.027244605123996735, -0.11704404652118683, 0.06953798234462738, -0.06516306847333908, 0.19453643262386322, 0.08216002583503723, 0.07716286927461624, -0.044351931661367416, -0.021396581083536148, -0.06242620199918747, -0.042115986347198486, 0.17284613847732544, 0.003747866954654455, 0.032461341470479965, 0.008940202184021473, 0.27116674184799194, 0.08280275017023087, -0.0007950533763505518, 0.07650277763605118, 0.007533969823271036, 0.05447465181350708, -0.08042380958795547, 0.02754669263958931, -0.007192419841885567, -0.025037283077836037, 0.11652127653360367, -0.09047596901655197, -0.007830829359591007, -0.07360092550516129, -0.03963785618543625, -0.07356758415699005, 0.03318201005458832, -0.0026184776797890663, -0.048998430371284485, -0.04957396537065506, 0.027963820844888687, 0.0805455893278122, -0.0510936975479126, -0.01943330466747284, -0.035153359174728394, -0.05594256892800331, 0.23936344683170319, 0.11700337380170822, -0.02597968466579914, 0.0027199676260352135, 0.016878642141819, -0.02270723134279251, 0.09989113360643387, -0.15558630228042603, 0.01254607830196619, 0.05776435136795044, 0.009916920214891434, 0.12556415796279907, 0.008287526667118073, -0.14069601893424988, -0.009617261588573456, 0.035642024129629135, -0.1751815229654312, -0.030494876205921173, -0.07503741979598999, 0.049276817589998245, -0.007170799653977156, 0.006274781655520201, 0.08347580581903458, -0.07152559608221054, -0.049931809306144714, -0.013694355264306068, 0.025345539674162865, -0.00146165129262954, -0.0037304325960576534, -0.045028358697891235, -0.030609458684921265, -0.10929816961288452, -0.02132885903120041, -0.04100256785750389, -0.21921983361244202, 0.056166499853134155, 0.06079370528459549, -0.08246426284313202, -0.0582834929227829, 0.01936364732682705, 0.29404497146606445, -0.19307325780391693, 0.02806824818253517, -0.023906415328383446, -0.1544886976480484, -0.022258248180150986, 0.12467177212238312, 0.0954456776380539, -0.024593379348516464, -0.1404363214969635, 0.030229542404413223, -0.02304587885737419, 0.05352003872394562, 0.11052357405424118, 0.004536626860499382, -0.04704723507165909, -0.109016053378582, -0.02120431698858738, 0.013693992048501968, -0.04667602479457855, -0.06183773651719093, -0.07825471460819244, -0.06904077529907227, -0.16406649351119995, -0.012542525306344032, -0.058146119117736816, 0.07372001558542252, 0.03943605720996857, 0.012213186360895634, 0.06527350097894669, -0.05074858292937279, -0.08607477694749832, 0.03815251961350441, 0.022185711190104485, 0.04954727366566658, -0.070380300283432, -0.03450341150164604, 0.02945481613278389, -0.06861342489719391, 0.11098408699035645, 0.061547379940748215, -0.02315688692033291, 0.05119814723730087, -0.26901325583457947, -0.016146553680300713, 0.13711871206760406, -0.017868056893348694, 0.08060342073440552, -0.0436381958425045, -0.016930535435676575, -0.001965900883078575, 0.015276946127414703, -0.028847308829426765, 0.1873612403869629, -0.022948987782001495, 0.09007275849580765, 0.09213138371706009, -0.08143895864486694, -0.11221106350421906, 0.019573362544178963, 0.051669321954250336, 0.05131275951862335, 0.12979060411453247, -0.08743323385715485, 0.05484622344374657, -0.10270662605762482, 0.010318927466869354, 0.041053082793951035, -0.061184365302324295, -0.2136709839105606, -0.07355303317308426, 0.042940039187669754, 0.007196575403213501, 0.1701645702123642, 0.10899233818054199, 0.08643944561481476, -0.016447652131319046, 0.16549858450889587, 0.006999568082392216, -0.005511309020221233, -0.01422320306301117, 0.032809603959321976, -0.02959870919585228, -0.007119777612388134, 0.04312361404299736, 0.07987532764673233, -0.012161395512521267, 0.010204288177192211, -0.0335565023124218, 0.05100620910525322, 0.09639780223369598, 0.023876283317804337, 0.09755061566829681, -0.04609459638595581, -0.08771442621946335, -0.06932084262371063, 0.13955934345722198, -0.03112935833632946, 0.1943405419588089, 0.03199787065386772, -0.03606764227151871, 0.06716026365756989, -0.05042349919676781, -0.06837238371372223, -0.08097920566797256, -0.2751479148864746, -0.06441842764616013, -0.1637800633907318, 0.03565886244177818, -0.12541979551315308, 0.028453387320041656, -0.09813910722732544, 0.05151082202792168, -0.06654627621173859, 0.056864168494939804, -0.052406154572963715, -0.07657787948846817, 0.17868627607822418, 0.009532692842185497, -0.13943123817443848, -0.021790076047182083, 0.013793445192277431, -0.000810281082522124, 0.08483269065618515, 0.041173480451107025, 0.035394493490457535, -0.019902225583791733, 0.012312018312513828, -0.0765467956662178, -0.14773209393024445, -0.01742302067577839, -0.052873291075229645, -0.017484722658991814, 0.04266820102930069, -0.032132044434547424, 0.0044807312078773975, -0.0359954871237278, 0.10159332305192947, -0.09986559301614761, 0.14437665045261383, -0.12288777530193329, 0.20122291147708893, -0.1286548525094986, 0.07292997092008591, -0.015683883801102638, -0.08267943561077118, -0.0734853446483612, 0.18168962001800537, 0.08860146254301071, -0.07054264098405838, 0.0027478784322738647, -0.10764849185943604, 0.03241332992911339, -0.061920929700136185, 0.03431762754917145, 0.05082819238305092, -0.04148537293076515, -0.07966078072786331, 0.1485104113817215, -0.04795806109905243, -0.11011204123497009, 0.02085161954164505, -0.006817598827183247, 0.08358749747276306, 0.0012338570086285472, -0.11302261054515839, 0.0686025395989418, -0.09359396249055862, -0.11624788492918015, 0.1153474897146225, -0.11182571947574615, -0.040194373577833176, -0.015919415280222893, 0.0893755778670311, 0.14065919816493988, 0.11308817565441132, 0.009414928033947945, -0.017739620059728622, 0.18110695481300354, -0.005285756662487984, -0.14209416508674622, -0.04643199220299721, 0.13129884004592896, -0.25397053360939026, 0.03385443612933159, 0.0009518279111944139, 0.003355909138917923, 0.07944285124540329, 0.040329668670892715, -0.05282524600625038, 0.04989849403500557, -0.0058122784830629826, -0.1550990641117096, -0.02997083030641079, 0.19420774281024933, 0.0075323390774428844, 0.04222850874066353, -0.002144339494407177, -0.24399468302726746, 0.03491741046309471, -0.071846142411232, -0.01814817450940609, -0.04334733262658119, 0.04876234382390976, -0.06094012409448624, 0.0943102166056633, 0.15767870843410492, -0.016167979687452316, -0.039723772555589676, -0.057374801486730576, -0.015983272343873978, -0.0038007053080946207, -0.11902210861444473, 0.017769096419215202, -0.06841042637825012, -0.022728830575942993, 0.17090682685375214, 0.02467295154929161, -0.21318630874156952, -0.03698364645242691, -0.050419002771377563, 0.016850251704454422, -0.010066967457532883, 0.07579634338617325, 0.03157433494925499, 0.06591060012578964, 0.0022785367909818888, -0.18926464021205902, 0.08003178238868713, 0.10619169473648071, -0.034325167536735535, -0.03357170522212982 ]
null
null
transformers
# rickbot Dialo-GPT
{"tags": ["conversational"]}
text-generation
avinashshrangee/DialoGPT-small-Ricky
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# rickbot Dialo-GPT
[ "# rickbot Dialo-GPT" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# rickbot Dialo-GPT" ]
[ 51, 9 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# rickbot Dialo-GPT" ]
[ -0.00890159048140049, 0.0815073698759079, -0.0066546304151415825, 0.02727660909295082, 0.14936144649982452, 0.00967015977948904, 0.1361241638660431, 0.15373891592025757, 0.015581551939249039, -0.02409518137574196, 0.12058112025260925, 0.22871136665344238, 0.0018438296392560005, 0.09797441214323044, -0.05322940647602081, -0.2865822911262512, 0.06262137740850449, 0.04382408410310745, -0.017062421888113022, 0.12026268243789673, 0.09235977381467819, -0.032900165766477585, 0.08834689110517502, -0.014367916621267796, -0.1693214774131775, 0.003940476104617119, 0.0506846159696579, -0.13178037106990814, 0.11156795918941498, 0.06259684264659882, 0.04318529739975929, 0.012979418970644474, -0.06731945276260376, -0.11197543144226074, 0.0353698655962944, 0.01676083728671074, -0.04539467394351959, 0.0564439594745636, -0.004963364917784929, -0.08179918676614761, 0.1284571886062622, 0.09748780727386475, 0.0003606786485761404, 0.05917423963546753, -0.15369950234889984, -0.06398938596248627, -0.03670995682477951, 0.06694691628217697, 0.036705970764160156, 0.10874508321285248, -0.0349806472659111, 0.16018660366535187, -0.051994238048791885, 0.11133366823196411, 0.14716321229934692, -0.33272784948349, -0.019677240401506424, 0.1404191255569458, 0.0554250068962574, 0.06045207381248474, -0.03918703645467758, 0.08457404375076294, 0.0023936545476317406, 0.004099670797586441, -0.06524451076984406, -0.07710503041744232, -0.1102038249373436, 0.004559176042675972, -0.07797041535377502, -0.04310060665011406, 0.2458500862121582, -0.048988789319992065, 0.06429578363895416, -0.057959578931331635, -0.07440301775932312, -0.041025325655937195, -0.0322471559047699, -0.01756012812256813, -0.09504628926515579, 0.0718923807144165, -0.03339775651693344, -0.10481049120426178, -0.11564578115940094, -0.03809959068894386, -0.17699232697486877, 0.18134206533432007, 0.04335613176226616, 0.04676050692796707, -0.19572073221206665, 0.08764449506998062, 0.01425981242209673, -0.08614479005336761, 0.0009361235424876213, -0.08003265410661697, 0.01693739742040634, 0.01562820002436638, -0.006140718702226877, -0.01100525539368391, 0.11517935246229172, 0.13378466665744781, -0.014423706568777561, 0.028684496879577637, -0.03617803752422333, 0.0612257644534111, 0.01826908253133297, 0.08141545951366425, -0.007534825708717108, 0.00023949053138494492, 0.07520618289709091, -0.13648591935634613, 0.005331330932676792, -0.07162231206893921, -0.1834970861673355, -0.00754373986274004, 0.04121732711791992, 0.08257578313350677, 0.05396924912929535, 0.13104914128780365, -0.002638587960973382, -0.035694170743227005, 0.042987190186977386, -0.045237936079502106, -0.006338424514979124, 0.003633941989392042, 0.008852201513946056, 0.11249765753746033, 0.010092301294207573, 0.025105977430939674, -0.14049497246742249, -0.011632162146270275, -0.049273695796728134, -0.003008487168699503, -0.03037283942103386, -0.041250329464673996, 0.014100132510066032, -0.02324768528342247, -0.012397333979606628, -0.164638489484787, -0.12902602553367615, -0.0023389337584376335, -0.013901285827159882, -0.0444142147898674, -0.1079728901386261, -0.0685032308101654, -0.01551440916955471, 0.03475569933652878, -0.06027122586965561, -0.03347735479474068, -0.04477924481034279, 0.12063781917095184, -0.029627691954374313, 0.09997887164354324, -0.1001875028014183, 0.06328949332237244, -0.08874991536140442, -0.050521478056907654, -0.12108279764652252, 0.12580174207687378, -0.02738279663026333, 0.08329804241657257, -0.04757985845208168, -0.0025342535227537155, -0.07068078219890594, 0.06470146775245667, -0.04831327870488167, 0.21021100878715515, -0.07818182557821274, -0.10016241669654846, 0.3416871428489685, -0.06889300048351288, -0.14127831161022186, 0.12289705872535706, 0.0038081686943769455, 0.07869209349155426, 0.1243860274553299, 0.18055181205272675, 0.01312235463410616, -0.017781920731067657, 0.047317296266555786, 0.0774398222565651, -0.10274933278560638, -0.03693215921521187, 0.012422840110957623, -0.023849958553910255, -0.05957121029496193, 0.03017471171915531, 0.10177447646856308, 0.07719460874795914, -0.037943217903375626, -0.021851979196071625, -0.011169109493494034, 0.012724953703582287, 0.05065817013382912, -0.017767585813999176, 0.10469906032085419, -0.05546741187572479, -0.08462046831846237, -0.0680401548743248, 0.028042905032634735, -0.03323547542095184, 0.03231019899249077, -0.09040408581495285, 0.07871139794588089, 0.002006636932492256, 0.07614494115114212, -0.13897794485092163, -0.12861210107803345, -0.029086988419294357, 0.1735243946313858, 0.08352918922901154, 0.06436270475387573, 0.05939723178744316, -0.07140430063009262, 0.008528050035238266, 0.026146668940782547, 0.17880438268184662, -0.030734572559595108, -0.05608346313238144, -0.08764305710792542, 0.11963928490877151, -0.0527319461107254, 0.07085329294204712, -0.029741745442152023, 0.020559802651405334, 0.07414128631353378, 0.12597008049488068, -0.012940989807248116, 0.018252748996019363, 0.014936812222003937, -0.013967443257570267, -0.05683703348040581, -0.02589830756187439, 0.07882234454154968, 0.006542436312884092, -0.0838664323091507, 0.23497329652309418, -0.1738751381635666, 0.12409135699272156, 0.19194376468658447, -0.20209380984306335, -0.012007046490907669, -0.07341687381267548, -0.021834038197994232, 0.01817040890455246, 0.04731811583042145, -0.0253558661788702, 0.19114437699317932, -0.003812632290646434, 0.1578192412853241, -0.03853379189968109, -0.04169050231575966, -0.024481579661369324, -0.07597297430038452, 0.006928789895027876, 0.09140373021364212, 0.10092125833034515, -0.1423201560974121, 0.19011947512626648, 0.06233326345682144, 0.03858935087919235, 0.20700646936893463, 0.025934487581253052, 0.02993733063340187, 0.05680325627326965, 0.023858970031142235, -0.018066244199872017, -0.07449664175510406, -0.23275378346443176, -0.027651702985167503, 0.07589344680309296, 0.035592179745435715, 0.09904105961322784, -0.10122398287057877, -0.033858124166727066, -0.0342395193874836, -0.030355559661984444, 0.06944359838962555, 0.131292462348938, 0.029469849541783333, 0.13573025166988373, -0.006529252044856548, -0.035037439316511154, 0.07874161005020142, 0.031415995210409164, -0.0705534815788269, 0.17797964811325073, -0.10885831713676453, -0.3646329343318939, -0.11388835310935974, -0.1974858194589615, -0.060813967138528824, 0.047129563987255096, 0.1280462145805359, -0.1640487164258957, -0.012251666747033596, 0.03736591711640358, 0.08337146788835526, -0.06711234152317047, 0.009085515514016151, -0.041123468428850174, 0.00414077565073967, -0.11542318761348724, -0.09824596345424652, -0.06377395987510681, -0.041316479444503784, -0.112151138484478, 0.14525772631168365, -0.11890920996665955, 0.037856243550777435, 0.21912351250648499, 0.03082382306456566, 0.07140220701694489, -0.046491336077451706, 0.17464914917945862, -0.11472119390964508, 0.007070933934301138, 0.2176927775144577, -0.010368889197707176, 0.056967414915561676, 0.10844351351261139, -0.01216966100037098, -0.09802498668432236, 0.026755837723612785, -0.0016287126345559955, -0.08681747317314148, -0.2405051290988922, -0.12135352194309235, -0.11196054518222809, 0.0661383867263794, 0.026356035843491554, 0.06405682861804962, 0.16846387088298798, 0.07280991971492767, -0.046640899032354355, 0.029674114659428596, 0.038152582943439484, 0.09783633798360825, 0.24335892498493195, -0.06171194463968277, 0.13966211676597595, -0.04837142676115036, -0.14445817470550537, 0.08735872060060501, 0.08486756682395935, 0.08985790610313416, 0.04785217344760895, 0.07958680391311646, 0.011308308690786362, 0.041056226938962936, 0.11411003768444061, 0.03848817199468613, 0.022578220814466476, -0.03239505738019943, -0.020139263942837715, -0.027479834854602814, -0.06456100940704346, 0.010104041546583176, 0.037413619458675385, -0.17863363027572632, 0.00205021258443594, -0.004038363695144653, 0.08311571925878525, 0.0330156646668911, 0.045852698385715485, -0.17027175426483154, -0.036130718886852264, 0.06191907078027725, -0.006051910109817982, -0.13006319105625153, 0.07477587461471558, 0.006824449170380831, -0.13764947652816772, 0.027440516278147697, -0.02765842340886593, 0.10892626643180847, -0.11904376745223999, 0.06142464280128479, -0.10844694823026657, -0.02456236258149147, 0.00007157027721405029, 0.11942675709724426, -0.2908390760421753, 0.1778879463672638, -0.008749104104936123, -0.03576286509633064, -0.12925903499126434, 0.0030134529806673527, 0.010048306547105312, 0.06613421440124512, 0.120755136013031, 0.003950454294681549, -0.05502857267856598, -0.011420777067542076, -0.07334607094526291, 0.033989232033491135, 0.08522196114063263, -0.038692619651556015, -0.026387467980384827, -0.03939871862530708, -0.000563579611480236, -0.029193514958024025, -0.10996408760547638, 0.031565725803375244, -0.1796351820230484, 0.09353981167078018, 0.10465142130851746, 0.07875864207744598, 0.02353157475590706, -0.012726625427603722, -0.05260153114795685, 0.2882421910762787, -0.01351880468428135, -0.1059127151966095, -0.10419327020645142, -0.0003387331962585449, 0.0323350764811039, -0.09219582378864288, 0.015146953985095024, -0.08285350352525711, 0.029172182083129883, -0.067306287586689, -0.1609160304069519, 0.1357487291097641, -0.08346674591302872, -0.02869018353521824, -0.0231779757887125, 0.20794923603534698, -0.0432368703186512, 0.010672476142644882, 0.03228606656193733, -0.005603097379207611, -0.09821572154760361, -0.10980522632598877, 0.03154990077018738, 0.0035563651472330093, -0.002707253908738494, 0.03121224045753479, -0.04100721329450607, -0.04626888036727905, -0.07337665557861328, -0.04023002088069916, 0.34542107582092285, 0.140069380402565, -0.041641734540462494, 0.16272948682308197, 0.07728812098503113, -0.0632505714893341, -0.2909311354160309, -0.11139822006225586, -0.08305023610591888, -0.07137535512447357, -0.12761490046977997, -0.19104674458503723, 0.07015315443277359, -0.010947578586637974, -0.02332664653658867, 0.10749930888414383, -0.24520817399024963, -0.09330083429813385, 0.16807708144187927, -0.0222132820636034, 0.3952651023864746, -0.12077606469392776, -0.08812063187360764, -0.04551239311695099, -0.1573183238506317, 0.17550544440746307, 0.017852526158094406, 0.11935380101203918, -0.003850414417684078, 0.19892863929271698, 0.050656337291002274, 0.00005161578883416951, 0.06399565935134888, 0.004039391875267029, -0.0486612543463707, -0.10825341939926147, -0.06550462543964386, 0.006332400254905224, 0.024469764903187752, 0.0318533256649971, -0.04529967159032822, 0.012598483823239803, -0.12255710363388062, -0.046386417001485825, -0.09297195822000504, 0.04288104176521301, 0.03612024337053299, -0.07072491198778152, -0.010932168923318386, -0.043201837688684464, -0.021154992282390594, 0.01627092994749546, 0.18928144872188568, -0.09368468821048737, 0.16298866271972656, 0.03140731528401375, 0.10047957301139832, -0.11105209589004517, -0.021548202261328697, -0.056601643562316895, -0.05869057774543762, 0.07274187356233597, -0.1264297068119049, 0.05027387663722038, 0.08291409909725189, -0.04217144101858139, 0.07602261006832123, 0.08426139503717422, -0.03202803060412407, 0.015913769602775574, 0.1064189225435257, -0.24266429245471954, -0.060369595885276794, -0.055411603301763535, 0.05804112181067467, 0.09530272334814072, 0.10321993380784988, 0.21187539398670197, 0.018716618418693542, -0.06238582730293274, 0.0043049403466284275, 0.02287164144217968, -0.03164453059434891, 0.07607007026672363, -0.008284146897494793, 0.02211715839803219, -0.15616396069526672, 0.0508871003985405, 0.03160223737359047, -0.10562913864850998, 0.054960958659648895, 0.17974784970283508, -0.10643400251865387, -0.1285153180360794, -0.04878600314259529, 0.09519313275814056, -0.13735073804855347, -0.003008528146892786, -0.053487617522478104, -0.10136730968952179, 0.07071103155612946, 0.09810495376586914, 0.03625602275133133, 0.056969426572322845, -0.05162336677312851, -0.025706449523568153, -0.06022496521472931, -0.015211719088256359, 0.017192507162690163, 0.0028863348998129368, -0.06666891276836395, 0.06586922705173492, -0.02776043675839901, 0.11565238982439041, -0.07972242683172226, -0.09763384610414505, -0.18344280123710632, 0.02236086130142212, -0.03455692529678345, -0.08797409385442734, -0.07886038720607758, -0.02799299918115139, 0.00046232063323259354, -0.040812794119119644, -0.029650017619132996, -0.02921629697084427, -0.1163771003484726, 0.013556803576648235, -0.04323334991931915, 0.013864753767848015, -0.08943917602300644, 0.019881298765540123, 0.06791353225708008, -0.023644810542464256, 0.17640171945095062, 0.14861774444580078, -0.09017820656299591, 0.08637912571430206, -0.16519515216350555, -0.04759880527853966, 0.11209218949079514, 0.014201316051185131, 0.048976801335811615, 0.05524056777358055, 0.031076371669769287, 0.053941767662763596, 0.05915273353457451, 0.058900073170661926, 0.03862825036048889, -0.10739199072122574, 0.06685130298137665, -0.05019306391477585, -0.12192682921886444, -0.03560705482959747, -0.0019091602880507708, 0.040228210389614105, 0.036408089101314545, 0.0813569650053978, -0.06774868071079254, 0.08131438493728638, -0.05310627445578575, 0.028316672891378403, 0.015001521445810795, -0.17129793763160706, 0.020272541791200638, -0.0674624890089035, 0.042005106806755066, 0.01551832165569067, 0.16965706646442413, 0.05189540237188339, -0.03604042902588844, 0.03249737620353699, 0.08931052684783936, 0.032809846103191376, -0.01604866422712803, 0.12605388462543488, 0.09557680040597916, -0.05276251211762428, -0.05967254936695099, 0.06976452469825745, 0.03793814405798912, 0.03542026877403259, 0.1245683953166008, -0.028732217848300934, 0.01033438928425312, 0.052373483777046204, 0.004065222106873989, 0.01687709428369999, -0.10472496598958969, -0.10712479054927826, -0.046430572867393494, 0.03501254320144653, -0.050667695701122284, 0.07078768312931061, 0.14960148930549622, -0.0017245551571249962, 0.01058112271130085, -0.024053970351815224, -0.05088517814874649, -0.18352080881595612, -0.19419948756694794, -0.08193299174308777, -0.14664407074451447, 0.017230965197086334, -0.11158008873462677, 0.05417042225599289, 0.008138729259371758, 0.07885850220918655, -0.04726981371641159, 0.09065236151218414, 0.018653184175491333, -0.10273414105176926, 0.038575172424316406, -0.0317155197262764, 0.06465325504541397, -0.018326494842767715, 0.012363510206341743, -0.07680509984493256, 0.03239038586616516, 0.030291138216853142, 0.06388211995363235, -0.022330155596137047, 0.02995169907808304, -0.15074686706066132, -0.07856905460357666, -0.05684714764356613, 0.06378791481256485, -0.014431370422244072, 0.15566974878311157, 0.024981670081615448, -0.03026868775486946, 0.03983297199010849, 0.2429601401090622, -0.060073066502809525, -0.0956934466958046, -0.06688141077756882, 0.17435267567634583, -0.00395948113873601, 0.08003787696361542, -0.05926486849784851, 0.006742642726749182, -0.07236688584089279, 0.331343412399292, 0.27552270889282227, -0.05485304817557335, 0.009799845516681671, -0.010692475363612175, 0.04171661660075188, 0.10187152028083801, 0.11669018864631653, 0.129445418715477, 0.27527952194213867, -0.05744016915559769, -0.04287269338965416, -0.009941516444087029, -0.01587585359811783, -0.10140544176101685, 0.0472479872405529, 0.03406533598899841, -0.0476858988404274, -0.020353399217128754, 0.09225169569253922, -0.24208074808120728, 0.0826031044125557, -0.11048878729343414, -0.18727442622184753, -0.07940401136875153, 0.009244399145245552, 0.09566988050937653, 0.02211848460137844, 0.10203203558921814, 0.014010714367032051, -0.09862406551837921, 0.06046474725008011, 0.025634337216615677, -0.23420298099517822, -0.01629374921321869, 0.06475675106048584, -0.0583508126437664, 0.023496776819229126, -0.032499466091394424, 0.07297383248806, 0.09535618126392365, 0.026735052466392517, -0.032150957733392715, 0.034866560250520706, -0.016611335799098015, -0.06898951530456543, 0.025716807693243027, 0.04629160463809967, 0.0027601029723882675, -0.06635551899671555, 0.08279188722372055, -0.11967107653617859, 0.015745442360639572, -0.015144294127821922, 0.0016120974905788898, -0.024715179577469826, 0.04185955226421356, -0.08141392469406128, 0.07370835542678833, 0.08167718350887299, -0.017675034701824188, 0.00234060175716877, -0.05448093265295029, -0.0189832653850317, -0.0062657869420945644, -0.08579414337873459, -0.11466328054666519, -0.16684934496879578, -0.12536461651325226, 0.07908150553703308, 0.011286891996860504, -0.17560383677482605, 0.023233657702803612, -0.11535480618476868, 0.05935998260974884, -0.15454533696174622, 0.10524857044219971, 0.07884759455919266, 0.019195739179849625, -0.00028656492941081524, -0.008156108669936657, 0.041265737265348434, 0.0845412164926529, -0.1166185587644577, -0.06798484921455383 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.2125 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.2637 | 1.0 | 5533 | 1.2125 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["squad"]}
question-answering
avioo1/distilbert-base-uncased-finetuned-squad
[ "transformers", "pytorch", "tensorboard", "distilbert", "question-answering", "generated_from_trainer", "dataset:squad", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-squad ======================================= This model is a fine-tuned version of distilbert-base-uncased on the squad dataset. It achieves the following results on the evaluation set: * Loss: 1.2125 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.10.2 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ 56, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ -0.1081218421459198, 0.08665481954813004, -0.0019734499510377645, 0.11980724334716797, 0.1639399528503418, 0.023396607488393784, 0.09885699301958084, 0.12226859480142593, -0.09512254595756531, 0.031825821846723557, 0.13206203281879425, 0.17120274901390076, 0.0016580113442614675, 0.06159606948494911, -0.04702337458729744, -0.21845468878746033, -0.0200874675065279, 0.05637078359723091, -0.0998520702123642, 0.14527656137943268, 0.08514146506786346, -0.14992168545722961, 0.07109404355287552, 0.004433309193700552, -0.21555066108703613, 0.014025719836354256, 0.004071226343512535, -0.03754277154803276, 0.14436578750610352, 0.00319460011087358, 0.12367971986532211, -0.0004823615017812699, 0.06790561228990555, -0.18166689574718475, 0.014116760343313217, 0.05067618936300278, 0.010513479821383953, 0.08459962904453278, 0.04910590127110481, 0.00584397604689002, 0.10436712950468063, -0.08712361752986908, 0.040377020835876465, 0.027254557237029076, -0.12948258221149445, -0.2573572099208832, -0.10456980019807816, 0.014765523374080658, 0.06322389841079712, 0.12137169390916824, -0.005105495452880859, 0.15751895308494568, -0.11296563595533371, 0.08846994489431381, 0.25568053126335144, -0.29454702138900757, -0.06910262256860733, 0.030112653970718384, 0.025582637637853622, 0.06914079934358597, -0.1076497957110405, -0.03316973149776459, 0.05347539111971855, 0.052304819226264954, 0.10938668251037598, -0.04493562504649162, -0.12636876106262207, 0.03803937882184982, -0.14875921607017517, -0.044769804924726486, 0.1436338722705841, 0.04647607356309891, -0.026044320315122604, -0.023581044748425484, -0.06251414865255356, -0.1332535296678543, -0.024055421352386475, -0.018303345888853073, 0.047980282455682755, -0.04684578999876976, -0.08232725411653519, -0.0037009334191679955, -0.10676536709070206, -0.08301205188035965, -0.07542102038860321, 0.13565035164356232, 0.04100826010107994, 0.03072817251086235, -0.050032276660203934, 0.1018606424331665, 0.0073599060997366905, -0.13160544633865356, 0.015266901813447475, 0.036453112959861755, -0.017884666100144386, -0.03383035585284233, -0.06546676903963089, -0.05835319310426712, 0.02570483647286892, 0.12147531658411026, -0.07889924943447113, 0.031884320080280304, 0.05587765574455261, 0.042132697999477386, -0.08704063296318054, 0.18057988584041595, -0.07136484235525131, -0.004164637997746468, -0.005728185176849365, 0.03805986046791077, 0.0012731902534142137, 0.0018262439407408237, -0.09711797535419464, -0.007312555331736803, 0.09382573515176773, 0.020405929535627365, -0.03521427884697914, 0.05215555801987648, -0.050382889807224274, -0.025741178542375565, -0.0011978837428614497, -0.08454166352748871, 0.028766319155693054, -0.002292028395459056, -0.09215663373470306, -0.013462753966450691, 0.008795255795121193, 0.013695497997105122, -0.015977857634425163, 0.09721996635198593, -0.09437794983386993, 0.03848237544298172, -0.09642738848924637, -0.09927541017532349, 0.02544553205370903, -0.0888921469449997, 0.02771683596074581, -0.07935667783021927, -0.15798419713974, -0.011141447350382805, 0.05134735256433487, -0.02508920058608055, -0.05001704767346382, -0.03313067555427551, -0.09069399535655975, -0.012242726050317287, -0.018107281997799873, 0.15977491438388824, -0.058648571372032166, 0.11928404122591019, 0.03992374241352081, 0.06538859754800797, -0.044009238481521606, 0.0600406751036644, -0.10696934163570404, 0.015341261401772499, -0.16976580023765564, 0.037300679832696915, -0.055807095021009445, 0.062445610761642456, -0.10233812034130096, -0.12232377380132675, 0.024136003106832504, -0.019169503822922707, 0.08626240491867065, 0.09793485701084137, -0.1692907065153122, -0.06082156300544739, 0.1447843611240387, -0.052574507892131805, -0.14480900764465332, 0.12083259969949722, -0.0549415722489357, 0.033790186047554016, 0.06326605379581451, 0.1686031073331833, 0.05496174842119217, -0.08920158445835114, 0.01113475114107132, -0.0004416894807945937, 0.03185898810625076, -0.07942872494459152, 0.07141144573688507, -0.009497782215476036, 0.028299592435359955, 0.02309240773320198, -0.05806373804807663, 0.0566115528345108, -0.11394322663545609, -0.09684354066848755, -0.05239700525999069, -0.10580629855394363, 0.043173160403966904, 0.0889706164598465, 0.06952431052923203, -0.09907278418540955, -0.07060075551271439, 0.07938767224550247, 0.07113073766231537, -0.055893294513225555, 0.03146578371524811, -0.06048053503036499, 0.06982652097940445, -0.07094525545835495, -0.02904580719769001, -0.19921325147151947, -0.02969740331172943, 0.004683692939579487, -0.0020540624391287565, 0.01240891870111227, 0.04712724685668945, 0.07515227794647217, 0.04037543758749962, -0.05255388095974922, -0.021131521090865135, -0.052795231342315674, -0.007310677319765091, -0.1329045295715332, -0.1885766088962555, -0.038912829011678696, -0.013390184380114079, 0.10179212689399719, -0.17873089015483856, 0.024664180353283882, -0.013663223944604397, 0.06831377744674683, -0.007877247408032417, -0.01327646616846323, -0.037831224501132965, 0.08453086018562317, -0.018599778413772583, -0.0496964268386364, 0.07715803384780884, -0.003927280660718679, -0.09106199443340302, -0.06069771945476532, -0.06030641123652458, 0.14625611901283264, 0.131173238158226, -0.12017800658941269, -0.06362443417310715, 0.008861520327627659, -0.07353880256414413, -0.03901531174778938, -0.04373374953866005, 0.04130929708480835, 0.17824123799800873, -0.0027841750998049974, 0.12313341349363327, -0.0825551226735115, -0.05014725401997566, 0.012605486437678337, -0.0345822274684906, 0.04358314722776413, 0.1301945596933365, 0.12293878197669983, -0.0606050081551075, 0.14348477125167847, 0.15506622195243835, -0.09422797709703445, 0.10001020133495331, -0.06509991735219955, -0.0882989764213562, -0.033961221575737, 0.001190973212942481, -0.006756638642400503, 0.12274196743965149, -0.14880073070526123, 0.018888024613261223, 0.03367273136973381, 0.019786687567830086, 0.025890421122312546, -0.2328421175479889, -0.0611376129090786, 0.018306899815797806, -0.0507877841591835, -0.031325314193964005, -0.003642821218818426, 0.02237456664443016, 0.10104469209909439, -0.005124136805534363, -0.05913940817117691, 0.04532894119620323, -0.0004813089908566326, -0.06607043743133545, 0.21950885653495789, -0.073747918009758, -0.12041930854320526, -0.09713555127382278, -0.043125398457050323, -0.049855198711156845, -0.00621896842494607, 0.06714016944169998, -0.09321453422307968, -0.014379733242094517, -0.04691339284181595, 0.02238789014518261, -0.010215063579380512, 0.02161562815308571, -0.0005535397212952375, -0.004709654953330755, 0.07697850465774536, -0.12056206911802292, 0.006512482184916735, -0.05737905576825142, -0.07690393179655075, 0.0545320101082325, 0.0569911003112793, 0.13066992163658142, 0.14442875981330872, -0.015943095088005066, 0.007530089467763901, -0.020732488483190536, 0.2513790428638458, -0.07096070051193237, -0.03867470100522041, 0.1433461457490921, 0.015521069057285786, 0.0599655918776989, 0.10294348001480103, 0.0752970427274704, -0.09161365032196045, 0.005034042522311211, 0.03880704194307327, -0.039083972573280334, -0.2450471669435501, -0.033955540508031845, -0.05854189395904541, -0.025747135281562805, 0.07264456152915955, 0.024175848811864853, 0.03541618213057518, 0.07230924069881439, 0.04315287619829178, 0.043026525527238846, -0.07340579479932785, 0.04067510738968849, 0.10978464037179947, 0.04478071630001068, 0.11445321887731552, -0.04783743619918823, -0.05945519730448723, 0.02840142324566841, -0.0013929489068686962, 0.24557280540466309, -0.010373773984611034, 0.1492546945810318, 0.08255577832460403, 0.21919982135295868, -0.015662292018532753, 0.08202389627695084, -0.011222978122532368, -0.04746970534324646, -0.004286656156182289, -0.034542474895715714, -0.02797832153737545, 0.005259878467768431, -0.0394345298409462, 0.07226461172103882, -0.09800007194280624, -0.01714972034096718, 0.056878507137298584, 0.27427563071250916, 0.027786672115325928, -0.2979797422885895, -0.08845202624797821, -0.01001039333641529, -0.023565778508782387, -0.0058179013431072235, 0.01871062070131302, 0.1225421130657196, -0.09387841820716858, -0.0011919678654521704, -0.06878715008497238, 0.1002921462059021, -0.012080804444849491, 0.03788299858570099, 0.0765715166926384, 0.09072401374578476, 0.016117878258228302, 0.09232646971940994, -0.31108158826828003, 0.27099379897117615, 0.0013447041856124997, 0.07491617649793625, -0.07760810852050781, -0.014324107207357883, 0.008295329287648201, 0.034085486084222794, 0.08296117186546326, -0.006285442505031824, 0.014393053017556667, -0.17444519698619843, -0.04367279261350632, 0.03750208020210266, 0.08126477897167206, -0.01888776384294033, 0.09429117292165756, -0.011811821721494198, 0.015899036079645157, 0.07323213666677475, 0.00270519545301795, -0.04621556028723717, -0.07692700624465942, -0.015135341323912144, 0.012021761387586594, -0.05156589299440384, -0.07029179483652115, -0.10441756248474121, -0.11893629282712936, 0.12350057810544968, -0.004377948120236397, -0.04088025167584419, -0.10730171203613281, 0.09088611602783203, 0.10854675620794296, -0.08851844817399979, 0.03394761681556702, 0.012049267068505287, 0.03181932494044304, 0.0428953655064106, -0.06132921576499939, 0.09757273644208908, -0.05803560093045235, -0.15640266239643097, -0.04401674494147301, 0.1124778538942337, 0.05272740125656128, 0.06434956938028336, -0.015271629206836224, 0.0166313536465168, -0.05667118355631828, -0.1037694588303566, 0.022547651082277298, -0.035953495651483536, 0.09055644273757935, 0.025926468893885612, -0.029216010123491287, 0.05813150852918625, -0.05906000733375549, -0.02282133139669895, 0.19183550775051117, 0.2307959944009781, -0.10240183025598526, 0.014129981398582458, 0.03359711170196533, -0.04860355332493782, -0.18610155582427979, 0.05305757746100426, 0.0688682571053505, -0.01094765029847622, 0.05082383006811142, -0.16024133563041687, 0.1548338085412979, 0.11830127239227295, -0.006534809712320566, 0.11852594465017319, -0.37225884199142456, -0.11422576755285263, 0.09655865281820297, 0.16073371469974518, 0.11982813477516174, -0.16493381559848785, -0.022821074351668358, 0.0008885537972673774, -0.16979210078716278, 0.11129388213157654, -0.09309817850589752, 0.11733362078666687, -0.0414404459297657, 0.10592722147703171, -0.001169731141999364, -0.07317209243774414, 0.12963533401489258, 0.03898831456899643, 0.10103604942560196, -0.04704327508807182, -0.03086920641362667, 0.08543924987316132, -0.01985197514295578, 0.02131439559161663, -0.060324300080537796, 0.040571726858615875, -0.1036156713962555, -0.011054204776883125, -0.10940517485141754, 0.04329053312540054, -0.04328712821006775, -0.0564606711268425, -0.04059953615069389, 0.023238033056259155, 0.04456917196512222, -0.012785718776285648, 0.12127390503883362, 0.03220447897911072, 0.1454075425863266, 0.0869474783539772, 0.06670873612165451, -0.06385386735200882, -0.12035218626260757, -0.016045207157731056, -0.0049170833081007, 0.0543304979801178, -0.13645364344120026, 0.024005519226193428, 0.15614794194698334, 0.05192888528108597, 0.12022751569747925, 0.07693500816822052, -0.028910323977470398, 0.01180249359458685, 0.037836432456970215, -0.16220593452453613, -0.13762053847312927, 0.014298103749752045, -0.0718918964266777, -0.11780636757612228, 0.06009784713387489, 0.06362926959991455, -0.052895549684762955, -0.01326073333621025, -0.005142338573932648, -0.0039098565466701984, -0.06673404574394226, 0.2082156091928482, 0.08383913338184357, 0.05164876580238342, -0.11374913156032562, 0.0798710286617279, 0.04599163681268692, -0.0929858535528183, -0.011240682564675808, 0.06079668924212456, -0.06972391158342361, -0.04396234452724457, 0.09480777382850647, 0.15585274994373322, -0.06101103499531746, -0.03741893917322159, -0.1308615356683731, -0.10986242443323135, 0.07779307663440704, 0.14409682154655457, 0.11349353939294815, 0.011756113730370998, -0.04643531143665314, 0.010767620988190174, -0.12299423664808273, 0.08372405171394348, 0.036913927644491196, 0.06221906095743179, -0.12288957834243774, 0.13897201418876648, 0.006175137124955654, 0.06337575614452362, -0.01743401773273945, 0.03210284560918808, -0.09809472411870956, 0.030463434755802155, -0.14294403791427612, -0.036050133407115936, -0.03166112303733826, -0.004696040414273739, -0.010352849029004574, -0.08228690177202225, -0.062434904277324677, 0.023924555629491806, -0.1229245588183403, -0.023349441587924957, 0.04866015166044235, 0.044385869055986404, -0.14252646267414093, -0.04566698148846626, 0.03419461101293564, -0.04975327104330063, 0.0664161965250969, 0.06847658008337021, 0.011976595036685467, 0.053476136177778244, -0.13489589095115662, -0.01781543344259262, 0.050109706819057465, 0.014086516574025154, 0.08260184526443481, -0.09380419552326202, -0.020469026640057564, 0.00736933434382081, 0.06879912316799164, 0.016679419204592705, 0.04268522188067436, -0.14330565929412842, -0.01261149812489748, -0.02738211490213871, -0.07681136578321457, -0.07512322813272476, 0.01057496853172779, 0.10057606548070908, 0.03270503878593445, 0.19699257612228394, -0.05683174356818199, 0.06238596886396408, -0.2191339135169983, -0.009507371112704277, -0.013630986213684082, -0.09526431560516357, -0.10931959748268127, -0.042956508696079254, 0.06804070621728897, -0.05870629847049713, 0.12628938257694244, -0.00819423608481884, 0.05801478028297424, 0.023015646263957024, -0.01209248136729002, 0.029393712058663368, 0.013943925499916077, 0.2383275181055069, 0.019439268857240677, -0.03170947730541229, 0.08223096281290054, 0.06038239225745201, 0.0726848766207695, 0.11793136596679688, 0.2072349488735199, 0.17512357234954834, 0.008602574467658997, 0.07085277885198593, 0.03218189626932144, -0.047193095088005066, -0.1358916163444519, 0.04539402201771736, -0.023872215300798416, 0.08175479620695114, -0.01617514155805111, 0.24275150895118713, 0.05781218037009239, -0.17501667141914368, 0.06012890487909317, -0.05895395949482918, -0.09345562011003494, -0.08303852379322052, -0.024595966562628746, -0.06639083474874496, -0.14998646080493927, 0.010731574147939682, -0.12337371706962585, 0.017469631507992744, 0.13779999315738678, 0.009309954009950161, -0.031756170094013214, 0.18237464129924774, 0.05386216938495636, 0.028323635458946228, 0.038396693766117096, 0.00015905086183920503, -0.023278551176190376, -0.08477957546710968, -0.04328535124659538, -0.0012371264165267348, -0.024252571165561676, 0.04320726543664932, -0.0495903454720974, -0.07993009686470032, 0.032723721116781235, -0.03843919187784195, -0.0944444015622139, 0.005481781903654337, 0.034556902945041656, 0.06903433799743652, 0.051854051649570465, 0.017446743324398994, 0.03269825130701065, -0.023876242339611053, 0.2190386950969696, -0.07389110326766968, -0.08762068301439285, -0.08995087444782257, 0.24980758130550385, 0.03202006220817566, -0.022533588111400604, 0.04377151280641556, -0.0696730986237526, 0.0051801553927361965, 0.24647904932498932, 0.17791016399860382, -0.1007506251335144, -0.013419854454696178, 0.007877500727772713, -0.011119559407234192, -0.03629665449261665, 0.08729850500822067, 0.14436599612236023, 0.04726633056998253, -0.10995219647884369, -0.04340929165482521, -0.07759195566177368, -0.012929943390190601, -0.043325379490852356, 0.050642356276512146, 0.050543058663606644, -0.0048489137552678585, -0.04221436753869057, 0.06570209562778473, -0.06705861538648605, -0.12491151690483093, 0.07768802344799042, -0.1936170905828476, -0.15915779769420624, -0.019773516803979874, 0.1128520742058754, 0.0045839399099349976, 0.06400315463542938, -0.03425842151045799, 0.012641509994864464, 0.07460948824882507, -0.015347023494541645, -0.09812062233686447, -0.07958126813173294, 0.1167924627661705, -0.12642234563827515, 0.18590541183948517, -0.04057910665869713, 0.08584952354431152, 0.1289622038602829, 0.060171157121658325, -0.0875810831785202, 0.06347353011369705, 0.05980094522237778, -0.09678692370653152, 0.013633696362376213, 0.07948533445596695, -0.010029087774455547, 0.03690804913640022, 0.03698736056685448, -0.12027401477098465, 0.011925666593015194, -0.040093835443258286, -0.023373931646347046, -0.06548340618610382, -0.039182644337415695, -0.05795421823859215, 0.12280209362506866, 0.21267710626125336, -0.03936230391263962, 0.014934880658984184, -0.08123168349266052, 0.008982906118035316, 0.05511922389268875, 0.025991395115852356, -0.0755775198340416, -0.20985467731952667, 0.028076400980353355, 0.05395406484603882, -0.03414290398359299, -0.20089465379714966, -0.08983114361763, 0.02398407645523548, -0.08847593516111374, -0.06494580209255219, 0.06005621328949928, 0.07082279026508331, 0.059760719537734985, -0.04549173638224602, -0.06823209673166275, -0.08886721730232239, 0.16068138182163239, -0.15138830244541168, -0.08447614312171936 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-squad2-finetuned-squad This model is a fine-tuned version of [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 5.0220 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 74 | 1.7148 | | No log | 2.0 | 148 | 1.6994 | | No log | 3.0 | 222 | 1.7922 | | No log | 4.0 | 296 | 1.9947 | | No log | 5.0 | 370 | 2.0753 | | No log | 6.0 | 444 | 2.2096 | | 0.9547 | 7.0 | 518 | 2.3070 | | 0.9547 | 8.0 | 592 | 2.6947 | | 0.9547 | 9.0 | 666 | 2.7169 | | 0.9547 | 10.0 | 740 | 2.8503 | | 0.9547 | 11.0 | 814 | 3.1990 | | 0.9547 | 12.0 | 888 | 3.4931 | | 0.9547 | 13.0 | 962 | 3.6575 | | 0.3191 | 14.0 | 1036 | 3.1863 | | 0.3191 | 15.0 | 1110 | 3.7922 | | 0.3191 | 16.0 | 1184 | 3.6336 | | 0.3191 | 17.0 | 1258 | 4.1156 | | 0.3191 | 18.0 | 1332 | 4.1353 | | 0.3191 | 19.0 | 1406 | 3.9888 | | 0.3191 | 20.0 | 1480 | 4.4290 | | 0.1904 | 21.0 | 1554 | 4.0473 | | 0.1904 | 22.0 | 1628 | 4.5048 | | 0.1904 | 23.0 | 1702 | 4.4026 | | 0.1904 | 24.0 | 1776 | 4.2864 | | 0.1904 | 25.0 | 1850 | 4.3941 | | 0.1904 | 26.0 | 1924 | 4.4921 | | 0.1904 | 27.0 | 1998 | 4.9139 | | 0.1342 | 28.0 | 2072 | 4.8914 | | 0.1342 | 29.0 | 2146 | 5.0148 | | 0.1342 | 30.0 | 2220 | 5.0220 | ### Framework versions - Transformers 4.11.0 - Pytorch 1.9.0+cu102 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "roberta-base-squad2-finetuned-squad", "results": []}]}
question-answering
avioo1/roberta-base-squad2-finetuned-squad
[ "transformers", "pytorch", "tensorboard", "roberta", "question-answering", "generated_from_trainer", "license:cc-by-4.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #question-answering #generated_from_trainer #license-cc-by-4.0 #endpoints_compatible #region-us
roberta-base-squad2-finetuned-squad =================================== This model is a fine-tuned version of deepset/roberta-base-squad2 on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 5.0220 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0001 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 30 ### Training results ### Framework versions * Transformers 4.11.0 * Pytorch 1.9.0+cu102 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #question-answering #generated_from_trainer #license-cc-by-4.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 50, 97, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #roberta #question-answering #generated_from_trainer #license-cc-by-4.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 30### Training results### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.09406963735818863, 0.04242470860481262, -0.0017767370445653796, 0.11381895840167999, 0.17902567982673645, 0.028026478365063667, 0.11019888520240784, 0.09754886478185654, -0.10873712599277496, 0.043134111911058426, 0.12573936581611633, 0.16602644324302673, -0.00579972006380558, 0.048085927963256836, -0.05289237201213837, -0.23571202158927917, -0.036072321236133575, 0.04895747825503349, -0.10335659980773926, 0.13871072232723236, 0.07473985105752945, -0.16192837059497833, 0.06590133160352707, -0.0017634426476433873, -0.26464876532554626, 0.01761934719979763, 0.020339027047157288, -0.042714498937129974, 0.15058134496212006, 0.00426477799192071, 0.16577577590942383, 0.006694907788187265, 0.09752406924962997, -0.1647469699382782, 0.019062642008066177, 0.056625355035066605, 0.002818516455590725, 0.08689891546964645, 0.05664391070604324, -0.0004730811924673617, 0.11459105461835861, -0.088060162961483, 0.05745360255241394, 0.016974033787846565, -0.13231292366981506, -0.21899476647377014, -0.09627409279346466, 0.011552765034139156, 0.04448622092604637, 0.11402047425508499, -0.004412128124386072, 0.17107954621315002, -0.11306751519441605, 0.08834784477949142, 0.24315489828586578, -0.2914748191833496, -0.08329875767230988, 0.07464201748371124, 0.02010667324066162, 0.08523689955472946, -0.1194101870059967, -0.01863471232354641, 0.06159435212612152, 0.04511438310146332, 0.11408121138811111, -0.043075717985630035, -0.10637883096933365, 0.039546556770801544, -0.1549232006072998, -0.0220597293227911, 0.06566011160612106, 0.03702383115887642, -0.026825621724128723, -0.0008026765426620841, -0.06387614458799362, -0.14777031540870667, -0.039229948073625565, -0.02109694294631481, 0.05213010311126709, -0.05785985291004181, -0.13252514600753784, -0.0031259276438504457, -0.10573360323905945, -0.08631199598312378, -0.08155660331249237, 0.17110508680343628, 0.04037303104996681, 0.028844086453318596, -0.039903104305267334, 0.10699184238910675, -0.011711268685758114, -0.13114981353282928, 0.036151185631752014, 0.038204438984394073, -0.03279220685362816, -0.05677204951643944, -0.06362365186214447, -0.0756438598036766, 0.016127096489071846, 0.09581225365400314, -0.07399842143058777, 0.03296200931072235, 0.04982186481356621, 0.04977969825267792, -0.10240131616592407, 0.1834980696439743, -0.06097592040896416, -0.003008500672876835, -0.0032905403058975935, 0.04303227365016937, -0.014731236733496189, 0.0010085429530590773, -0.10218128561973572, -0.0098393764346838, 0.08819124102592468, 0.012864305637776852, -0.04920068010687828, 0.05449981987476349, -0.031055333092808723, -0.010209644213318825, -0.03991411253809929, -0.07906235754489899, 0.03979339823126793, -0.006328369956463575, -0.08932360261678696, 0.003220826620236039, 0.010270611383020878, 0.005455743987113237, 0.002028018468990922, 0.15150754153728485, -0.10629364103078842, 0.05065777897834778, -0.11819648742675781, -0.11240491271018982, 0.013864073902368546, -0.0713539570569992, 0.01760539598762989, -0.0904775932431221, -0.11939392238855362, -0.02366311103105545, 0.05329688638448715, -0.03210170939564705, -0.02533145435154438, -0.03281063213944435, -0.09979840368032455, -0.011378387920558453, -0.015088462270796299, 0.17518801987171173, -0.050912536680698395, 0.13524241745471954, 0.04559943825006485, 0.06659266352653503, -0.04935294762253761, 0.04302217438817024, -0.08389157056808472, 0.004841319285333157, -0.1995842307806015, 0.02607138082385063, -0.062837615609169, 0.07148560136556625, -0.08735035359859467, -0.12938468158245087, 0.012804996222257614, -0.011717468500137329, 0.09724882990121841, 0.08487673103809357, -0.16113078594207764, -0.06652548164129257, 0.15187819302082062, -0.04620172455906868, -0.11809081584215164, 0.12388405948877335, -0.06871569901704788, 0.03183436021208763, 0.06575355678796768, 0.15273188054561615, 0.03435808792710304, -0.11958210170269012, 0.018021566793322563, -0.02039162814617157, 0.03454591706395149, -0.07471822202205658, 0.04407474026083946, -0.0007022269419394433, 0.021076442673802376, 0.015469778329133987, -0.05782822519540787, 0.05245387181639671, -0.13243544101715088, -0.08699891716241837, -0.04729869216680527, -0.10208710283041, 0.03062458150088787, 0.08894111216068268, 0.08013120293617249, -0.10410383343696594, -0.07357213646173477, 0.0916602611541748, 0.05851832404732704, -0.05844302847981453, 0.026343287900090218, -0.06363179534673691, 0.07107128202915192, -0.07985322177410126, -0.03767490014433861, -0.19631880521774292, -0.03480639308691025, -0.009748687036335468, 0.03988715633749962, 0.044197723269462585, 0.04852336645126343, 0.08105350285768509, 0.04527309909462929, -0.06351146101951599, 0.00029867360717616975, -0.05788235366344452, -0.010883807204663754, -0.13859345018863678, -0.202894389629364, -0.030712660402059555, -0.00857169646769762, 0.07189665734767914, -0.1919100135564804, 0.018160227686166763, -0.019207103177905083, 0.07302700728178024, -0.0016667143208906054, -0.011990062892436981, -0.06015263497829437, 0.10426528751850128, -0.005442062392830849, -0.04432878643274307, 0.06719785183668137, -0.009570191614329815, -0.07800492644309998, -0.07797885686159134, -0.08763391524553299, 0.17105865478515625, 0.14587488770484924, -0.15566398203372955, -0.09417161345481873, 0.0167611762881279, -0.06534525007009506, -0.02150038257241249, -0.05232752859592438, 0.05186242237687111, 0.1878892034292221, -0.008382616564631462, 0.1325201690196991, -0.07662267982959747, -0.0450720340013504, 0.013908904045820236, -0.03267839923501015, 0.05911296606063843, 0.11548249423503876, 0.14421850442886353, -0.06744486838579178, 0.12779806554317474, 0.1338960826396942, -0.12950094044208527, 0.09849406778812408, -0.05166483670473099, -0.09023638069629669, -0.03553035855293274, -0.0023659199941903353, 0.008839521557092667, 0.13654154539108276, -0.11082549393177032, -0.00386830884963274, 0.009578163735568523, 0.008279509842395782, 0.027930203825235367, -0.2460235357284546, -0.06372284144163132, 0.01533031091094017, -0.033193331211805344, -0.02348012663424015, -0.011980254203081131, 0.02281000465154648, 0.11032447963953018, -0.0005339960916899145, -0.07071958482265472, 0.019913503900170326, -0.0030166665092110634, -0.0604800209403038, 0.2191140055656433, -0.04924655333161354, -0.06283410638570786, -0.09508664160966873, -0.05707574635744095, -0.029839521273970604, -0.0038621751591563225, 0.04222021624445915, -0.09882689267396927, -0.017584580928087234, -0.03199571371078491, 0.028631769120693207, -0.013457090593874454, 0.03894655033946037, 0.0005893866764381528, 0.00456228107213974, 0.06759107857942581, -0.11751603335142136, 0.00858035497367382, -0.07593918591737747, -0.09033092856407166, 0.0595308281481266, 0.06626590341329575, 0.13899526000022888, 0.16177518665790558, -0.03747430071234703, 0.0022771877702325583, -0.02455778606235981, 0.24962545931339264, -0.08303231000900269, -0.05626114830374718, 0.11548037081956863, 0.0024502617307007313, 0.06009283289313316, 0.09848236292600632, 0.09647353738546371, -0.10162299126386642, 0.01046848390251398, 0.040493547916412354, -0.04314684867858887, -0.2246202826499939, -0.02601279877126217, -0.057510100305080414, -0.052752893418073654, 0.07138890773057938, 0.0242652278393507, 0.026342052966356277, 0.0726151168346405, 0.0570095032453537, 0.06791368126869202, -0.07164685428142548, 0.05436026677489281, 0.09324698150157928, 0.05558593198657036, 0.12135065346956253, -0.04912814125418663, -0.08154264092445374, 0.01849021576344967, -0.023185137659311295, 0.25321850180625916, -0.005225470755249262, 0.1146511435508728, 0.07873951643705368, 0.20970208942890167, 0.009099009446799755, 0.07111851871013641, -0.006910139229148626, -0.06417142599821091, 0.0019592640455812216, -0.038750894367694855, -0.016471995040774345, -0.0026363711804151535, -0.029180720448493958, 0.05946236848831177, -0.09513475745916367, -0.02896931581199169, 0.05637873709201813, 0.24126701056957245, 0.011157186701893806, -0.30388301610946655, -0.0731375589966774, -0.01789357140660286, -0.02746463008224964, 0.004583083558827639, 0.006656333804130554, 0.12286857515573502, -0.08461174368858337, -0.007301919627934694, -0.07708089798688889, 0.0999283418059349, -0.010843868367373943, 0.036073729395866394, 0.0638163685798645, 0.1149991899728775, -0.004625000990927219, 0.06537855416536331, -0.3115754723548889, 0.2932099401950836, 0.00902959518134594, 0.09515944868326187, -0.0616946667432785, -0.027803895995020866, 0.01525940652936697, 0.02202385663986206, 0.06005009636282921, -0.022824252024292946, 0.018157223239541054, -0.19333067536354065, -0.018708789721131325, 0.03733548894524574, 0.12831763923168182, -0.002773882821202278, 0.10788879543542862, -0.00995351281017065, 0.00793538149446249, 0.08297289162874222, -0.010985812172293663, -0.047875139862298965, -0.0632045567035675, -0.027290010824799538, -0.0081156175583601, -0.06815242767333984, -0.05849704518914223, -0.11081317067146301, -0.14786170423030853, 0.1189437061548233, -0.001257095718756318, -0.012984328903257847, -0.11123333871364594, 0.11065644025802612, 0.10138645023107529, -0.07874772697687149, 0.04178126901388168, 0.022285543382167816, 0.026984956115484238, 0.030772510915994644, -0.053412552922964096, 0.10360098630189896, -0.050535935908555984, -0.15313787758350372, -0.05012376978993416, 0.0970134437084198, 0.044887036085128784, 0.07091809809207916, -0.015780862420797348, 0.0225539468228817, -0.04391346499323845, -0.1083398163318634, 0.028106877580285072, -0.05893541872501373, 0.08565684407949448, 0.02043965831398964, -0.02509397640824318, 0.03712327033281326, -0.05611513555049896, -0.010227678343653679, 0.17263683676719666, 0.2511765658855438, -0.0998162254691124, -0.01701190508902073, 0.03288964554667473, -0.04695209488272667, -0.18686364591121674, 0.09310439974069595, 0.07275041937828064, -0.0050689601339399815, 0.04916774854063988, -0.1350022554397583, 0.13966532051563263, 0.10855622589588165, -0.004757722839713097, 0.11066677421331406, -0.3518671989440918, -0.12417694181203842, 0.07093600183725357, 0.18888325989246368, 0.12914884090423584, -0.16288349032402039, -0.01846197433769703, -0.0000988005122053437, -0.15870386362075806, 0.10779517143964767, -0.08585190027952194, 0.11996515095233917, -0.02582416869699955, 0.11925461888313293, 0.005836689379066229, -0.0797254741191864, 0.11579836905002594, 0.009940942749381065, 0.11801311373710632, -0.05142604187130928, -0.051330819725990295, 0.08182519674301147, -0.01220253948122263, -0.007698139175772667, -0.03127490356564522, 0.02814456634223461, -0.05159112811088562, -0.01667197234928608, -0.10340756922960281, 0.03781129792332649, -0.042348530143499374, -0.059786632657051086, -0.04714375361800194, 0.024020269513130188, 0.03201385587453842, -0.025519929826259613, 0.11389032006263733, 0.011302166618406773, 0.16641269624233246, 0.06700623780488968, 0.071835458278656, -0.05477248877286911, -0.0989617258310318, 0.009491453878581524, -0.009576477110385895, 0.061493877321481705, -0.1493910551071167, 0.014015070162713528, 0.1587696671485901, 0.054673194885253906, 0.11077358573675156, 0.08856991678476334, -0.03545476868748665, 0.02115252986550331, 0.05720829591155052, -0.1548675149679184, -0.11717259883880615, 0.02440483123064041, -0.08273105323314667, -0.09836500138044357, 0.055339690297842026, 0.06451098620891571, -0.0674586221575737, -0.005395210348069668, 0.004140303935855627, -0.02961229346692562, -0.07153575867414474, 0.21044619381427765, 0.09308239817619324, 0.04787071421742439, -0.10392377525568008, 0.06096505746245384, 0.05295921117067337, -0.10194015502929688, -0.022999797016382217, 0.06565828621387482, -0.06616674363613129, -0.020207326859235764, 0.1081719920039177, 0.17592893540859222, -0.0865960642695427, -0.006635175086557865, -0.14768604934215546, -0.10769100487232208, 0.0646696537733078, 0.19568225741386414, 0.10822293907403946, -0.005776757374405861, -0.04216858372092247, 0.03711191192269325, -0.14917419850826263, 0.07208448648452759, 0.03165751323103905, 0.07408196479082108, -0.12250201404094696, 0.19318795204162598, 0.0020920548122376204, 0.04687854275107384, -0.026551654562354088, 0.04698725789785385, -0.11904917657375336, 0.037554509937763214, -0.12491800636053085, -0.043439216911792755, -0.018518203869462013, -0.013050179928541183, -0.010149393230676651, -0.08273563534021378, -0.0711965411901474, 0.016350291669368744, -0.13048067688941956, -0.014679946005344391, 0.05896945297718048, 0.01983153074979782, -0.1304175853729248, -0.03653926029801369, 0.030169278383255005, -0.04545648768544197, 0.04653819650411606, 0.05420153588056564, 0.016072651371359825, 0.07125953584909439, -0.17407342791557312, -0.015656549483537674, 0.052768953144550323, 0.002511059632524848, 0.09125425666570663, -0.05547181889414787, -0.009848162531852722, -0.010967493057250977, 0.10595173388719559, 0.025902289897203445, 0.04483994096517563, -0.13859352469444275, -0.004269328899681568, -0.03912092372775078, -0.0801997035741806, -0.06957542151212692, 0.0165261123329401, 0.0855158194899559, 0.032729119062423706, 0.17845310270786285, -0.07427546381950378, 0.05705076456069946, -0.21860480308532715, -0.01556551642715931, -0.013441815041005611, -0.08938828855752945, -0.10344886779785156, -0.05250668153166771, 0.08043178170919418, -0.06394634395837784, 0.11614985018968582, 0.01191363763064146, 0.05895737558603287, 0.032096948474645615, -0.015533443540334702, 0.01935647986829281, 0.016877291724085808, 0.24111658334732056, 0.02868587337434292, -0.028023002669215202, 0.0658063292503357, 0.07043294608592987, 0.07886607199907303, 0.09612764418125153, 0.22141405940055847, 0.17705121636390686, 0.0021003829315304756, 0.0667356625199318, 0.05044236034154892, -0.04720302298665047, -0.12233880162239075, 0.03954247757792473, -0.04063142091035843, 0.06275736540555954, -0.03464983403682709, 0.2146158218383789, 0.07419583946466446, -0.1641249805688858, 0.05126047879457474, -0.05917982757091522, -0.09810306131839752, -0.08776208758354187, -0.003723694011569023, -0.0735606774687767, -0.15074023604393005, 0.03322120010852814, -0.11168154329061508, 0.023990273475646973, 0.1390489935874939, 0.013961726799607277, -0.021417584270238876, 0.19210542738437653, 0.0387880764901638, 0.03232570365071297, 0.05335615947842598, 0.007568936329334974, -0.014047733508050442, -0.08736195415258408, -0.04510989785194397, -0.01907455548644066, -0.03961476311087608, 0.03837640956044197, -0.05543500557541847, -0.088562972843647, 0.02914046309888363, -0.021316781640052795, -0.09675249457359314, 0.0194771196693182, 0.03934618458151817, 0.08972924947738647, 0.05188753083348274, -0.0003099326859228313, 0.026709582656621933, -0.03661458194255829, 0.23112648725509644, -0.0839538499712944, -0.0915762335062027, -0.0834033340215683, 0.24848893284797668, 0.02604314312338829, -0.008974535390734673, 0.029634445905685425, -0.07062368094921112, 0.023064710199832916, 0.2445010393857956, 0.17808489501476288, -0.12368007749319077, -0.003463940927758813, 0.016285618767142296, -0.014092518016695976, -0.04559547081589699, 0.11016539484262466, 0.12846168875694275, 0.06660293787717819, -0.11747605353593826, -0.05492148548364639, -0.07381361722946167, -0.0016268163453787565, -0.029096512123942375, 0.03781260922551155, 0.05202646926045418, 0.0004406342632137239, -0.05534401908516884, 0.07542070746421814, -0.07515842467546463, -0.14741966128349304, 0.10725413262844086, -0.20356632769107819, -0.15794210135936737, -0.011254234239459038, 0.11618871241807938, -0.01106355432420969, 0.0672764778137207, -0.041821759194135666, 0.005528777837753296, 0.03709254041314125, -0.020748969167470932, -0.08018513768911362, -0.08986613154411316, 0.10769883543252945, -0.10077732801437378, 0.17692416906356812, -0.043994612991809845, 0.10050135850906372, 0.12862055003643036, 0.06427479535341263, -0.0637199804186821, 0.06651237607002258, 0.06438721716403961, -0.14426924288272858, 0.013191716745495796, 0.10302330553531647, -0.019900623708963394, 0.049879614263772964, 0.03835655003786087, -0.12629221379756927, 0.02919360063970089, -0.06470083445310593, -0.02802572213113308, -0.07008949667215347, -0.05148179084062576, -0.06315001100301743, 0.1219344213604927, 0.22224409878253937, -0.021837834268808365, 0.03234759345650673, -0.07656009495258331, 0.012176846154034138, 0.06589414924383163, 0.05481071397662163, -0.09182879328727722, -0.2341296523809433, 0.02234431728720665, 0.07208668440580368, -0.04345187172293663, -0.21263402700424194, -0.0904976949095726, 0.033093031495809555, -0.07773891091346741, -0.07553213834762573, 0.075504831969738, 0.08890771865844727, 0.06528367102146149, -0.04188437759876251, -0.11356256902217865, -0.08706662803888321, 0.1629825085401535, -0.16189034283161163, -0.08263864368200302 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.4981 - Matthews Correlation: 0.4218 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5248 | 1.0 | 535 | 0.4981 | 0.4218 | ### Framework versions - Transformers 4.9.1 - Pytorch 1.9.0+cu102 - Datasets 1.10.2 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model_index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"name": "Text Classification", "type": "text-classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metric": {"name": "Matthews Correlation", "type": "matthews_correlation", "value": 0.42176824452830747}}]}]}
text-classification
avneet/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.4981 * Matthews Correlation: 0.4218 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.9.1 * Pytorch 1.9.0+cu102 * Datasets 1.10.2 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.10.2\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.10.2\n* Tokenizers 0.10.3" ]
[ 63, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.10.2\n* Tokenizers 0.10.3" ]
[ -0.10041800886392593, 0.09187408536672592, -0.0017998735420405865, 0.12316817045211792, 0.17497141659259796, 0.030917715281248093, 0.11766374111175537, 0.12837880849838257, -0.09793481975793839, 0.021454423666000366, 0.12565848231315613, 0.17039784789085388, 0.010765883140265942, 0.11671925336122513, -0.05836978927254677, -0.26020652055740356, -0.01399797573685646, 0.047568097710609436, -0.07857032865285873, 0.1329323947429657, 0.09647627174854279, -0.12376619130373001, 0.09279442578554153, -0.005557360593229532, -0.21547169983386993, 0.006550864782184362, 0.012406528927385807, -0.05170246958732605, 0.14593574404716492, 0.034369274973869324, 0.12206467241048813, 0.004261396359652281, 0.08353578299283981, -0.1929604411125183, 0.011002568528056145, 0.04981108754873276, 0.0004636645026039332, 0.088859423995018, 0.04993332922458649, -0.005932452157139778, 0.12560312449932098, -0.10478220134973526, 0.047965601086616516, 0.017689641565084457, -0.12023469805717468, -0.21158799529075623, -0.07731970399618149, 0.02311687171459198, 0.06836913526058197, 0.10707113146781921, -0.008116200566291809, 0.12643027305603027, -0.07986330986022949, 0.09931811690330505, 0.22172631323337555, -0.28908634185791016, -0.06621575355529785, 0.04610060155391693, 0.012032978236675262, 0.05805329233407974, -0.10586149990558624, -0.03007293865084648, 0.05700961500406265, 0.04893903434276581, 0.1250201314687729, -0.039536550641059875, -0.12539702653884888, 0.016339484602212906, -0.13900694251060486, -0.017075415700674057, 0.16145958006381989, 0.0324421226978302, -0.02381230518221855, -0.04417557641863823, -0.057809505611658096, -0.13473296165466309, -0.03139549121260643, -0.001416380051523447, 0.04636354744434357, -0.027420053258538246, -0.07408644258975983, -0.006489279214292765, -0.1112237274646759, -0.05544321611523628, -0.07999242842197418, 0.12144825607538223, 0.029385894536972046, 0.009552819654345512, -0.028467997908592224, 0.09915084391832352, -0.006063870154321194, -0.12679511308670044, 0.021355479955673218, 0.0340057797729969, -0.001206237473525107, -0.054310377687215805, -0.06804459542036057, -0.05557920038700104, 0.01499414723366499, 0.1214403361082077, -0.0552542619407177, 0.046684570610523224, 0.04643268510699272, 0.04566217213869095, -0.08449062705039978, 0.18987902998924255, -0.04141984134912491, -0.01882397010922432, 0.010017924942076206, 0.04024771600961685, 0.019366368651390076, -0.014849066734313965, -0.12844984233379364, 0.005714333150535822, 0.0989694893360138, 0.007235745433717966, -0.06419622153043747, 0.07706768065690994, -0.055387333035469055, -0.019162438809871674, -0.006297029089182615, -0.09685219824314117, 0.034816674888134, -0.0038100359961390495, -0.08036570996046066, -0.021701650694012642, 0.022042816504836082, 0.011858990415930748, -0.018125412985682487, 0.12355764210224152, -0.09039249271154404, 0.041099585592746735, -0.09301173686981201, -0.11494047194719315, 0.02373599447309971, -0.0889897421002388, 0.026660697534680367, -0.09519593417644501, -0.18084761500358582, -0.01998050883412361, 0.0598062165081501, -0.025528520345687866, -0.05700720474123955, -0.058759041130542755, -0.0791688859462738, 0.01318331714719534, -0.01762595772743225, 0.12290996313095093, -0.06654569506645203, 0.0974808931350708, 0.0309133380651474, 0.06348340213298798, -0.06466590613126755, 0.05445391684770584, -0.10252637416124344, 0.009154544211924076, -0.16590702533721924, 0.04758230969309807, -0.04685644432902336, 0.07767967879772186, -0.08151054382324219, -0.1050412654876709, 0.023440955206751823, 0.0038184195291250944, 0.06311875581741333, 0.08573096990585327, -0.19587484002113342, -0.08091865479946136, 0.15240196883678436, -0.06829318404197693, -0.12507720291614532, 0.11799484491348267, -0.06720536947250366, 0.056657660752534866, 0.07249327003955841, 0.1850377768278122, 0.06831374764442444, -0.07342251390218735, -0.0062988209538161755, 0.026235194876790047, 0.05144308879971504, -0.07209300249814987, 0.06183478981256485, 0.011636469513177872, 0.02003420516848564, 0.029572904109954834, -0.023534193634986877, 0.06511012464761734, -0.09338251501321793, -0.0915030762553215, -0.041604459285736084, -0.0891525074839592, 0.037097811698913574, 0.07872863113880157, 0.07367397844791412, -0.10103686153888702, -0.07551629841327667, 0.07852797955274582, 0.07862936705350876, -0.06481653451919556, 0.030340859666466713, -0.055876921862363815, 0.05887269601225853, -0.02183513343334198, -0.0212620310485363, -0.18801823258399963, -0.03682823106646538, 0.0038607583846896887, 0.010319691151380539, 0.02934558503329754, 0.024695467203855515, 0.06556295603513718, 0.052292417734861374, -0.05783342570066452, -0.009018835611641407, -0.009092966094613075, 0.0022893415298312902, -0.13097329437732697, -0.20112940669059753, -0.022544045001268387, -0.02388080582022667, 0.15572431683540344, -0.21747541427612305, 0.04394819587469101, 0.0038682136218994856, 0.06915296614170074, 0.01128359604626894, -0.003836678806692362, -0.044004812836647034, 0.08235437422990799, -0.045891374349594116, -0.05543505772948265, 0.073673777282238, 0.01904948428273201, -0.09871391206979752, -0.05167798697948456, -0.10863251984119415, 0.1557130217552185, 0.1390971541404724, -0.12154051661491394, -0.08272817730903625, -0.004004188813269138, -0.057750213891267776, -0.028204254806041718, -0.04992824047803879, 0.03189495578408241, 0.17613361775875092, -0.009659809060394764, 0.15505492687225342, -0.06711085885763168, -0.04894189536571503, 0.013975011184811592, -0.0406288281083107, 0.02369306981563568, 0.12660491466522217, 0.11989562213420868, -0.07376528531312943, 0.1532750427722931, 0.15223468840122223, -0.1007736399769783, 0.13154472410678864, -0.040487486869096756, -0.06284482777118683, -0.021337702870368958, -0.027980918064713478, -0.005462690256536007, 0.0953056588768959, -0.12414542585611343, -0.0004383292980492115, 0.021249821409583092, 0.02020823210477829, 0.026426924392580986, -0.22902162373065948, -0.03901813179254532, 0.028298847377300262, -0.04131123051047325, -0.010080356150865555, -0.020411545410752296, 0.004382716026157141, 0.10299480706453323, 0.00649286201223731, -0.08438300341367722, 0.045834776014089584, 0.003917197231203318, -0.08299095183610916, 0.2197130173444748, -0.09605570882558823, -0.1668400913476944, -0.1227995827794075, -0.07994429767131805, -0.044267237186431885, 0.00855969823896885, 0.061342332512140274, -0.09074791520833969, -0.032627783715724945, -0.06780200451612473, 0.019676165655255318, -0.002601837506517768, 0.02854214236140251, 0.006594790145754814, 0.006969159003347158, 0.06698118150234222, -0.10637456923723221, -0.010675739496946335, -0.05260933190584183, -0.05867180973291397, 0.05234045162796974, 0.032572995871305466, 0.11060429364442825, 0.15689243376255035, -0.018806632608175278, 0.009577740915119648, -0.036766860634088516, 0.23979203402996063, -0.06298448145389557, -0.020830219611525536, 0.12594269216060638, -0.00815015472471714, 0.048584215342998505, 0.12216746807098389, 0.0719132348895073, -0.08920086920261383, 0.0098697729408741, 0.03435599058866501, -0.03804447129368782, -0.22370994091033936, -0.058035969734191895, -0.056371353566646576, 0.003053137566894293, 0.09371253103017807, 0.03015008382499218, 0.03331061825156212, 0.06415164470672607, 0.048893269151449203, 0.08310278505086899, -0.034768592566251755, 0.055177900940179825, 0.12865015864372253, 0.04185185581445694, 0.1258823424577713, -0.04849233850836754, -0.06215395778417587, 0.04630482569336891, -0.015154440887272358, 0.22378969192504883, -0.000886137830093503, 0.13816937804222107, 0.04938024654984474, 0.15828262269496918, -0.008856663480401039, 0.07402543723583221, -0.004153151996433735, -0.042593494057655334, -0.013074282556772232, -0.03891763091087341, -0.03328578174114227, 0.023511188104748726, -0.06809063255786896, 0.047040414065122604, -0.11741584539413452, 0.0039068711921572685, 0.0666024461388588, 0.2237580567598343, 0.045514851808547974, -0.3226272761821747, -0.09491825103759766, 0.0065155536867678165, -0.026597561314702034, -0.017879795283079147, 0.024990912526845932, 0.10253192484378815, -0.0845203623175621, 0.043942976742982864, -0.06646975874900818, 0.09786922484636307, -0.0541909858584404, 0.053942736238241196, 0.08336091786623001, 0.09575205296278, 0.007795863784849644, 0.0892275795340538, -0.30986952781677246, 0.2702886164188385, 0.007419167086482048, 0.06487283855676651, -0.07511432468891144, 0.000871701049618423, 0.03237006440758705, 0.06733806431293488, 0.06690167635679245, -0.0115983160212636, -0.04549287259578705, -0.18731458485126495, -0.05154705420136452, 0.031141070649027824, 0.08642479032278061, -0.015904927626252174, 0.08752357214689255, -0.03167708218097687, 0.006835493259131908, 0.08087625354528427, -0.026021353900432587, -0.04462859034538269, -0.1103147640824318, -0.010455415584146976, 0.03327735513448715, -0.053259897977113724, -0.05801377072930336, -0.11076731979846954, -0.13096573948860168, 0.16943955421447754, -0.029300201684236526, -0.031914882361888885, -0.10776127129793167, 0.09976936876773834, 0.06709963083267212, -0.08976428210735321, 0.03410027176141739, 0.010387159883975983, 0.07125640660524368, 0.03187572956085205, -0.07662557810544968, 0.11000455915927887, -0.0637795552611351, -0.1492954045534134, -0.058762624859809875, 0.09942290186882019, 0.0303806122392416, 0.0672077164053917, -0.01177572924643755, 0.0148955462500453, -0.05002004653215408, -0.0920470803976059, 0.010348375886678696, -0.01362479105591774, 0.06849057972431183, 0.03202947601675987, -0.06297567486763, 0.009045197628438473, -0.05817518010735512, -0.03525881469249725, 0.2042563557624817, 0.2326991707086563, -0.10180642455816269, 0.011091076768934727, 0.02340918965637684, -0.07227852195501328, -0.21036966145038605, 0.04755167290568352, 0.0514562763273716, 0.002969205379486084, 0.03148770332336426, -0.17819714546203613, 0.14336206018924713, 0.11067391186952591, -0.008122633211314678, 0.10153565555810928, -0.3171425759792328, -0.12998312711715698, 0.11766942590475082, 0.1308688521385193, 0.1157570332288742, -0.1425764560699463, -0.018778659403324127, -0.03378121554851532, -0.11945401877164841, 0.13359268009662628, -0.10943881422281265, 0.12029775232076645, -0.023171018809080124, 0.08387299627065659, 0.0005684139905497432, -0.056921232491731644, 0.12249056249856949, 0.039181485772132874, 0.09957252442836761, -0.061339154839515686, -0.037827055901288986, 0.04639703035354614, -0.03778168931603432, 0.020333176478743553, -0.09518802911043167, 0.02552787773311138, -0.09452233463525772, -0.020768078044056892, -0.06795525550842285, 0.04135217145085335, -0.036730047315359116, -0.0626441016793251, -0.04560517892241478, 0.02633962593972683, 0.04718917980790138, -0.012256861664354801, 0.13631828129291534, 0.020231541246175766, 0.14470645785331726, 0.10413989424705505, 0.07258670032024384, -0.08221485465765, -0.05656382068991661, -0.008773869834840298, -0.014354662969708443, 0.04806140065193176, -0.14905856549739838, 0.029915401712059975, 0.15433886647224426, 0.019725650548934937, 0.1488981395959854, 0.08756951987743378, -0.01106821559369564, 0.0023593390360474586, 0.06695582717657089, -0.16568143665790558, -0.09227000921964645, -0.004010491073131561, -0.07201182842254639, -0.11694513261318207, 0.04586157575249672, 0.09646008908748627, -0.06068021059036255, -0.004669808316975832, -0.0027742807287722826, 0.007091234438121319, -0.059203095734119415, 0.1930038034915924, 0.06765357404947281, 0.04254591837525368, -0.10652004182338715, 0.06990062445402145, 0.05173873528838158, -0.08142077922821045, 0.004759574308991432, 0.08656732738018036, -0.08492020517587662, -0.05077390745282173, 0.08807110786437988, 0.18612973392009735, -0.06662914156913757, -0.048556383699178696, -0.1379210650920868, -0.12274498492479324, 0.07784494757652283, 0.15883810818195343, 0.12285467982292175, 0.019979601725935936, -0.05903337523341179, 0.00364125007763505, -0.12390539050102234, 0.08823470026254654, 0.047047208994627, 0.06731384247541428, -0.14347706735134125, 0.16567479074001312, 0.00775969447568059, 0.05158763378858566, -0.024689868092536926, 0.029208583757281303, -0.10862531512975693, 0.009984218515455723, -0.10589565336704254, -0.00962116476148367, -0.037082694470882416, 0.009586487896740437, 0.006466344930231571, -0.05006425827741623, -0.05537882074713707, 0.0052664438262581825, -0.11266795545816422, -0.02297920547425747, 0.02787833660840988, 0.05687223747372627, -0.11541324108839035, -0.04553835093975067, 0.02058243937790394, -0.05866514518857002, 0.07484627515077591, 0.04245281592011452, 0.011535332538187504, 0.06486064940690994, -0.15083833038806915, 0.0015109420055523515, 0.07581590861082077, 0.023464716970920563, 0.07091418653726578, -0.07957907766103745, -0.008847996592521667, 0.00432796124368906, 0.060914792120456696, 0.025270136073231697, 0.07909148931503296, -0.1333031952381134, 0.008405223488807678, -0.03550156578421593, -0.08304344862699509, -0.06850036978721619, 0.037134088575839996, 0.08698856085538864, 0.021525397896766663, 0.20387326180934906, -0.08401969820261002, 0.03701129928231239, -0.20669837296009064, 0.006539537105709314, -0.014193656854331493, -0.11950641125440598, -0.11368824541568756, -0.07019766420125961, 0.0651252493262291, -0.05561024695634842, 0.1474165916442871, 0.03421015664935112, 0.03780185058712959, 0.02990436553955078, -0.007347329054027796, 0.0028849770314991474, 0.021476637572050095, 0.21079862117767334, 0.03175308182835579, -0.042044512927532196, 0.055925190448760986, 0.05150838568806648, 0.11006495356559753, 0.11284393817186356, 0.20646686851978302, 0.1377871334552765, -0.013296850025653839, 0.0944439247250557, 0.03292014077305794, -0.06521227955818176, -0.14624156057834625, 0.058382563292980194, -0.048982709646224976, 0.10409432649612427, -0.02545270510017872, 0.20936626195907593, 0.05051792412996292, -0.16985879838466644, 0.042929355055093765, -0.06493119150400162, -0.09706878662109375, -0.11181750148534775, -0.032795339822769165, -0.08750470727682114, -0.1364011913537979, -0.008014756254851818, -0.11971766501665115, 0.007264505140483379, 0.11717516928911209, 0.0075271413661539555, -0.02369200624525547, 0.15159909427165985, 0.007651979569345713, 0.02268468216061592, 0.05553872510790825, 0.004416924435645342, -0.03122895583510399, -0.11445702612400055, -0.05970478057861328, -0.022480789572000504, -0.010734658688306808, 0.0334610715508461, -0.05267271026968956, -0.03963937982916832, 0.03730347752571106, -0.0377950482070446, -0.09514646232128143, 0.01117195375263691, 0.030395641922950745, 0.05967599153518677, 0.04120979830622673, 0.009893695823848248, 0.007579626981168985, -0.004266524221748114, 0.21891868114471436, -0.08027279376983643, -0.08825135976076126, -0.1003011018037796, 0.25660526752471924, 0.03769070282578468, -0.011248022317886353, 0.028761446475982666, -0.07068607211112976, -0.0005062112468294799, 0.24264316260814667, 0.21856454014778137, -0.09904588758945465, -0.00422821706160903, 0.004117398988455534, -0.0045724110677838326, -0.020304996520280838, 0.10836180299520493, 0.13619457185268402, 0.05538675934076309, -0.09631659090518951, -0.05285907909274101, -0.05166875571012497, -0.012768933549523354, -0.04755297303199768, 0.051787205040454865, 0.043311141431331635, 0.007391277700662613, -0.03679182380437851, 0.05968480184674263, -0.07410793751478195, -0.09347275644540787, 0.057483021169900894, -0.22103223204612732, -0.16184628009796143, -0.011745045892894268, 0.1102326363325119, 0.0009563847561366856, 0.06431415677070618, -0.028878148645162582, -0.00012531227548606694, 0.06997157633304596, -0.021442001685500145, -0.10136649012565613, -0.059716202318668365, 0.08901814371347427, -0.12041039019823074, 0.19281058013439178, -0.049165330827236176, 0.06927259266376495, 0.1326725333929062, 0.06714607775211334, -0.07058152556419373, 0.06696251779794693, 0.0382753424346447, -0.052780307829380035, 0.03230620548129082, 0.0738477110862732, -0.04346172511577606, 0.05991750210523605, 0.050269272178411484, -0.13513000309467316, 0.020845767110586166, -0.05358114838600159, -0.05995842441916466, -0.04704571142792702, -0.04494743421673775, -0.058727581053972244, 0.12542568147182465, 0.21562181413173676, -0.02751288004219532, 0.006098936777561903, -0.06687735766172409, -0.0004073968739248812, 0.05497436597943306, 0.03065263107419014, -0.06872787326574326, -0.2185104340314865, 0.01211613230407238, 0.05043087527155876, -0.013951906003057957, -0.2362850457429886, -0.08787836879491806, -0.003890622640028596, -0.06874100863933563, -0.09629415720701218, 0.09182994812726974, 0.0827857106924057, 0.04769261181354523, -0.05954857915639877, -0.05265124887228012, -0.07743755728006363, 0.15188424289226532, -0.14308449625968933, -0.09820015728473663 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-sst2 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.3651 - Accuracy: 0.9151 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.1902 | 1.0 | 4210 | 0.3102 | 0.9117 | | 0.1293 | 2.0 | 8420 | 0.3672 | 0.9048 | | 0.084 | 3.0 | 12630 | 0.3651 | 0.9151 | | 0.0682 | 4.0 | 16840 | 0.3971 | 0.9037 | | 0.0438 | 5.0 | 21050 | 0.4720 | 0.9117 | ### Framework versions - Transformers 4.9.1 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["accuracy"], "model_index": [{"name": "distilbert-base-uncased-finetuned-sst2", "results": [{"task": {"name": "Text Classification", "type": "text-classification"}, "dataset": {"name": "glue", "type": "glue", "args": "sst2"}, "metric": {"name": "Accuracy", "type": "accuracy", "value": 0.9151376146788991}}]}]}
text-classification
avneet/distilbert-base-uncased-finetuned-sst2
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-sst2 ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.3651 * Accuracy: 0.9151 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.9.1 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ 63, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.9.1\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ -0.10138363391160965, 0.09181274473667145, -0.0017480016686022282, 0.12251314520835876, 0.17487236857414246, 0.030555732548236847, 0.11795240640640259, 0.1284196972846985, -0.0991346538066864, 0.021271787583827972, 0.12584805488586426, 0.17002835869789124, 0.010741710662841797, 0.11769048124551773, -0.05881334841251373, -0.2584119737148285, -0.014017499051988125, 0.04791738837957382, -0.07845775783061981, 0.13301406800746918, 0.09648574143648148, -0.12414977699518204, 0.09311668574810028, -0.004832270089536905, -0.2168150246143341, 0.0075859264470636845, 0.012418114580214024, -0.052767205983400345, 0.1444525122642517, 0.034126028418540955, 0.12158221751451492, 0.005013753660023212, 0.08290760964155197, -0.1915157288312912, 0.011272239498794079, 0.05034958943724632, 0.0004795369750354439, 0.08983013778924942, 0.04984158277511597, -0.004389860667288303, 0.12477509677410126, -0.10553069412708282, 0.04841882735490799, 0.017974091693758965, -0.11980147659778595, -0.21046559512615204, -0.07755997031927109, 0.023209288716316223, 0.06895510852336884, 0.10723888874053955, -0.008464409969747066, 0.1267317682504654, -0.07952189445495605, 0.09968549013137817, 0.22137734293937683, -0.2899080514907837, -0.06648627668619156, 0.04634452611207962, 0.012798751704394817, 0.058051660656929016, -0.1062227264046669, -0.029340026900172234, 0.05857633799314499, 0.04805753380060196, 0.12578821182250977, -0.039213500916957855, -0.12572668492794037, 0.015777021646499634, -0.1394176036119461, -0.01676696352660656, 0.16315938532352448, 0.03296953812241554, -0.023761456832289696, -0.04513244703412056, -0.05685555189847946, -0.1354190558195114, -0.031068751588463783, -0.002263233996927738, 0.046718157827854156, -0.028431862592697144, -0.07347904145717621, -0.0067316628992557526, -0.11192303895950317, -0.05554114282131195, -0.07968578487634659, 0.11954854428768158, 0.028946293517947197, 0.009572487324476242, -0.030037324875593185, 0.09903988242149353, -0.00625665532425046, -0.1259600818157196, 0.020809059962630272, 0.03279375284910202, -0.0003051985113415867, -0.05526058375835419, -0.0685529038310051, -0.057281743735075, 0.014994512312114239, 0.1212097778916359, -0.05445388704538345, 0.046691011637449265, 0.046145811676979065, 0.045070964843034744, -0.08390553295612335, 0.18968869745731354, -0.04098469763994217, -0.019244709983468056, 0.011217089369893074, 0.041044097393751144, 0.018894728273153305, -0.013744070194661617, -0.12753142416477203, 0.004806716460734606, 0.09884382039308548, 0.007470232900232077, -0.06463329493999481, 0.07715759426355362, -0.05452563613653183, -0.019385728985071182, -0.0054857442155480385, -0.09680405259132385, 0.03441823273897171, -0.004249462857842445, -0.08126911520957947, -0.022060025483369827, 0.022231625393033028, 0.011360249482095242, -0.018155450001358986, 0.12288933247327805, -0.0906168669462204, 0.040340036153793335, -0.09327813982963562, -0.11538615822792053, 0.02385558933019638, -0.09099278599023819, 0.026915371417999268, -0.09574364870786667, -0.18126414716243744, -0.02119799703359604, 0.05838366597890854, -0.02496919222176075, -0.057531844824552536, -0.05914877727627754, -0.07909619808197021, 0.012710535898804665, -0.01752087101340294, 0.12487196922302246, -0.0661722719669342, 0.09737953543663025, 0.03186080604791641, 0.06287983059883118, -0.06483995914459229, 0.05454431101679802, -0.10234236717224121, 0.007735768798738718, -0.1652398407459259, 0.04769985377788544, -0.046658217906951904, 0.07866405695676804, -0.08223975449800491, -0.1051115170121193, 0.02311071567237377, 0.0033810080494731665, 0.06313920766115189, 0.08614415675401688, -0.19414222240447998, -0.08049311488866806, 0.15221811830997467, -0.06840579956769943, -0.12562373280525208, 0.11699496954679489, -0.06706288456916809, 0.05698290467262268, 0.07371602952480316, 0.184627965092659, 0.06854979693889618, -0.07383659482002258, -0.006559441797435284, 0.025592148303985596, 0.05278046429157257, -0.07360737025737762, 0.06177666038274765, 0.012096193619072437, 0.018694767728447914, 0.030097441747784615, -0.02470897138118744, 0.06522233784198761, -0.0925702303647995, -0.09115505963563919, -0.042404405772686005, -0.08998403698205948, 0.03639598563313484, 0.078729048371315, 0.07467003166675568, -0.10150585323572159, -0.07369990646839142, 0.07819622755050659, 0.07913773506879807, -0.0637807846069336, 0.02981165610253811, -0.055730391293764114, 0.05935884267091751, -0.02315233089029789, -0.020711883902549744, -0.1876806765794754, -0.0353415384888649, 0.0033801691606640816, 0.010064490139484406, 0.02905278652906418, 0.02397819235920906, 0.06553194671869278, 0.05169892683625221, -0.05748188868165016, -0.009254573844373226, -0.00924514140933752, 0.0017017362406477332, -0.13137264549732208, -0.201021209359169, -0.02204478532075882, -0.024205349385738373, 0.15560223162174225, -0.21743567287921906, 0.04454241320490837, 0.002680771751329303, 0.06929411739110947, 0.011735393665730953, -0.003736986080184579, -0.042548440396785736, 0.08219271153211594, -0.045021288096904755, -0.05508570000529289, 0.0734293982386589, 0.01929287798702717, -0.09789107739925385, -0.052520621567964554, -0.1082654818892479, 0.15701158344745636, 0.13954690098762512, -0.12095533311367035, -0.08241265267133713, -0.004869589116424322, -0.057637132704257965, -0.027924181893467903, -0.04948529973626137, 0.03248227760195732, 0.17407584190368652, -0.01073733251541853, 0.15568917989730835, -0.06790310144424438, -0.049282900989055634, 0.014606268145143986, -0.039691586047410965, 0.023940667510032654, 0.12615559995174408, 0.1206049844622612, -0.0747140496969223, 0.15334181487560272, 0.1534608006477356, -0.10042717307806015, 0.13248319923877716, -0.040576107800006866, -0.0631931722164154, -0.020310979336500168, -0.02775336056947708, -0.006297021172940731, 0.09525734186172485, -0.12458211183547974, -0.0019712201319634914, 0.021205568686127663, 0.019204357638955116, 0.026035606861114502, -0.22809220850467682, -0.03826599195599556, 0.0281943641602993, -0.04012012109160423, -0.01114169042557478, -0.019353274255990982, 0.003671730402857065, 0.10307865589857101, 0.006109531037509441, -0.08333727717399597, 0.04527758061885834, 0.00433516688644886, -0.08271829783916473, 0.2196817696094513, -0.09577154368162155, -0.16733883321285248, -0.12176884710788727, -0.07893921434879303, -0.044794365763664246, 0.007620422635227442, 0.061704300343990326, -0.09212178736925125, -0.03240298107266426, -0.06761292368173599, 0.019370902329683304, -0.0023205296602100134, 0.028749944642186165, 0.00725533114746213, 0.007390500977635384, 0.06709158420562744, -0.10672678798437119, -0.009718425571918488, -0.053294528275728226, -0.059197794646024704, 0.05104880407452583, 0.032916001975536346, 0.11025732010602951, 0.15554381906986237, -0.018638934940099716, 0.009264804422855377, -0.0379490852355957, 0.24054202437400818, -0.06337197124958038, -0.02255132421851158, 0.12653496861457825, -0.007734472397714853, 0.046888548880815506, 0.12052290141582489, 0.07125964015722275, -0.08900165557861328, 0.009850301779806614, 0.03428591787815094, -0.036994948983192444, -0.22528907656669617, -0.05716115981340408, -0.056467991322278976, 0.0030881159473210573, 0.0939449816942215, 0.030201366171240807, 0.034760039299726486, 0.0645616203546524, 0.04835493117570877, 0.08317936211824417, -0.03455687686800957, 0.055070631206035614, 0.12929177284240723, 0.04124240577220917, 0.12607607245445251, -0.04883788898587227, -0.0621289499104023, 0.045546237379312515, -0.015357374213635921, 0.22370217740535736, -0.0004287105402909219, 0.13657733798027039, 0.04890884831547737, 0.15842336416244507, -0.008791405707597733, 0.07418705523014069, -0.0037982272915542126, -0.042784661054611206, -0.012841083109378815, -0.038446925580501556, -0.032840095460414886, 0.023165127262473106, -0.06864646822214127, 0.04783427715301514, -0.1182292252779007, 0.003483798587694764, 0.06672365963459015, 0.22510957717895508, 0.045347969979047775, -0.3229960799217224, -0.09459678828716278, 0.005747778806835413, -0.026898711919784546, -0.018383393064141273, 0.024982361122965813, 0.10111680626869202, -0.08431806415319443, 0.044699277728796005, -0.06669696420431137, 0.0978696271777153, -0.05325652286410332, 0.053714096546173096, 0.0833359807729721, 0.09653782844543457, 0.0074645220302045345, 0.0882764682173729, -0.31208133697509766, 0.26951196789741516, 0.007898678071796894, 0.06543229520320892, -0.07460381835699081, 0.000865441164933145, 0.03253268077969551, 0.06775101274251938, 0.06726422160863876, -0.012305173091590405, -0.04567352309823036, -0.1856955885887146, -0.0524708554148674, 0.03176290914416313, 0.08576912432909012, -0.016058573499321938, 0.08660877496004105, -0.031276553869247437, 0.006710614077746868, 0.08061738312244415, -0.02759450301527977, -0.047168929129838943, -0.11016498506069183, -0.010352971963584423, 0.0342746302485466, -0.05324574559926987, -0.057463500648736954, -0.11063649505376816, -0.13110534846782684, 0.16762825846672058, -0.030626121908426285, -0.031653039157390594, -0.10816682875156403, 0.10112731903791428, 0.0664503276348114, -0.08856356143951416, 0.03485357016324997, 0.010295683518052101, 0.0708969384431839, 0.03165864944458008, -0.07747503370046616, 0.109786257147789, -0.06427845358848572, -0.14875121414661407, -0.05849020555615425, 0.09909399598836899, 0.02966337651014328, 0.06653016805648804, -0.011604848317801952, 0.014306565746665001, -0.04986817389726639, -0.09121626615524292, 0.0112402169033885, -0.012907742522656918, 0.06797314435243607, 0.03146473690867424, -0.0636986792087555, 0.006366269197314978, -0.05857650190591812, -0.03542293608188629, 0.20417991280555725, 0.23033764958381653, -0.10128122568130493, 0.01128136646002531, 0.023579541593790054, -0.07304376363754272, -0.20903372764587402, 0.04996675252914429, 0.052024539560079575, 0.0038596957456320524, 0.03059072233736515, -0.17878276109695435, 0.1426609456539154, 0.11043008416891098, -0.008285046555101871, 0.10300042480230331, -0.3162187933921814, -0.1298130452632904, 0.11809475719928741, 0.13054122030735016, 0.1176321730017662, -0.1425286829471588, -0.0183709729462862, -0.03309715539216995, -0.11702732741832733, 0.1329037845134735, -0.10983089357614517, 0.11971795558929443, -0.023416943848133087, 0.08235824108123779, 0.0005208451766520739, -0.0567815899848938, 0.12149792909622192, 0.03761216253042221, 0.10023183375597, -0.060792677104473114, -0.036266520619392395, 0.04596136510372162, -0.03819906711578369, 0.02092742919921875, -0.09577959030866623, 0.025361867621541023, -0.0936150848865509, -0.020406529307365417, -0.06852612644433975, 0.042476836591959, -0.03654692322015762, -0.061134614050388336, -0.04469207674264908, 0.026457475498318672, 0.04644780233502388, -0.012331503443419933, 0.13658395409584045, 0.02035389095544815, 0.14555925130844116, 0.10335679352283478, 0.07351463288068771, -0.08133154362440109, -0.056079600006341934, -0.008207510225474834, -0.014590260572731495, 0.04897656664252281, -0.14960558712482452, 0.028569240123033524, 0.1536419540643692, 0.019732961431145668, 0.1482411026954651, 0.0869421511888504, -0.010772149078547955, 0.002147610764950514, 0.06737595051527023, -0.1666678488254547, -0.0909922644495964, -0.004821413196623325, -0.07190437614917755, -0.11649473756551743, 0.04683881253004074, 0.09706324338912964, -0.061846498399972916, -0.005686305463314056, -0.0030022866558283567, 0.007164324168115854, -0.05895275995135307, 0.19428816437721252, 0.06646456569433212, 0.042164549231529236, -0.10693242400884628, 0.07034572213888168, 0.051408663392066956, -0.08282521367073059, 0.00593935139477253, 0.08518387377262115, -0.08517700433731079, -0.051019828766584396, 0.08751039206981659, 0.18584555387496948, -0.06616263836622238, -0.04911690950393677, -0.13821348547935486, -0.12204624712467194, 0.07799609750509262, 0.15999682247638702, 0.12266577035188675, 0.019017066806554794, -0.05772237107157707, 0.004102302715182304, -0.12431347370147705, 0.08867493271827698, 0.0465819425880909, 0.06777802109718323, -0.14385131001472473, 0.1651107370853424, 0.008049125783145428, 0.05041046813130379, -0.02450399287045002, 0.029895659536123276, -0.10823038220405579, 0.009272205643355846, -0.10413534939289093, -0.010283788666129112, -0.036030080169439316, 0.009663371369242668, 0.006098243873566389, -0.04915362223982811, -0.05547851696610451, 0.005855679977685213, -0.11320266872644424, -0.022768666967749596, 0.02856818586587906, 0.05749751627445221, -0.11522112786769867, -0.04550881311297417, 0.019614633172750473, -0.058190491050481796, 0.07416874170303345, 0.042481135576963425, 0.011530384421348572, 0.0648583173751831, -0.15035967528820038, 0.0011729872785508633, 0.07639433443546295, 0.02306401915848255, 0.0702202171087265, -0.07973466068506241, -0.007887307554483414, 0.005099095869809389, 0.060839470475912094, 0.024969816207885742, 0.07799719274044037, -0.1327405720949173, 0.00647034402936697, -0.035032033920288086, -0.08092737942934036, -0.06851889938116074, 0.03686651960015297, 0.08677248656749725, 0.021485108882188797, 0.20365601778030396, -0.08378603309392929, 0.035802118480205536, -0.20686383545398712, 0.006670020055025816, -0.014568416401743889, -0.12046866118907928, -0.11601729691028595, -0.06955434381961823, 0.06604958325624466, -0.05596913769841194, 0.1477183699607849, 0.03519124165177345, 0.037841230630874634, 0.030449194833636284, -0.006695518270134926, 0.002068361733108759, 0.021651513874530792, 0.21045580506324768, 0.03141474723815918, -0.04134353622794151, 0.05728282406926155, 0.052612900733947754, 0.1101500615477562, 0.11330829560756683, 0.20648503303527832, 0.1380019187927246, -0.013899591751396656, 0.095646433532238, 0.033273931592702866, -0.0642935112118721, -0.1443653106689453, 0.059243205934762955, -0.048629723489284515, 0.10570532828569412, -0.02623470313847065, 0.20827357470989227, 0.051297031342983246, -0.16901373863220215, 0.04210517555475235, -0.06468261033296585, -0.09619585424661636, -0.11231566220521927, -0.03238716721534729, -0.0875634029507637, -0.13717755675315857, -0.00764749851077795, -0.12037774175405502, 0.00704254349693656, 0.11477746814489365, 0.007749590557068586, -0.02439321205019951, 0.15134622156620026, 0.00909509602934122, 0.022123757749795914, 0.05812042951583862, 0.003955349326133728, -0.03175492212176323, -0.11609432846307755, -0.05957282334566116, -0.021430084481835365, -0.011019710451364517, 0.03345385938882828, -0.05343703553080559, -0.0408368855714798, 0.038032352924346924, -0.037218619138002396, -0.09469655156135559, 0.010968920774757862, 0.03042767010629177, 0.06000256910920143, 0.04168739542365074, 0.009817145764827728, 0.007382037118077278, -0.004258687607944012, 0.21944472193717957, -0.08034026622772217, -0.08558638393878937, -0.10149785131216049, 0.25397300720214844, 0.03819878771901131, -0.012124216184020042, 0.028620058670639992, -0.07035787403583527, 0.00025873054983094335, 0.24288126826286316, 0.21838118135929108, -0.09676580131053925, -0.004494017921388149, 0.0034187587443739176, -0.004584383685141802, -0.020861610770225525, 0.10863317549228668, 0.13634389638900757, 0.053747471421957016, -0.09628306329250336, -0.05230684578418732, -0.051461752504110336, -0.012838656082749367, -0.045591581612825394, 0.05112669616937637, 0.04321906715631485, 0.006692682392895222, -0.035356249660253525, 0.06000112369656563, -0.07312456518411636, -0.09315123409032822, 0.0566585436463356, -0.22041015326976776, -0.1617775559425354, -0.012369958683848381, 0.10987871140241623, 0.0014037297805771232, 0.06387913227081299, -0.02875623106956482, 0.0002558798005338758, 0.07067462056875229, -0.02238454483449459, -0.10012708604335785, -0.06309697031974792, 0.08920733630657196, -0.12066013365983963, 0.19246868789196014, -0.048799995332956314, 0.06887868046760559, 0.13211356103420258, 0.06647348403930664, -0.07096020132303238, 0.06567104905843735, 0.037423670291900635, -0.052818406373262405, 0.032912615686655045, 0.07441634684801102, -0.04309992119669914, 0.061217714101076126, 0.04962910711765289, -0.1339409351348877, 0.020919783040881157, -0.054605819284915924, -0.05957318842411041, -0.047798026353120804, -0.042255524545907974, -0.05862826108932495, 0.12552937865257263, 0.21632321178913116, -0.027486182749271393, 0.005953099578619003, -0.06732727587223053, 0.00007510215073125437, 0.05533996596932411, 0.030002247542142868, -0.06856279820203781, -0.21824997663497925, 0.012374167330563068, 0.05057553946971893, -0.01405921671539545, -0.23624102771282196, -0.08679673075675964, -0.004863446578383446, -0.06928899884223938, -0.09631688892841339, 0.09118016064167023, 0.08410878479480743, 0.047422777861356735, -0.05942878499627113, -0.05266397446393967, -0.07673587650060654, 0.15150220692157745, -0.14311285316944122, -0.09775520861148834 ]
null
null
transformers
---- tags: - conversational --- #Rick DialoGPT model
{}
text-generation
avnish100/DialoGPT-small-rick
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
---- tags: - conversational --- #Rick DialoGPT model
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 47 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.027653997763991356, 0.02414041943848133, -0.0068230400793254375, 0.010564634576439857, 0.18164798617362976, 0.033704131841659546, 0.08821956068277359, 0.13570955395698547, -0.0068973456509411335, -0.013526750728487968, 0.1547490805387497, 0.20799952745437622, -0.0026462990790605545, 0.0791444480419159, -0.0664469450712204, -0.2753458023071289, 0.05913490429520607, 0.0680282786488533, -0.007687992881983519, 0.12075648456811905, 0.07187031954526901, -0.0549883171916008, 0.0886516347527504, -0.02030559629201889, -0.17324471473693848, 0.01953965798020363, 0.04816993698477745, -0.12518654763698578, 0.1176358312368393, 0.05111858248710632, 0.09795232862234116, 0.008365745656192303, -0.06405694782733917, -0.13635118305683136, 0.022147029638290405, 0.03033585101366043, -0.058860234916210175, 0.0636059120297432, 0.1087222546339035, -0.09939044713973999, 0.09311723709106445, 0.08541663736104965, -0.0255570225417614, 0.05364618077874184, -0.15825888514518738, -0.06378549337387085, -0.02499648556113243, 0.007804732769727707, 0.06256697326898575, 0.10073644667863846, -0.017566369846463203, 0.10258800536394119, -0.0975269079208374, 0.10333853214979172, 0.1500675231218338, -0.3112771809101105, 0.009987793862819672, 0.09499151259660721, 0.04119991883635521, 0.03931105509400368, -0.02533094584941864, 0.05045793950557709, 0.025268254801630974, 0.027277586981654167, 0.007437177933752537, -0.0750175341963768, -0.1137726753950119, 0.049895867705345154, -0.09199702739715576, -0.07458660751581192, 0.22324641048908234, -0.07399588078260422, 0.060080595314502716, -0.025852523744106293, -0.11121725291013718, -0.05274823680520058, -0.013890148140490055, 0.018784796819090843, -0.06587869673967361, 0.08765926212072372, 0.024050135165452957, -0.06755640357732773, -0.1323474794626236, -0.04128742218017578, -0.18628640472888947, 0.17943057417869568, 0.015332846902310848, 0.05883103236556053, -0.1924149990081787, 0.11635245382785797, -0.004000017885118723, -0.08559784293174744, 0.024640021845698357, -0.09488005936145782, 0.03717249631881714, -0.005796557758003473, -0.06343648582696915, -0.07624655961990356, 0.078512042760849, 0.13449318706989288, -0.0038929670117795467, 0.031459223479032516, -0.03913462534546852, 0.08946967869997025, 0.023094916716217995, 0.11019261926412582, -0.01329297386109829, -0.00601809611544013, 0.043852973729372025, -0.14449132978916168, -0.008341594599187374, -0.06913956254720688, -0.1527271568775177, -0.05108632892370224, 0.05306483805179596, 0.08953460305929184, 0.008545879274606705, 0.09067165106534958, -0.04840036481618881, -0.026439275592565536, 0.06191498041152954, -0.07166212797164917, -0.0057375445030629635, 0.0005479406099766493, 0.020326290279626846, 0.12346802651882172, -0.006863993126899004, 0.01816580630838871, -0.1344953328371048, 0.07597071677446365, -0.0810447409749031, 0.0016609809827059507, -0.037295255810022354, -0.051307324320077896, 0.016753138974308968, -0.09774310886859894, 0.014272624626755714, -0.15190516412258148, -0.18175770342350006, 0.015764877200126648, 0.0044948384165763855, -0.03198384866118431, -0.035312067717313766, -0.03263629972934723, -0.023609675467014313, 0.04306609928607941, -0.06790579855442047, 0.009302832186222076, -0.05678845942020416, 0.10395034402608871, -0.032171644270420074, 0.06649759411811829, -0.10738259553909302, 0.0829162523150444, -0.12368609756231308, -0.004673504736274481, -0.09571383893489838, 0.07571588456630707, -0.0049130916595458984, 0.11728651076555252, -0.028541911393404007, -0.03454771637916565, -0.07556727528572083, 0.04999465495347977, -0.02550712786614895, 0.18951213359832764, -0.060080599039793015, -0.12557648122310638, 0.2583121061325073, -0.07503679394721985, -0.1294521689414978, 0.09354755282402039, 0.013357079587876797, 0.03000263124704361, 0.08708256483078003, 0.17770351469516754, 0.03385210409760475, 0.011724604293704033, 0.08526027947664261, 0.1101398766040802, -0.11245359480381012, -0.0934135690331459, 0.01582467369735241, -0.04410967230796814, -0.14348545670509338, 0.0551721565425396, 0.06396481394767761, 0.08126390725374222, -0.04889657348394394, -0.02648499235510826, -0.04211905598640442, 0.005280596204102039, 0.08378548920154572, 0.011136471293866634, 0.12981148064136505, -0.04937934875488281, -0.03142275661230087, -0.018193937838077545, -0.012411710806190968, -0.03191297501325607, 0.03591127321124077, -0.019667068496346474, 0.13700194656848907, -0.048340748995542526, 0.053371917456388474, -0.18971459567546844, -0.07922437787055969, 0.0010099048959091306, 0.123023621737957, -0.014106693677604198, 0.08013445883989334, 0.05753817409276962, -0.018720267340540886, -0.004700321704149246, -0.01032867468893528, 0.1544346958398819, -0.021616755053400993, -0.06661882251501083, -0.04162381589412689, 0.0662311464548111, -0.05831345543265343, -0.0033040468115359545, -0.05776660889387131, 0.013589667156338692, 0.05048443749547005, 0.10443682968616486, -0.0023575187660753727, 0.03253777325153351, -0.02123248018324375, 0.018250472843647003, -0.07885172218084335, -0.0028943256475031376, 0.09839999675750732, -0.003195167751982808, -0.06114937365055084, 0.191707044839859, -0.16508106887340546, 0.2123199850320816, 0.18989497423171997, -0.2840019166469574, 0.008855658583343029, -0.07930868119001389, -0.03107025846838951, 0.019292673096060753, 0.04051336646080017, -0.035391807556152344, 0.12321244925260544, 0.0030509934294968843, 0.1893225461244583, -0.05120055004954338, -0.054668959230184555, -0.0003608512051869184, -0.05736381933093071, 0.0013126746052876115, 0.06707432866096497, 0.11558198183774948, -0.12564630806446075, 0.1973772495985031, 0.17830142378807068, 0.02446782775223255, 0.16028088331222534, 0.003589105326682329, -0.02908729389309883, 0.07800903916358948, 0.001039333757944405, -0.03403163328766823, -0.08341804146766663, -0.19453173875808716, -0.01920945756137371, 0.08615871518850327, 0.05208343267440796, 0.11178864538669586, -0.1340440809726715, -0.039688125252723694, -0.016580121591687202, -0.013963420875370502, 0.004052120726555586, 0.08927994221448898, 0.05621529743075371, 0.11766386777162552, -0.008479462936520576, 0.004914911463856697, 0.11690844595432281, 0.024292193353176117, -0.0974007099866867, 0.20369629561901093, -0.12859489023685455, -0.35919657349586487, -0.17192909121513367, -0.16941924393177032, -0.046767693012952805, 0.06603047996759415, 0.10566895455121994, -0.11921820044517517, -0.03283723443746567, 0.01984371617436409, 0.10511579364538193, -0.0874844342470169, 0.025252653285861015, -0.07854585349559784, 0.039858005940914154, -0.08228866755962372, -0.07852846384048462, -0.058627899736166, -0.02397638000547886, -0.06844961643218994, 0.15293799340724945, -0.10580270737409592, 0.04606963321566582, 0.19703397154808044, 0.035209350287914276, 0.05708123743534088, -0.03352535888552666, 0.19375872611999512, -0.09711813181638718, -0.014181635342538357, 0.20692157745361328, -0.04432303458452225, 0.08276087045669556, 0.10658510029315948, -0.0009211950236931443, -0.0905555859208107, 0.023672347888350487, -0.03327333554625511, -0.09995128959417343, -0.2413795441389084, -0.12423769384622574, -0.12672755122184753, 0.07157120853662491, 0.06113129481673241, 0.06719478219747543, 0.1604551076889038, 0.09354656934738159, -0.019843624904751778, 0.04505275562405586, -0.0036725422833114862, 0.07906411588191986, 0.20365294814109802, -0.0204415675252676, 0.13615357875823975, -0.050657231360673904, -0.13334059715270996, 0.09257177263498306, 0.06900633871555328, 0.15225820243358612, 0.054498545825481415, 0.05270633473992348, 0.006767008453607559, 0.06716175377368927, 0.1454283893108368, 0.13071000576019287, 0.014545821584761143, -0.016409022733569145, -0.021825823932886124, -0.011036834679543972, -0.05876464396715164, 0.04085689038038254, 0.02777833305299282, -0.1610528975725174, -0.05520197004079819, -0.12001585215330124, 0.08774644136428833, 0.09219257533550262, 0.06569026410579681, -0.2342914491891861, 0.007060535252094269, 0.08197256177663803, -0.028898365795612335, -0.1258426308631897, 0.08190665394067764, -0.021697908639907837, -0.14926569163799286, 0.0494246669113636, -0.061497997492551804, 0.12161173671483994, -0.07084709405899048, 0.08109014481306076, -0.03937468305230141, -0.062106676399707794, 0.020281726494431496, 0.1271398812532425, -0.29730626940727234, 0.20356124639511108, -0.001819691271521151, -0.05869410187005997, -0.11437822878360748, 0.01959572173655033, 0.01367559190839529, 0.11016108095645905, 0.10386832803487778, 0.005328167695552111, -0.0475030355155468, -0.12364684045314789, -0.022924374788999557, 0.024910306558012962, 0.12441114336252213, -0.05739542469382286, -0.008891535922884941, -0.044362228363752365, -0.0058176638558506966, -0.028876133263111115, -0.053936153650283813, 0.025268638506531715, -0.16888569295406342, 0.08389513194561005, 0.017658868804574013, 0.09978678822517395, 0.01261826977133751, -0.013697084039449692, -0.09944134950637817, 0.23519866168498993, -0.07718266546726227, -0.11035529524087906, -0.1205357164144516, -0.04611735790967941, 0.0686027929186821, -0.0741099938750267, 0.0634869635105133, -0.08208895474672318, 0.024847982451319695, -0.047674816101789474, -0.21411024034023285, 0.1248590424656868, -0.09078147262334824, -0.047217957675457, -0.038028888404369354, 0.1873915195465088, -0.07860055565834045, 0.003835690440610051, 0.01727161929011345, 0.03052649088203907, -0.11501652747392654, -0.10535892844200134, 0.02131424844264984, -0.005508285015821457, 0.06073078140616417, 0.04357268661260605, -0.06716573983430862, 0.01641303487122059, -0.022389056161046028, -0.006917606573551893, 0.32454678416252136, 0.14079391956329346, -0.04770330339670181, 0.17363035678863525, 0.11376409232616425, -0.08209476619958878, -0.31482723355293274, -0.08535979688167572, -0.09984239190816879, -0.03735451400279999, -0.06232178583741188, -0.21656104922294617, 0.09480288624763489, 0.04200942441821098, -0.015409117564558983, 0.1568077802658081, -0.24411429464817047, -0.0795927420258522, 0.15950311720371246, -0.007333407178521156, 0.3560895025730133, -0.12491796165704727, -0.11301901936531067, -0.05532994866371155, -0.1397564709186554, 0.15002089738845825, -0.009417316876351833, 0.11106741428375244, -0.03287123143672943, 0.10856477171182632, 0.048215944319963455, -0.05544896051287651, 0.09160676598548889, 0.026295991614460945, -0.003711326979100704, -0.10597866773605347, -0.01747799478471279, 0.043585844337940216, 0.006319248117506504, 0.031217962503433228, -0.03127649053931236, 0.033463045954704285, -0.12691029906272888, -0.04727448150515556, -0.08006873726844788, 0.05846472829580307, 0.052333541214466095, -0.0737200528383255, -0.0010956452460959554, -0.06611854583024979, -0.016030769795179367, 0.003143493551760912, 0.19045160710811615, -0.03460016846656799, 0.14779594540596008, 0.0818052664399147, 0.09073434770107269, -0.1361592561006546, -0.0061243316158652306, -0.06888517737388611, -0.057741593569517136, 0.08706554025411606, -0.10988334566354752, 0.06429524719715118, 0.11854783445596695, -0.04650293290615082, 0.07134203612804413, 0.11840200424194336, 0.015247469767928123, -0.0033181030303239822, 0.13015136122703552, -0.2568117082118988, 0.019211336970329285, -0.0754370167851448, -0.03775216266512871, 0.08088402450084686, 0.07995659112930298, 0.16486960649490356, 0.036187540739774704, -0.042049095034599304, -0.003924929536879063, 0.009187355637550354, -0.039663419127464294, 0.08243577927350998, 0.012240500189363956, 0.023174172267317772, -0.15248477458953857, 0.071900375187397, 0.015580810606479645, -0.12336304783821106, 0.011253113858401775, 0.1477922946214676, -0.13801799714565277, -0.11707340180873871, -0.03374985232949257, 0.08742405474185944, -0.14541642367839813, -0.0241269338876009, -0.04783749580383301, -0.12825986742973328, 0.09339214116334915, 0.11613135039806366, 0.07497538626194, 0.10595441609621048, -0.0529337078332901, -0.02668607421219349, -0.03682107478380203, -0.022537073120474815, -0.0017330512637272477, 0.032638516277074814, -0.08304216712713242, 0.0579586885869503, -0.020800847560167313, 0.14298540353775024, -0.08964299410581589, -0.07169508188962936, -0.1581236720085144, 0.03564200550317764, -0.12593989074230194, -0.07035141438245773, -0.08840593695640564, -0.05227470397949219, -0.007837125100195408, -0.01494099572300911, -0.0388214997947216, -0.04472146928310394, -0.12364204227924347, 0.01879296824336052, -0.05806630104780197, 0.02100815810263157, -0.07383234053850174, 0.00039667764212936163, 0.08932872861623764, -0.0410015694797039, 0.13851116597652435, 0.13557660579681396, -0.08107975125312805, 0.11907198280096054, -0.13537484407424927, -0.0908876284956932, 0.1157127171754837, 0.013428857550024986, 0.03907458856701851, 0.06849293410778046, 0.037317484617233276, 0.06514574587345123, 0.016511039808392525, 0.05237346887588501, 0.006972990930080414, -0.1299850195646286, 0.03433857858181, -0.042786743491888046, -0.1481933295726776, -0.05744143947958946, -0.05092177540063858, 0.039562974125146866, 0.02438235841691494, 0.10801149904727936, -0.03665049374103546, 0.11085481196641922, -0.058541763573884964, 0.01499281544238329, 0.004919432103633881, -0.18287403881549835, -0.044654008001089096, -0.07792776077985764, 0.02775009535253048, 0.022204352542757988, 0.2720205783843994, 0.0410233810544014, 0.020275471732020378, 0.017097288742661476, 0.11327627301216125, 0.057128578424453735, 0.015525308437645435, 0.214890718460083, 0.11996994912624359, -0.06049320101737976, -0.10806480050086975, 0.0858595222234726, 0.02164783701300621, 0.007426374591886997, 0.14070266485214233, 0.008503482677042484, -0.015597577206790447, 0.0887407436966896, -0.03357330709695816, 0.0031263602431863546, -0.11658911406993866, -0.13779941201210022, -0.028487415984272957, 0.0629650130867958, -0.0040870243683457375, 0.0956285297870636, 0.13609373569488525, -0.026881180703639984, 0.03953414782881737, -0.007877747528254986, -0.054916199296712875, -0.1785028725862503, -0.15742821991443634, -0.0790708139538765, -0.13561099767684937, 0.014744875021278858, -0.10368648171424866, 0.04369770362973213, 0.09560346603393555, 0.055915698409080505, -0.05440305173397064, 0.10839936882257462, 0.060064028948545456, -0.1045473963022232, 0.056569941341876984, -0.032912541180849075, 0.06427399069070816, -0.001812951872125268, -0.02503552846610546, -0.09098561853170395, 0.0020124134607613087, 0.0017788249533623457, 0.0514003150165081, -0.05152478814125061, 0.024474015459418297, -0.15132632851600647, -0.09570280462503433, -0.04949872940778732, 0.07316448539495468, -0.06007300689816475, 0.1162300780415535, -0.001420395914465189, -0.017011309042572975, 0.03990921378135681, 0.2064858227968216, -0.07188161462545395, -0.04990030825138092, -0.047407180070877075, 0.22449158132076263, 0.04847963526844978, 0.10619479417800903, -0.013415440917015076, -0.00436578830704093, -0.07670432329177856, 0.36612021923065186, 0.2802904546260834, -0.06149837002158165, 0.012722660787403584, 0.03524370491504669, 0.030115660279989243, 0.13885097205638885, 0.1454230099916458, 0.09396251291036606, 0.27579233050346375, -0.08266803622245789, -0.052018675953149796, -0.015770163387060165, -0.020211221650242805, -0.09714096784591675, 0.11003416776657104, 0.04697350785136223, -0.06982195377349854, -0.044631510972976685, 0.09750646352767944, -0.24107815325260162, 0.1615772694349289, -0.07760030031204224, -0.15214353799819946, -0.06177033111453056, 0.012448563240468502, 0.10150322318077087, 0.00011545186134753749, 0.08784360438585281, -0.009687529876828194, -0.10291683673858643, 0.05749227851629257, 0.02730483002960682, -0.23568211495876312, -0.007146455347537994, 0.053680915385484695, -0.04540037736296654, 0.013332240283489227, -0.01917567476630211, 0.04910791665315628, 0.06717875599861145, 0.055140718817710876, -0.0426395982503891, 0.03817736729979515, -0.010196289978921413, -0.05020907521247864, 0.029649224132299423, 0.044778332114219666, 0.017814766615629196, -0.13065220415592194, 0.05277646332979202, -0.13968263566493988, 0.041911475360393524, -0.029653942212462425, -0.027413733303546906, -0.004670299123972654, -0.019546283408999443, -0.06313455104827881, 0.057941507548093796, 0.08424945920705795, 0.001472705160267651, -0.007915833964943886, -0.08050897717475891, -0.011023934930562973, -0.012819311581552029, -0.08308050036430359, -0.10086389631032944, -0.1384236365556717, -0.10634621232748032, 0.12701933085918427, -0.017066750675439835, -0.19125573337078094, 0.01284839678555727, -0.09708964824676514, 0.060041818767786026, -0.1797112077474594, 0.0843181237578392, 0.06071038171648979, 0.01623542606830597, -0.004114143084734678, -0.029135411605238914, 0.039420004934072495, 0.08210206776857376, -0.10779064148664474, -0.09044761955738068 ]
null
null
transformers
## ะžะฟะธัะฐะฝะธะต ะผะพะดะตะปะธ ะญั‚ะพั‚ ั‡ะฐั‚ะฑะพั‚ - ะดะธะฟะปะพะผะฝะฐั ั€ะฐะฑะพั‚ะฐ ัั‚ัƒะดะตะฝั‚ะฐ ะะฝะดั€ะตั ะ’ะพั€ะพะถะบะพ ะฒ ะฃะ˜ะ˜ (ะฃะฝะธะฒะตั€ัะธั‚ะตั‚ ะ˜ัะบัƒััั‚ะฒะตะฝะฝะพะณะพ ะ˜ะฝั‚ะตะปะปะตะบั‚ะฐ). ะžะบะพะฝั‡ะฐะฝะธะต ะพะฑัƒั‡ะตะฝะธั - ะผะฐั€ั‚ 2022 ะณะพะดะฐ. ะงะฐั‚ะฑะพั‚ ัะดะตะปะฐะฝ ะฝะฐ ะพัะฝะพะฒะต ะผะพะดะตะปะธ [Kirili4ik/ruDialoGpt3-medium-finetuned-telegram](https://huggingface.co/Kirili4ik/ruDialoGpt3-medium-finetuned-telegram) ะขะตะฟะตั€ัŒ ะผะพะดะตะปัŒ ะดะพะพะฑัƒั‡ะตะฝะฐ ะฝะฐ ะพัะฝะพะฒะต 27000 ะฐะฝะตะบะดะพั‚ะพะฒ (14 ัะฟะพั…, ัะบะพั€ะพัั‚ัŒ ะพะฑัƒั‡ะตะฝะธั ะฒ ะบะพะปะฐะฑะต 2-6 ั‡ะฐัะพะฒ ะฝะฐ ัะฟะพั…ัƒ) ะธ ัƒะผะตะตั‚ ะฟะพะฝะธะผะฐั‚ัŒ ะบะพะฝั‚ะตะบัั‚ ั€ะฐะทะณะพะฒะพั€ะฐ. ะžะดะฝะฐะบะพ ะบะพะฝั‚ะตะบัั‚ ะฟั€ะธั…ะพะดะธั‚ัั ะพะณั€ะฐะฝะธั‡ะธะฒะฐั‚ัŒ ะฝะตัะบะพะปัŒะบะธะผะธ ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ ะฟะพั‚ะพะผัƒ ั‡ั‚ะพ ั‡ะตะผ ะฑะพะปัŒัˆะต ะบะพะฝั‚ะตะบัั‚ะฐ ั‚ะตะผ ะผะตะดะปะตะฝะฝะตะต ะผะพะดะตะปัŒ ั€ะฐะฑะพั‚ะฐะตั‚, ะฐ ะบะพะฝั‚ะตะบัั‚ ั€ะฐัั‚ะตั‚ ะบะฐะบ ัะฝะตะถะฝั‹ะน ะบะพะผ ะฒ ะฟั€ะพั†ะตััะต ั€ะฐะทะณะพะฒะพั€ะฐ. ะ˜ะฝั„ะตั€ะตะฝั ะฝะฐั…ะพะดะธั‚ัั ะฒ [spaces](https://huggingface.co/spaces/avorozhko/funbot): ะขะฐะผ ั ะฑะพั‚ะพะผ ะผะพะถะฝะพ ะฟะพะณะพะฒะพั€ะธั‚ัŒ. ะšะพะฝั‚ะตะบัั‚ ะพะณั€ะฐะฝะธั‡ะตะฝ 10 ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ. ะจัƒั‚ะบะธ ะฑะพั‚ ะฒั‹ะดะฐะตั‚, ะฝะพ ะฟะพะบะฐ ัะบะพั€ะตะต ัะปัƒั‡ะฐะนะฝะพ, ั‡ะตะผ ะฝะฐะผะตั€ะตะฝะฝะพ. ะžะดะฝะฐะบะพ ั€ะฐะทะณะพะฒะพั€ ะฟะพะดะดะตั€ะถะฐั‚ัŒ ัะฟะพัะพะฑะตะฝ ะธ ะดะฐะถะต ะฝะตะผะฝะพะณะพ ั€ะฐะทะฒะปะตั‡ัŒ. ะขะฐะบ ะบะฐะบ ัั‚ะพ ะณะตะฝะตั€ะฐั†ะธั ั‚ะตะบัั‚ะฐ, ั‚ะพ ะฝะฐ ะพะดะฝัƒ ะธ ั‚ัƒ ะถะต ั„ั€ะฐะทัƒ ะฑะพั‚ ะฒัะตะณะดะฐ ะฑัƒะดะตั‚ ะฒั‹ะดะฐะฒะฐั‚ัŒ ั€ะฐะทะฝั‹ะต ะพั‚ะฒะตั‚ั‹. ะขะฐะบะถะต ะดะปั ะพะฟั€ะตะดะตะปะตะฝะธั ะบะฐั‡ะตัั‚ะฒะฐ ะดะฐะฝะฝะพะน ะผะพะดะตะปะธ ะธัะฟะพะปัŒะทะพะฒะฐะปะฐััŒ ะบะฐัั‚ะพะผะฝะฐั ะผะตั‚ั€ะธะบะฐ - ัƒะณะปะพะฒะพะต ั€ะฐััั‚ะพัะฝะธั ะผะตะถะดัƒ ัะผะฑะตะดะดะธะฝะณะฐะผะธ y_train ะธ ะฟั€ะตะดะธะบั‚ะฐ. ะขะพ ะตัั‚ัŒ ะผั‹ ะฒะทัะปะธ ะฟะตั€ะฒั‹ะน ัะปะพะน ัะผะฑะตะดะดะธะฝะณะฐ ะผะพะดะตะปะธ ะธ ะฟั€ะพะณะพะฝัะปะธ ะฟั€ะตะดะธะบั‚ั‹ ะธ ะปะตะนะฑะปั‹, ะฟะพะปัƒั‡ะธะปะธ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ. ะŸะพั‚ะพะผ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ ััƒะผะผะธั€ะพะฒะฐะปะธ ะธ ะฟะพะปัƒั‡ะธะปะธ ะพะฑั‰ะธะต (ััƒะผะผะฐั€ะฝั‹ะต) ะฒะตะบั‚ะพั€ะฐ ะปะตะนะฑะปะพะฒ ะธ ะฟั€ะตะดะธะบั‚ะพะฒ. ะงะตะผ ะผะตะฝัŒัˆะต ัƒะณะพะป ะผะตะถะดัƒ ะฝะธะผะธ, ั‚ะตะผ ะปัƒั‡ัˆะต. ะŸั€ะธ ั€ะฐััั‡ะตั‚ะฐั… ะพั€ะธะตะฝั‚ะธั€ะพะฒะฐะปะธััŒ ะฝะฐ ะบะพัะธะฝัƒั ัั‚ะพะณะพ ัƒะณะปะฐ, ั‚ะฐะบ ะบะฐะบ cos 0 = 1, ั‚ะพ ัั‚ะพ ะพั‡ะตะฝัŒ ัƒะดะพะฑะฝะพ - ั‡ะตะผ ะฑะปะธะถะต ะฟะพะบะฐะทะฐั‚ะตะปัŒ ะบ 1, ั‚ะตะผ ะปัƒั‡ัˆะต. ะ’ะพั‚ ั‚ะฐะบะพะต ั€ะฐัะฟั€ะตะดะตะปะตะฝะธะต ัั‚ะธั… ะทะฝะฐั‡ะตะฝะธะน ะฟะพะปัƒั‡ะธะปะพััŒ ะฟะพ ัะฟะพั…ะฐะผ ะฝะฐ ะŸะ ะžะ’ะ•ะ ะžะงะะžะ™ ะฒั‹ะฑะพั€ะบะต (1406 ะฐะฝะตะบะดะพั‚ะพะฒ): ``` {1: tensor(0.9357, device='cuda:0', grad_fn=<DivBackward0>), 2: tensor(0.9390, device='cuda:0', grad_fn=<DivBackward0>), 3: tensor(0.9417, device='cuda:0', grad_fn=<DivBackward0>), 4: tensor(0.9439, device='cuda:0', grad_fn=<DivBackward0>), 5: tensor(0.9470, device='cuda:0', grad_fn=<DivBackward0>), 6: tensor(0.9537, device='cuda:0', grad_fn=<DivBackward0>), 7: tensor(0.9568, device='cuda:0', grad_fn=<DivBackward0>), 8: tensor(0.9592, device='cuda:0', grad_fn=<DivBackward0>), 9: tensor(0.9610, device='cuda:0', grad_fn=<DivBackward0>), 10: tensor(0.9622, device='cuda:0', grad_fn=<DivBackward0>), 11: tensor(0.9628, device='cuda:0', grad_fn=<DivBackward0>), 12: tensor(0.9632, device='cuda:0', grad_fn=<DivBackward0>), 13: tensor(0.9630, device='cuda:0', grad_fn=<DivBackward0>), 14: tensor(0.9634, device='cuda:0', grad_fn=<DivBackward0>), 15: tensor(0.9634, device='cuda:0', grad_fn=<DivBackward0>)} ``` ะ”ะปั ะธะฝั„ะตั€ะตะฝัะฐ ะฒั‹ะฑั€ะฐะฝะฐ 14-ั ัะฟะพั…ะฐ ั ั‚ะพั‡ะฝะพัั‚ัŒัŽ 0.9634. ะ”ะฐะปะตะต, ััƒะดั ะฟะพ ะฒัะตะผัƒ ะธะดะตั‚ ัƒะถะต ะฟะตั€ะตะพะฑัƒั‡ะตะฝะธะต.
{}
text-generation
avorozhko/ruDialoGpt3-medium-finetuned-context
[ "transformers", "pytorch", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
## ะžะฟะธัะฐะฝะธะต ะผะพะดะตะปะธ ะญั‚ะพั‚ ั‡ะฐั‚ะฑะพั‚ - ะดะธะฟะปะพะผะฝะฐั ั€ะฐะฑะพั‚ะฐ ัั‚ัƒะดะตะฝั‚ะฐ ะะฝะดั€ะตั ะ’ะพั€ะพะถะบะพ ะฒ ะฃะ˜ะ˜ (ะฃะฝะธะฒะตั€ัะธั‚ะตั‚ ะ˜ัะบัƒััั‚ะฒะตะฝะฝะพะณะพ ะ˜ะฝั‚ะตะปะปะตะบั‚ะฐ). ะžะบะพะฝั‡ะฐะฝะธะต ะพะฑัƒั‡ะตะฝะธั - ะผะฐั€ั‚ 2022 ะณะพะดะฐ. ะงะฐั‚ะฑะพั‚ ัะดะตะปะฐะฝ ะฝะฐ ะพัะฝะพะฒะต ะผะพะดะตะปะธ Kirili4ik/ruDialoGpt3-medium-finetuned-telegram ะขะตะฟะตั€ัŒ ะผะพะดะตะปัŒ ะดะพะพะฑัƒั‡ะตะฝะฐ ะฝะฐ ะพัะฝะพะฒะต 27000 ะฐะฝะตะบะดะพั‚ะพะฒ (14 ัะฟะพั…, ัะบะพั€ะพัั‚ัŒ ะพะฑัƒั‡ะตะฝะธั ะฒ ะบะพะปะฐะฑะต 2-6 ั‡ะฐัะพะฒ ะฝะฐ ัะฟะพั…ัƒ) ะธ ัƒะผะตะตั‚ ะฟะพะฝะธะผะฐั‚ัŒ ะบะพะฝั‚ะตะบัั‚ ั€ะฐะทะณะพะฒะพั€ะฐ. ะžะดะฝะฐะบะพ ะบะพะฝั‚ะตะบัั‚ ะฟั€ะธั…ะพะดะธั‚ัั ะพะณั€ะฐะฝะธั‡ะธะฒะฐั‚ัŒ ะฝะตัะบะพะปัŒะบะธะผะธ ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ ะฟะพั‚ะพะผัƒ ั‡ั‚ะพ ั‡ะตะผ ะฑะพะปัŒัˆะต ะบะพะฝั‚ะตะบัั‚ะฐ ั‚ะตะผ ะผะตะดะปะตะฝะฝะตะต ะผะพะดะตะปัŒ ั€ะฐะฑะพั‚ะฐะตั‚, ะฐ ะบะพะฝั‚ะตะบัั‚ ั€ะฐัั‚ะตั‚ ะบะฐะบ ัะฝะตะถะฝั‹ะน ะบะพะผ ะฒ ะฟั€ะพั†ะตััะต ั€ะฐะทะณะพะฒะพั€ะฐ. ะ˜ะฝั„ะตั€ะตะฝั ะฝะฐั…ะพะดะธั‚ัั ะฒ spaces: ะขะฐะผ ั ะฑะพั‚ะพะผ ะผะพะถะฝะพ ะฟะพะณะพะฒะพั€ะธั‚ัŒ. ะšะพะฝั‚ะตะบัั‚ ะพะณั€ะฐะฝะธั‡ะตะฝ 10 ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ. ะจัƒั‚ะบะธ ะฑะพั‚ ะฒั‹ะดะฐะตั‚, ะฝะพ ะฟะพะบะฐ ัะบะพั€ะตะต ัะปัƒั‡ะฐะนะฝะพ, ั‡ะตะผ ะฝะฐะผะตั€ะตะฝะฝะพ. ะžะดะฝะฐะบะพ ั€ะฐะทะณะพะฒะพั€ ะฟะพะดะดะตั€ะถะฐั‚ัŒ ัะฟะพัะพะฑะตะฝ ะธ ะดะฐะถะต ะฝะตะผะฝะพะณะพ ั€ะฐะทะฒะปะตั‡ัŒ. ะขะฐะบ ะบะฐะบ ัั‚ะพ ะณะตะฝะตั€ะฐั†ะธั ั‚ะตะบัั‚ะฐ, ั‚ะพ ะฝะฐ ะพะดะฝัƒ ะธ ั‚ัƒ ะถะต ั„ั€ะฐะทัƒ ะฑะพั‚ ะฒัะตะณะดะฐ ะฑัƒะดะตั‚ ะฒั‹ะดะฐะฒะฐั‚ัŒ ั€ะฐะทะฝั‹ะต ะพั‚ะฒะตั‚ั‹. ะขะฐะบะถะต ะดะปั ะพะฟั€ะตะดะตะปะตะฝะธั ะบะฐั‡ะตัั‚ะฒะฐ ะดะฐะฝะฝะพะน ะผะพะดะตะปะธ ะธัะฟะพะปัŒะทะพะฒะฐะปะฐััŒ ะบะฐัั‚ะพะผะฝะฐั ะผะตั‚ั€ะธะบะฐ - ัƒะณะปะพะฒะพะต ั€ะฐััั‚ะพัะฝะธั ะผะตะถะดัƒ ัะผะฑะตะดะดะธะฝะณะฐะผะธ y_train ะธ ะฟั€ะตะดะธะบั‚ะฐ. ะขะพ ะตัั‚ัŒ ะผั‹ ะฒะทัะปะธ ะฟะตั€ะฒั‹ะน ัะปะพะน ัะผะฑะตะดะดะธะฝะณะฐ ะผะพะดะตะปะธ ะธ ะฟั€ะพะณะพะฝัะปะธ ะฟั€ะตะดะธะบั‚ั‹ ะธ ะปะตะนะฑะปั‹, ะฟะพะปัƒั‡ะธะปะธ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ. ะŸะพั‚ะพะผ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ ััƒะผะผะธั€ะพะฒะฐะปะธ ะธ ะฟะพะปัƒั‡ะธะปะธ ะพะฑั‰ะธะต (ััƒะผะผะฐั€ะฝั‹ะต) ะฒะตะบั‚ะพั€ะฐ ะปะตะนะฑะปะพะฒ ะธ ะฟั€ะตะดะธะบั‚ะพะฒ. ะงะตะผ ะผะตะฝัŒัˆะต ัƒะณะพะป ะผะตะถะดัƒ ะฝะธะผะธ, ั‚ะตะผ ะปัƒั‡ัˆะต. ะŸั€ะธ ั€ะฐััั‡ะตั‚ะฐั… ะพั€ะธะตะฝั‚ะธั€ะพะฒะฐะปะธััŒ ะฝะฐ ะบะพัะธะฝัƒั ัั‚ะพะณะพ ัƒะณะปะฐ, ั‚ะฐะบ ะบะฐะบ cos 0 = 1, ั‚ะพ ัั‚ะพ ะพั‡ะตะฝัŒ ัƒะดะพะฑะฝะพ - ั‡ะตะผ ะฑะปะธะถะต ะฟะพะบะฐะทะฐั‚ะตะปัŒ ะบ 1, ั‚ะตะผ ะปัƒั‡ัˆะต. ะ’ะพั‚ ั‚ะฐะบะพะต ั€ะฐัะฟั€ะตะดะตะปะตะฝะธะต ัั‚ะธั… ะทะฝะฐั‡ะตะฝะธะน ะฟะพะปัƒั‡ะธะปะพััŒ ะฟะพ ัะฟะพั…ะฐะผ ะฝะฐ ะŸะ ะžะ’ะ•ะ ะžะงะะžะ™ ะฒั‹ะฑะพั€ะบะต (1406 ะฐะฝะตะบะดะพั‚ะพะฒ): ะ”ะปั ะธะฝั„ะตั€ะตะฝัะฐ ะฒั‹ะฑั€ะฐะฝะฐ 14-ั ัะฟะพั…ะฐ ั ั‚ะพั‡ะฝะพัั‚ัŒัŽ 0.9634. ะ”ะฐะปะตะต, ััƒะดั ะฟะพ ะฒัะตะผัƒ ะธะดะตั‚ ัƒะถะต ะฟะตั€ะตะพะฑัƒั‡ะตะฝะธะต.
[ "## ะžะฟะธัะฐะฝะธะต ะผะพะดะตะปะธ\n\nะญั‚ะพั‚ ั‡ะฐั‚ะฑะพั‚ - ะดะธะฟะปะพะผะฝะฐั ั€ะฐะฑะพั‚ะฐ ัั‚ัƒะดะตะฝั‚ะฐ ะะฝะดั€ะตั ะ’ะพั€ะพะถะบะพ ะฒ ะฃะ˜ะ˜ (ะฃะฝะธะฒะตั€ัะธั‚ะตั‚ ะ˜ัะบัƒััั‚ะฒะตะฝะฝะพะณะพ ะ˜ะฝั‚ะตะปะปะตะบั‚ะฐ).\n\nะžะบะพะฝั‡ะฐะฝะธะต ะพะฑัƒั‡ะตะฝะธั - ะผะฐั€ั‚ 2022 ะณะพะดะฐ.\n\nะงะฐั‚ะฑะพั‚ ัะดะตะปะฐะฝ ะฝะฐ ะพัะฝะพะฒะต ะผะพะดะตะปะธ Kirili4ik/ruDialoGpt3-medium-finetuned-telegram\n\nะขะตะฟะตั€ัŒ ะผะพะดะตะปัŒ ะดะพะพะฑัƒั‡ะตะฝะฐ ะฝะฐ ะพัะฝะพะฒะต 27000 ะฐะฝะตะบะดะพั‚ะพะฒ (14 ัะฟะพั…, ัะบะพั€ะพัั‚ัŒ ะพะฑัƒั‡ะตะฝะธั ะฒ ะบะพะปะฐะฑะต 2-6 ั‡ะฐัะพะฒ ะฝะฐ ัะฟะพั…ัƒ) ะธ ัƒะผะตะตั‚ ะฟะพะฝะธะผะฐั‚ัŒ ะบะพะฝั‚ะตะบัั‚ ั€ะฐะทะณะพะฒะพั€ะฐ. ะžะดะฝะฐะบะพ ะบะพะฝั‚ะตะบัั‚ ะฟั€ะธั…ะพะดะธั‚ัั ะพะณั€ะฐะฝะธั‡ะธะฒะฐั‚ัŒ ะฝะตัะบะพะปัŒะบะธะผะธ ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ ะฟะพั‚ะพะผัƒ ั‡ั‚ะพ ั‡ะตะผ ะฑะพะปัŒัˆะต ะบะพะฝั‚ะตะบัั‚ะฐ ั‚ะตะผ ะผะตะดะปะตะฝะฝะตะต ะผะพะดะตะปัŒ ั€ะฐะฑะพั‚ะฐะตั‚, ะฐ ะบะพะฝั‚ะตะบัั‚ ั€ะฐัั‚ะตั‚ ะบะฐะบ ัะฝะตะถะฝั‹ะน ะบะพะผ ะฒ ะฟั€ะพั†ะตััะต ั€ะฐะทะณะพะฒะพั€ะฐ.\n\nะ˜ะฝั„ะตั€ะตะฝั ะฝะฐั…ะพะดะธั‚ัั ะฒ spaces:\n\nะขะฐะผ ั ะฑะพั‚ะพะผ ะผะพะถะฝะพ ะฟะพะณะพะฒะพั€ะธั‚ัŒ. ะšะพะฝั‚ะตะบัั‚ ะพะณั€ะฐะฝะธั‡ะตะฝ 10 ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ.\n\nะจัƒั‚ะบะธ ะฑะพั‚ ะฒั‹ะดะฐะตั‚, ะฝะพ ะฟะพะบะฐ ัะบะพั€ะตะต ัะปัƒั‡ะฐะนะฝะพ, ั‡ะตะผ ะฝะฐะผะตั€ะตะฝะฝะพ. ะžะดะฝะฐะบะพ ั€ะฐะทะณะพะฒะพั€ ะฟะพะดะดะตั€ะถะฐั‚ัŒ ัะฟะพัะพะฑะตะฝ ะธ ะดะฐะถะต ะฝะตะผะฝะพะณะพ ั€ะฐะทะฒะปะตั‡ัŒ.\n\nะขะฐะบ ะบะฐะบ ัั‚ะพ ะณะตะฝะตั€ะฐั†ะธั ั‚ะตะบัั‚ะฐ, ั‚ะพ ะฝะฐ ะพะดะฝัƒ ะธ ั‚ัƒ ะถะต ั„ั€ะฐะทัƒ ะฑะพั‚ ะฒัะตะณะดะฐ ะฑัƒะดะตั‚ ะฒั‹ะดะฐะฒะฐั‚ัŒ ั€ะฐะทะฝั‹ะต ะพั‚ะฒะตั‚ั‹.\n\nะขะฐะบะถะต ะดะปั ะพะฟั€ะตะดะตะปะตะฝะธั ะบะฐั‡ะตัั‚ะฒะฐ ะดะฐะฝะฝะพะน ะผะพะดะตะปะธ ะธัะฟะพะปัŒะทะพะฒะฐะปะฐััŒ ะบะฐัั‚ะพะผะฝะฐั ะผะตั‚ั€ะธะบะฐ - ัƒะณะปะพะฒะพะต ั€ะฐััั‚ะพัะฝะธั ะผะตะถะดัƒ ัะผะฑะตะดะดะธะฝะณะฐะผะธ y_train ะธ ะฟั€ะตะดะธะบั‚ะฐ.\n\nะขะพ ะตัั‚ัŒ ะผั‹ ะฒะทัะปะธ ะฟะตั€ะฒั‹ะน ัะปะพะน ัะผะฑะตะดะดะธะฝะณะฐ ะผะพะดะตะปะธ ะธ ะฟั€ะพะณะพะฝัะปะธ ะฟั€ะตะดะธะบั‚ั‹ ะธ ะปะตะนะฑะปั‹, ะฟะพะปัƒั‡ะธะปะธ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ. ะŸะพั‚ะพะผ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ ััƒะผะผะธั€ะพะฒะฐะปะธ ะธ ะฟะพะปัƒั‡ะธะปะธ ะพะฑั‰ะธะต (ััƒะผะผะฐั€ะฝั‹ะต) ะฒะตะบั‚ะพั€ะฐ ะปะตะนะฑะปะพะฒ ะธ ะฟั€ะตะดะธะบั‚ะพะฒ. ะงะตะผ ะผะตะฝัŒัˆะต ัƒะณะพะป ะผะตะถะดัƒ ะฝะธะผะธ, ั‚ะตะผ ะปัƒั‡ัˆะต. ะŸั€ะธ ั€ะฐััั‡ะตั‚ะฐั… ะพั€ะธะตะฝั‚ะธั€ะพะฒะฐะปะธััŒ ะฝะฐ ะบะพัะธะฝัƒั ัั‚ะพะณะพ ัƒะณะปะฐ, ั‚ะฐะบ ะบะฐะบ cos 0 = 1, ั‚ะพ ัั‚ะพ ะพั‡ะตะฝัŒ ัƒะดะพะฑะฝะพ - ั‡ะตะผ ะฑะปะธะถะต ะฟะพะบะฐะทะฐั‚ะตะปัŒ ะบ 1, ั‚ะตะผ ะปัƒั‡ัˆะต.\n\nะ’ะพั‚ ั‚ะฐะบะพะต ั€ะฐัะฟั€ะตะดะตะปะตะฝะธะต ัั‚ะธั… ะทะฝะฐั‡ะตะฝะธะน ะฟะพะปัƒั‡ะธะปะพััŒ ะฟะพ ัะฟะพั…ะฐะผ ะฝะฐ ะŸะ ะžะ’ะ•ะ ะžะงะะžะ™ ะฒั‹ะฑะพั€ะบะต (1406 ะฐะฝะตะบะดะพั‚ะพะฒ):\n\nะ”ะปั ะธะฝั„ะตั€ะตะฝัะฐ ะฒั‹ะฑั€ะฐะฝะฐ 14-ั ัะฟะพั…ะฐ ั ั‚ะพั‡ะฝะพัั‚ัŒัŽ 0.9634. ะ”ะฐะปะตะต, ััƒะดั ะฟะพ ะฒัะตะผัƒ ะธะดะตั‚ ัƒะถะต ะฟะตั€ะตะพะฑัƒั‡ะตะฝะธะต." ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n", "## ะžะฟะธัะฐะฝะธะต ะผะพะดะตะปะธ\n\nะญั‚ะพั‚ ั‡ะฐั‚ะฑะพั‚ - ะดะธะฟะปะพะผะฝะฐั ั€ะฐะฑะพั‚ะฐ ัั‚ัƒะดะตะฝั‚ะฐ ะะฝะดั€ะตั ะ’ะพั€ะพะถะบะพ ะฒ ะฃะ˜ะ˜ (ะฃะฝะธะฒะตั€ัะธั‚ะตั‚ ะ˜ัะบัƒััั‚ะฒะตะฝะฝะพะณะพ ะ˜ะฝั‚ะตะปะปะตะบั‚ะฐ).\n\nะžะบะพะฝั‡ะฐะฝะธะต ะพะฑัƒั‡ะตะฝะธั - ะผะฐั€ั‚ 2022 ะณะพะดะฐ.\n\nะงะฐั‚ะฑะพั‚ ัะดะตะปะฐะฝ ะฝะฐ ะพัะฝะพะฒะต ะผะพะดะตะปะธ Kirili4ik/ruDialoGpt3-medium-finetuned-telegram\n\nะขะตะฟะตั€ัŒ ะผะพะดะตะปัŒ ะดะพะพะฑัƒั‡ะตะฝะฐ ะฝะฐ ะพัะฝะพะฒะต 27000 ะฐะฝะตะบะดะพั‚ะพะฒ (14 ัะฟะพั…, ัะบะพั€ะพัั‚ัŒ ะพะฑัƒั‡ะตะฝะธั ะฒ ะบะพะปะฐะฑะต 2-6 ั‡ะฐัะพะฒ ะฝะฐ ัะฟะพั…ัƒ) ะธ ัƒะผะตะตั‚ ะฟะพะฝะธะผะฐั‚ัŒ ะบะพะฝั‚ะตะบัั‚ ั€ะฐะทะณะพะฒะพั€ะฐ. ะžะดะฝะฐะบะพ ะบะพะฝั‚ะตะบัั‚ ะฟั€ะธั…ะพะดะธั‚ัั ะพะณั€ะฐะฝะธั‡ะธะฒะฐั‚ัŒ ะฝะตัะบะพะปัŒะบะธะผะธ ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ ะฟะพั‚ะพะผัƒ ั‡ั‚ะพ ั‡ะตะผ ะฑะพะปัŒัˆะต ะบะพะฝั‚ะตะบัั‚ะฐ ั‚ะตะผ ะผะตะดะปะตะฝะฝะตะต ะผะพะดะตะปัŒ ั€ะฐะฑะพั‚ะฐะตั‚, ะฐ ะบะพะฝั‚ะตะบัั‚ ั€ะฐัั‚ะตั‚ ะบะฐะบ ัะฝะตะถะฝั‹ะน ะบะพะผ ะฒ ะฟั€ะพั†ะตััะต ั€ะฐะทะณะพะฒะพั€ะฐ.\n\nะ˜ะฝั„ะตั€ะตะฝั ะฝะฐั…ะพะดะธั‚ัั ะฒ spaces:\n\nะขะฐะผ ั ะฑะพั‚ะพะผ ะผะพะถะฝะพ ะฟะพะณะพะฒะพั€ะธั‚ัŒ. ะšะพะฝั‚ะตะบัั‚ ะพะณั€ะฐะฝะธั‡ะตะฝ 10 ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ.\n\nะจัƒั‚ะบะธ ะฑะพั‚ ะฒั‹ะดะฐะตั‚, ะฝะพ ะฟะพะบะฐ ัะบะพั€ะตะต ัะปัƒั‡ะฐะนะฝะพ, ั‡ะตะผ ะฝะฐะผะตั€ะตะฝะฝะพ. ะžะดะฝะฐะบะพ ั€ะฐะทะณะพะฒะพั€ ะฟะพะดะดะตั€ะถะฐั‚ัŒ ัะฟะพัะพะฑะตะฝ ะธ ะดะฐะถะต ะฝะตะผะฝะพะณะพ ั€ะฐะทะฒะปะตั‡ัŒ.\n\nะขะฐะบ ะบะฐะบ ัั‚ะพ ะณะตะฝะตั€ะฐั†ะธั ั‚ะตะบัั‚ะฐ, ั‚ะพ ะฝะฐ ะพะดะฝัƒ ะธ ั‚ัƒ ะถะต ั„ั€ะฐะทัƒ ะฑะพั‚ ะฒัะตะณะดะฐ ะฑัƒะดะตั‚ ะฒั‹ะดะฐะฒะฐั‚ัŒ ั€ะฐะทะฝั‹ะต ะพั‚ะฒะตั‚ั‹.\n\nะขะฐะบะถะต ะดะปั ะพะฟั€ะตะดะตะปะตะฝะธั ะบะฐั‡ะตัั‚ะฒะฐ ะดะฐะฝะฝะพะน ะผะพะดะตะปะธ ะธัะฟะพะปัŒะทะพะฒะฐะปะฐััŒ ะบะฐัั‚ะพะผะฝะฐั ะผะตั‚ั€ะธะบะฐ - ัƒะณะปะพะฒะพะต ั€ะฐััั‚ะพัะฝะธั ะผะตะถะดัƒ ัะผะฑะตะดะดะธะฝะณะฐะผะธ y_train ะธ ะฟั€ะตะดะธะบั‚ะฐ.\n\nะขะพ ะตัั‚ัŒ ะผั‹ ะฒะทัะปะธ ะฟะตั€ะฒั‹ะน ัะปะพะน ัะผะฑะตะดะดะธะฝะณะฐ ะผะพะดะตะปะธ ะธ ะฟั€ะพะณะพะฝัะปะธ ะฟั€ะตะดะธะบั‚ั‹ ะธ ะปะตะนะฑะปั‹, ะฟะพะปัƒั‡ะธะปะธ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ. ะŸะพั‚ะพะผ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ ััƒะผะผะธั€ะพะฒะฐะปะธ ะธ ะฟะพะปัƒั‡ะธะปะธ ะพะฑั‰ะธะต (ััƒะผะผะฐั€ะฝั‹ะต) ะฒะตะบั‚ะพั€ะฐ ะปะตะนะฑะปะพะฒ ะธ ะฟั€ะตะดะธะบั‚ะพะฒ. ะงะตะผ ะผะตะฝัŒัˆะต ัƒะณะพะป ะผะตะถะดัƒ ะฝะธะผะธ, ั‚ะตะผ ะปัƒั‡ัˆะต. ะŸั€ะธ ั€ะฐััั‡ะตั‚ะฐั… ะพั€ะธะตะฝั‚ะธั€ะพะฒะฐะปะธััŒ ะฝะฐ ะบะพัะธะฝัƒั ัั‚ะพะณะพ ัƒะณะปะฐ, ั‚ะฐะบ ะบะฐะบ cos 0 = 1, ั‚ะพ ัั‚ะพ ะพั‡ะตะฝัŒ ัƒะดะพะฑะฝะพ - ั‡ะตะผ ะฑะปะธะถะต ะฟะพะบะฐะทะฐั‚ะตะปัŒ ะบ 1, ั‚ะตะผ ะปัƒั‡ัˆะต.\n\nะ’ะพั‚ ั‚ะฐะบะพะต ั€ะฐัะฟั€ะตะดะตะปะตะฝะธะต ัั‚ะธั… ะทะฝะฐั‡ะตะฝะธะน ะฟะพะปัƒั‡ะธะปะพััŒ ะฟะพ ัะฟะพั…ะฐะผ ะฝะฐ ะŸะ ะžะ’ะ•ะ ะžะงะะžะ™ ะฒั‹ะฑะพั€ะบะต (1406 ะฐะฝะตะบะดะพั‚ะพะฒ):\n\nะ”ะปั ะธะฝั„ะตั€ะตะฝัะฐ ะฒั‹ะฑั€ะฐะฝะฐ 14-ั ัะฟะพั…ะฐ ั ั‚ะพั‡ะฝะพัั‚ัŒัŽ 0.9634. ะ”ะฐะปะตะต, ััƒะดั ะฟะพ ะฒัะตะผัƒ ะธะดะตั‚ ัƒะถะต ะฟะตั€ะตะพะฑัƒั‡ะตะฝะธะต." ]
[ 51, 418 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n## ะžะฟะธัะฐะฝะธะต ะผะพะดะตะปะธ\n\nะญั‚ะพั‚ ั‡ะฐั‚ะฑะพั‚ - ะดะธะฟะปะพะผะฝะฐั ั€ะฐะฑะพั‚ะฐ ัั‚ัƒะดะตะฝั‚ะฐ ะะฝะดั€ะตั ะ’ะพั€ะพะถะบะพ ะฒ ะฃะ˜ะ˜ (ะฃะฝะธะฒะตั€ัะธั‚ะตั‚ ะ˜ัะบัƒััั‚ะฒะตะฝะฝะพะณะพ ะ˜ะฝั‚ะตะปะปะตะบั‚ะฐ).\n\nะžะบะพะฝั‡ะฐะฝะธะต ะพะฑัƒั‡ะตะฝะธั - ะผะฐั€ั‚ 2022 ะณะพะดะฐ.\n\nะงะฐั‚ะฑะพั‚ ัะดะตะปะฐะฝ ะฝะฐ ะพัะฝะพะฒะต ะผะพะดะตะปะธ Kirili4ik/ruDialoGpt3-medium-finetuned-telegram\n\nะขะตะฟะตั€ัŒ ะผะพะดะตะปัŒ ะดะพะพะฑัƒั‡ะตะฝะฐ ะฝะฐ ะพัะฝะพะฒะต 27000 ะฐะฝะตะบะดะพั‚ะพะฒ (14 ัะฟะพั…, ัะบะพั€ะพัั‚ัŒ ะพะฑัƒั‡ะตะฝะธั ะฒ ะบะพะปะฐะฑะต 2-6 ั‡ะฐัะพะฒ ะฝะฐ ัะฟะพั…ัƒ) ะธ ัƒะผะตะตั‚ ะฟะพะฝะธะผะฐั‚ัŒ ะบะพะฝั‚ะตะบัั‚ ั€ะฐะทะณะพะฒะพั€ะฐ. ะžะดะฝะฐะบะพ ะบะพะฝั‚ะตะบัั‚ ะฟั€ะธั…ะพะดะธั‚ัั ะพะณั€ะฐะฝะธั‡ะธะฒะฐั‚ัŒ ะฝะตัะบะพะปัŒะบะธะผะธ ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ ะฟะพั‚ะพะผัƒ ั‡ั‚ะพ ั‡ะตะผ ะฑะพะปัŒัˆะต ะบะพะฝั‚ะตะบัั‚ะฐ ั‚ะตะผ ะผะตะดะปะตะฝะฝะตะต ะผะพะดะตะปัŒ ั€ะฐะฑะพั‚ะฐะตั‚, ะฐ ะบะพะฝั‚ะตะบัั‚ ั€ะฐัั‚ะตั‚ ะบะฐะบ ัะฝะตะถะฝั‹ะน ะบะพะผ ะฒ ะฟั€ะพั†ะตััะต ั€ะฐะทะณะพะฒะพั€ะฐ.\n\nะ˜ะฝั„ะตั€ะตะฝั ะฝะฐั…ะพะดะธั‚ัั ะฒ spaces:\n\nะขะฐะผ ั ะฑะพั‚ะพะผ ะผะพะถะฝะพ ะฟะพะณะพะฒะพั€ะธั‚ัŒ. ะšะพะฝั‚ะตะบัั‚ ะพะณั€ะฐะฝะธั‡ะตะฝ 10 ะฟะพัะปะตะดะฝะธะผะธ ัะพะพะฑั‰ะตะฝะธัะผะธ.\n\nะจัƒั‚ะบะธ ะฑะพั‚ ะฒั‹ะดะฐะตั‚, ะฝะพ ะฟะพะบะฐ ัะบะพั€ะตะต ัะปัƒั‡ะฐะนะฝะพ, ั‡ะตะผ ะฝะฐะผะตั€ะตะฝะฝะพ. ะžะดะฝะฐะบะพ ั€ะฐะทะณะพะฒะพั€ ะฟะพะดะดะตั€ะถะฐั‚ัŒ ัะฟะพัะพะฑะตะฝ ะธ ะดะฐะถะต ะฝะตะผะฝะพะณะพ ั€ะฐะทะฒะปะตั‡ัŒ.\n\nะขะฐะบ ะบะฐะบ ัั‚ะพ ะณะตะฝะตั€ะฐั†ะธั ั‚ะตะบัั‚ะฐ, ั‚ะพ ะฝะฐ ะพะดะฝัƒ ะธ ั‚ัƒ ะถะต ั„ั€ะฐะทัƒ ะฑะพั‚ ะฒัะตะณะดะฐ ะฑัƒะดะตั‚ ะฒั‹ะดะฐะฒะฐั‚ัŒ ั€ะฐะทะฝั‹ะต ะพั‚ะฒะตั‚ั‹.\n\nะขะฐะบะถะต ะดะปั ะพะฟั€ะตะดะตะปะตะฝะธั ะบะฐั‡ะตัั‚ะฒะฐ ะดะฐะฝะฝะพะน ะผะพะดะตะปะธ ะธัะฟะพะปัŒะทะพะฒะฐะปะฐััŒ ะบะฐัั‚ะพะผะฝะฐั ะผะตั‚ั€ะธะบะฐ - ัƒะณะปะพะฒะพะต ั€ะฐััั‚ะพัะฝะธั ะผะตะถะดัƒ ัะผะฑะตะดะดะธะฝะณะฐะผะธ y_train ะธ ะฟั€ะตะดะธะบั‚ะฐ.\n\nะขะพ ะตัั‚ัŒ ะผั‹ ะฒะทัะปะธ ะฟะตั€ะฒั‹ะน ัะปะพะน ัะผะฑะตะดะดะธะฝะณะฐ ะผะพะดะตะปะธ ะธ ะฟั€ะพะณะพะฝัะปะธ ะฟั€ะตะดะธะบั‚ั‹ ะธ ะปะตะนะฑะปั‹, ะฟะพะปัƒั‡ะธะปะธ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ. ะŸะพั‚ะพะผ ะฒะตะบั‚ะพั€ะฐ ัะปะพะฒ ััƒะผะผะธั€ะพะฒะฐะปะธ ะธ ะฟะพะปัƒั‡ะธะปะธ ะพะฑั‰ะธะต (ััƒะผะผะฐั€ะฝั‹ะต) ะฒะตะบั‚ะพั€ะฐ ะปะตะนะฑะปะพะฒ ะธ ะฟั€ะตะดะธะบั‚ะพะฒ. ะงะตะผ ะผะตะฝัŒัˆะต ัƒะณะพะป ะผะตะถะดัƒ ะฝะธะผะธ, ั‚ะตะผ ะปัƒั‡ัˆะต. ะŸั€ะธ ั€ะฐััั‡ะตั‚ะฐั… ะพั€ะธะตะฝั‚ะธั€ะพะฒะฐะปะธััŒ ะฝะฐ ะบะพัะธะฝัƒั ัั‚ะพะณะพ ัƒะณะปะฐ, ั‚ะฐะบ ะบะฐะบ cos 0 = 1, ั‚ะพ ัั‚ะพ ะพั‡ะตะฝัŒ ัƒะดะพะฑะฝะพ - ั‡ะตะผ ะฑะปะธะถะต ะฟะพะบะฐะทะฐั‚ะตะปัŒ ะบ 1, ั‚ะตะผ ะปัƒั‡ัˆะต.\n\nะ’ะพั‚ ั‚ะฐะบะพะต ั€ะฐัะฟั€ะตะดะตะปะตะฝะธะต ัั‚ะธั… ะทะฝะฐั‡ะตะฝะธะน ะฟะพะปัƒั‡ะธะปะพััŒ ะฟะพ ัะฟะพั…ะฐะผ ะฝะฐ ะŸะ ะžะ’ะ•ะ ะžะงะะžะ™ ะฒั‹ะฑะพั€ะบะต (1406 ะฐะฝะตะบะดะพั‚ะพะฒ):\n\nะ”ะปั ะธะฝั„ะตั€ะตะฝัะฐ ะฒั‹ะฑั€ะฐะฝะฐ 14-ั ัะฟะพั…ะฐ ั ั‚ะพั‡ะฝะพัั‚ัŒัŽ 0.9634. ะ”ะฐะปะตะต, ััƒะดั ะฟะพ ะฒัะตะผัƒ ะธะดะตั‚ ัƒะถะต ะฟะตั€ะตะพะฑัƒั‡ะตะฝะธะต." ]
[ 0.015265676192939281, 0.009312869980931282, -0.01757386326789856, 0.02179829776287079, 0.04184107482433319, -0.011989162303507328, 0.09330148994922638, 0.03334183618426323, 0.02116091549396515, 0.06668468564748764, 0.05700088292360306, 0.08391701430082321, 0.0359375923871994, 0.07228323817253113, 0.0005198165308684111, -0.13924869894981384, 0.05801243335008621, -0.045891325920820236, -0.009293947368860245, 0.061273910105228424, 0.07978726923465729, -0.07229633629322052, 0.04845258593559265, 0.028760069981217384, -0.0370023138821125, 0.00015626236563548446, 0.003514431416988373, -0.018103569746017456, 0.06412799656391144, 0.04671066999435425, 0.07415737956762314, 0.06519564241170883, -0.009020816534757614, -0.08950625360012054, 0.028430592268705368, 0.04296473413705826, 0.01185925304889679, 0.028158338740468025, -0.017735863104462624, -0.007076045498251915, 0.06915900111198425, -0.03425327315926552, 0.02715950645506382, 0.01450031716376543, -0.11285745352506638, 0.08009268343448639, -0.04164491966366768, -0.02667240798473358, 0.0714595839381218, 0.04582856222987175, -0.02593633346259594, 0.09751206636428833, -0.06419561058282852, 0.037463586777448654, 0.12510988116264343, -0.20541439950466156, -0.03922111168503761, 0.06020621210336685, 0.012952513061463833, 0.0025739583652466536, -0.055977657437324524, 0.013479873538017273, 0.04096995294094086, -0.0033207559026777744, 0.00023973442148417234, -0.003321430180221796, 0.06795147061347961, -0.013663469813764095, -0.0976528599858284, -0.05715922266244888, 0.129258394241333, 0.014699012041091919, -0.040573831647634506, -0.0746607705950737, -0.061572592705488205, -0.05094948783516884, -0.02416388876736164, -0.06468550115823746, 0.0019121807999908924, 0.01940496265888214, 0.010974458418786526, 0.02200266532599926, -0.07483047246932983, 0.012254645116627216, -0.07702239602804184, 0.12862791121006012, 0.02505405992269516, 0.02454131282866001, -0.062446411699056625, 0.04934510588645935, 0.06413210928440094, -0.06960836797952652, -0.01682206243276596, -0.04266781359910965, -0.061853084713220596, -0.02066291682422161, -0.03333088755607605, -0.034082602709531784, 0.05662687495350838, 0.14415858685970306, -0.02170671708881855, 0.07065407186746597, 0.027812577784061432, 0.030632847920060158, 0.10965896397829056, 0.044310636818408966, -0.048656165599823, 0.0044324966147542, 0.008197339251637459, -0.009480931796133518, 0.013390511274337769, -0.05269492790102959, -0.007790293078869581, -0.015201417729258537, 0.013314757496118546, 0.04704577475786209, -0.004116971045732498, 0.08292073011398315, -0.025725120678544044, -0.03830545023083687, -0.015553062781691551, -0.06780509650707245, -0.024362318217754364, -0.01017299760133028, -0.045962173491716385, 0.044480934739112854, -0.022249383851885796, 0.03628721460700035, -0.041607797145843506, 0.02971206046640873, -0.05400698259472847, -0.04291490837931633, -0.01743941567838192, -0.04341479018330574, 0.014844524674117565, 0.019331619143486023, -0.013369258493185043, -0.10050000250339508, -0.0275319442152977, -0.0631236582994461, 0.0361432321369648, -0.008243951015174389, -0.03468272089958191, -0.03337184712290764, -0.004853604361414909, 0.03623924031853676, -0.013228852301836014, 0.017694581300020218, -0.02328353375196457, -0.007299797143787146, -0.05048978328704834, 0.041289251297712326, 0.039684273302555084, 0.012129426933825016, -0.06517500430345535, 0.03493209183216095, -0.10470084100961685, 0.06038679927587509, -0.034605901688337326, 0.011171837337315083, -0.04688303917646408, 0.006506146863102913, -0.10414145886898041, 0.02156628482043743, 0.03550907224416733, 0.08248145133256912, -0.11510046571493149, -0.02116996981203556, 0.08496006578207016, -0.0704948678612709, -0.016652239486575127, 0.07267890125513077, -0.04244439676403999, 0.03365195542573929, 0.0520472526550293, 0.10144127160310745, 0.0364077165722847, -0.07249712944030762, -0.04376479610800743, -0.03359537571668625, -0.11445533484220505, 0.1317196488380432, 0.004903785884380341, 0.05719743296504021, -0.020105039700865746, 0.028161846101284027, 0.038970377296209335, 0.0016786366468295455, -0.020268304273486137, -0.016748936846852303, 0.010598194785416126, 0.008825154975056648, 0.03897795453667641, -0.004049984738230705, 0.009379247203469276, -0.03301647678017616, -0.038842763751745224, -0.02460019290447235, 0.06215157359838486, -0.009951408952474594, 0.014818361029028893, -0.10911489278078079, 0.030786115676164627, 0.04854840040206909, 0.011066404171288013, -0.0776808112859726, -0.03891841322183609, 0.015285971574485302, 0.046230509877204895, 0.06651005148887634, 0.14663006365299225, 0.03156188130378723, -0.0040550753474235535, 0.024073997512459755, 0.01526612788438797, 0.04703221470117569, -0.0008210610249079764, -0.051201872527599335, -0.0800662562251091, 0.025849957019090652, -0.02120007760822773, 0.1202758178114891, -0.01915845461189747, 0.015119514428079128, 0.0034122911747545004, 0.0719858705997467, -0.02784314937889576, -0.019435249269008636, -0.009148471057415009, -0.03200516104698181, -0.02972361259162426, -0.0064952438697218895, 0.0003529466048348695, 0.01105279941111803, 0.011788385920226574, 0.08269122987985611, -0.12778595089912415, -0.02715970017015934, 0.0808052122592926, -0.02999757044017315, -0.04278668016195297, -0.01604500412940979, -0.044165436178445816, 0.004537162836641073, 0.014619724825024605, -0.012442626990377903, 0.18181045353412628, 0.020752953365445137, 0.03133740648627281, -0.0358421765267849, -0.021951403468847275, 0.033775344491004944, -0.0457553006708622, -0.030088668689131737, 0.015320241451263428, 0.01961081475019455, -0.13256151974201202, 0.006538171321153641, 0.04979998990893364, 0.07312661409378052, 0.0820108950138092, 0.01841452345252037, -0.017550278455018997, -0.0761827751994133, 0.02489163726568222, -0.02120729349553585, 0.015521559864282608, -0.09533128142356873, 0.0038042436353862286, 0.00857713632285595, 0.005935258697718382, 0.03900466486811638, -0.03854886814951897, 0.008432069793343544, 0.003065949073061347, 0.014227217994630337, 0.03179868310689926, 0.015381786972284317, -0.00962811429053545, 0.07595202326774597, 0.022640708833932877, -0.02609163150191307, -0.04388698562979698, -0.01852058432996273, -0.044769544154405594, 0.08239573240280151, -0.04005729779601097, -0.16269192099571228, -0.011116699315607548, -0.01973249763250351, -0.02349746599793434, 0.03164439648389816, 0.011038959957659245, -0.036644719541072845, -0.05883105471730232, -0.03452543914318085, 0.04422139748930931, 0.012295406311750412, -0.0068689449690282345, 0.06052656099200249, 0.0008119246340356767, -0.008733801543712616, -0.08126052469015121, -0.025946632027626038, -0.06278035789728165, -0.039728835225105286, 0.029761405661702156, -0.01784503273665905, 0.0242962334305048, 0.11974118649959564, 0.0002146997139789164, -0.006097236182540655, -0.010703199543058872, 0.1868121176958084, -0.04378192499279976, 0.04584261402487755, 0.05999067425727844, 0.011307235807180405, 0.036940962076187134, 0.10786334425210953, -0.011956749483942986, -0.07381849735975266, 0.038205571472644806, 0.05267602950334549, -0.03398189693689346, -0.12991443276405334, -0.05808165669441223, -0.034621696919202805, 0.0055322907865047455, 0.050250012427568436, 0.01709410361945629, 0.03467368707060814, 0.03801335394382477, -0.07169745117425919, 0.0002018675731960684, 0.061686672270298004, 0.03370232507586479, 0.0759994313120842, -0.033875297755002975, 0.08519026637077332, -0.01316598616540432, -0.06526409089565277, 0.03969626501202583, -0.028610318899154663, 0.041240036487579346, 0.027766797691583633, 0.03467660769820213, 0.07682839781045914, 0.06146116182208061, 0.022988321259617805, 0.023336853832006454, 0.016437510028481483, -0.01606646738946438, -0.028776798397302628, -0.02734166942536831, -0.07560883462429047, 0.035685066133737564, 0.05395139381289482, -0.05995212122797966, -0.03136284649372101, 0.07164111733436584, 0.06043313816189766, 0.11616219580173492, 0.03964345529675484, -0.09895911812782288, -0.05108960345387459, 0.033830318599939346, -0.031689487397670746, -0.04915524646639824, 0.0392935685813427, 0.06586967408657074, -0.051470737904310226, 0.025118950754404068, 0.011519859544932842, 0.0487586110830307, -0.028078671544790268, 0.06249533221125603, -0.1034635603427887, 0.02523486502468586, -0.010944994166493416, 0.04089692607522011, -0.14056576788425446, 0.03472920134663582, 0.001454782672226429, 0.023348884657025337, -0.04907332360744476, 0.004983227699995041, 0.008451968431472778, -0.0147269107401371, 0.0633854866027832, 0.01772063598036766, -0.14332076907157898, -0.0587550550699234, -0.06684402376413345, 0.015732156112790108, 0.07314061373472214, -0.029600514099001884, 0.05300445482134819, -0.013823638670146465, 0.02029239945113659, 0.011199251748621464, 0.057311899960041046, -0.07011816650629044, -0.07543136924505234, 0.07067275792360306, 0.031102379783988, -0.04319214075803757, -0.004821081645786762, -0.012475108727812767, -0.08824753016233444, 0.13590757548809052, -0.031021827831864357, -0.009298579767346382, -0.03874863684177399, 0.008562912233173847, 0.06656010448932648, -0.03486857935786247, -0.011326299048960209, -0.0002671623951755464, 0.09494557976722717, -0.03527349978685379, -0.06558626145124435, 0.026804029941558838, -0.028241660445928574, -0.07364659011363983, -0.03267737850546837, 0.0745401382446289, 0.03859426826238632, 0.0218012984842062, 0.025073563680052757, 0.015871291980147362, 0.0018106605857610703, -0.08186796307563782, 0.029661085456609726, 0.010832416824996471, -0.02570623718202114, -0.004928308073431253, 0.011396754533052444, -0.05393439903855324, -0.07980320602655411, 0.03250248730182648, 0.0985075980424881, 0.11886674910783768, -0.04394575208425522, 0.07591887563467026, 0.08803358674049377, -0.039136629551649094, -0.1586003303527832, -0.04525246471166611, 0.03442757576704025, 0.025771228596568108, -0.02997291274368763, -0.07928507775068283, 0.024734996259212494, 0.03758348524570465, -0.0020469308365136385, -0.01552306953817606, -0.18626806139945984, -0.10645291209220886, 0.06272812932729721, -0.0185391865670681, 0.19388219714164734, -0.021482069045305252, -0.00799885019659996, -0.03260542079806328, -0.03692774847149849, 0.10201599448919296, -0.02879004366695881, 0.05850924551486969, 0.02673533745110035, 0.025098884478211403, 0.05624031275510788, -0.010773122310638428, 0.07662172615528107, 0.019306868314743042, 0.06168890744447708, -0.07450596243143082, -0.10760705173015594, 0.05811142921447754, -0.0312652587890625, 0.046913355588912964, 0.04291350394487381, 0.007235704455524683, -0.0017121433047577739, -0.018544889986515045, -0.04253324121236801, 0.04731309413909912, -0.02327399142086506, -0.029942099004983902, -0.06894729286432266, 0.02731315791606903, 0.05481354519724846, 0.020977087318897247, 0.046740081161260605, -0.03006710112094879, 0.012350489385426044, 0.11088652908802032, 0.05123811587691307, -0.05970609188079834, 0.0011851884191855788, 0.02738761156797409, 0.01082860492169857, 0.035875409841537476, -0.06127075478434563, 0.04251161590218544, 0.06752876192331314, -0.004546315874904394, 0.04885697737336159, 0.025928283110260963, -0.06216162070631981, 0.005061776842921972, 0.05406273528933525, -0.12276724725961685, -0.08353706449270248, 0.007127954158931971, 0.04173816367983818, 0.008652138523757458, 0.010867031291127205, 0.11401620507240295, -0.00879010558128357, -0.003918027505278587, 0.01799941435456276, 0.03582366183400154, 0.011240077205002308, 0.05985128507018089, -0.01121989730745554, 0.017028139904141426, -0.05945782735943794, 0.05222565680742264, 0.0718197301030159, -0.03941629081964493, 0.05258210003376007, -0.03443954139947891, -0.03942243009805679, -0.046350330114364624, -0.0672529861330986, 0.02289443463087082, 0.024335632100701332, 0.008300893940031528, -0.02160612680017948, -0.07208958268165588, 0.006260564550757408, 0.06268518418073654, 0.0007992147584445775, 0.07174519449472427, -0.012309090234339237, 0.020186305046081543, -0.03682010993361473, 0.013250153511762619, 0.027391454204916954, 0.007398150861263275, -0.033535800874233246, 0.10849788784980774, -0.007351911161094904, 0.025599893182516098, -0.04053715988993645, -0.0703633576631546, -0.05938540771603584, -0.0051988111808896065, -0.04478231072425842, 0.021985052153468132, -0.06887511909008026, -0.019260840490460396, 0.026155108585953712, -0.020403215661644936, -0.05280927196145058, 0.0029530671890825033, -0.021626397967338562, 0.02783169038593769, -0.011152097024023533, 0.046863507479429245, -0.04908840358257294, 0.0008154444512911141, 0.048271119594573975, -0.04701065272092819, 0.05407030135393143, 0.06846825033426285, -0.018303679302334785, 0.019583139568567276, -0.07435017079114914, -0.004437097813934088, 0.012955829501152039, 0.04359269514679909, -0.0010988132562488317, -0.04066750034689903, 0.005823375191539526, -0.0076653724536299706, 0.0010529209394007921, 0.011457614600658417, 0.022837962955236435, -0.04990185424685478, 0.042024802416563034, 0.04313287511467934, -0.0726548358798027, -0.05281401053071022, -0.024338707327842712, 0.024574287235736847, 0.020767489448189735, 0.08743150532245636, -0.014743486419320107, 0.018837496638298035, -0.10084731876850128, 0.0044113751500844955, 0.03594117611646652, -0.028510116040706635, 0.041552089154720306, -0.047309987246990204, 0.03199925273656845, -0.021744728088378906, 0.08232530206441879, 0.00867657084017992, -0.022904807701706886, 0.017863836139440536, 0.014778897166252136, -0.06252442300319672, 0.04824052378535271, -0.02086699940264225, 0.03750130906701088, -0.015010817907750607, -0.04061141237616539, -0.021077264100313187, 0.025044437497854233, -0.050627581775188446, 0.09108524769544601, 0.04683965817093849, 0.06906968355178833, 0.03106461651623249, 0.02322789840400219, -0.01827666535973549, -0.006753277499228716, 0.08342841267585754, -0.12941811978816986, 0.013873301446437836, -0.015569260343909264, -0.008806967176496983, 0.0931580439209938, -0.07307462394237518, 0.04252231866121292, -0.056384533643722534, -0.03187769278883934, -0.06989193707704544, -0.13756319880485535, -0.024747135117650032, -0.010224811732769012, 0.018455706536769867, -0.07797437906265259, 0.023534530773758888, 0.007070738822221756, 0.025822758674621582, -0.02521449513733387, 0.050202205777168274, -0.032660551369190216, -0.05509715527296066, 0.025856420397758484, -0.00010795673733809963, 0.007653338368982077, -0.030120903626084328, -0.015305474400520325, -0.0013849325478076935, 0.06059744954109192, 0.029034536331892014, 0.0315399095416069, 0.08421378582715988, -0.00780521659180522, -0.02394610270857811, -0.05653165280818939, -0.006392257753759623, -0.009968343190848827, -0.00947343185544014, -0.014314836822450161, 0.04620186984539032, 0.0067689139395952225, -0.015333825722336769, 0.12585163116455078, 0.015961647033691406, -0.05692511051893234, -0.09498655796051025, 0.08121301978826523, 0.02336703985929489, 0.05886277183890343, -0.00769630866125226, -0.01270014327019453, -0.020873360335826874, 0.13287101686000824, 0.11205470561981201, -0.03783431649208069, 0.006865696981549263, -0.008110595867037773, 0.020559553056955338, 0.016844473779201508, 0.0916619598865509, 0.019516756758093834, 0.15047411620616913, -0.021239541471004486, 0.029429320245981216, -0.01207707915455103, 0.010230634361505508, -0.03182937949895859, 0.05954582244157791, -0.038058776408433914, 0.020563911646604538, -0.047062475234270096, 0.04148440062999725, -0.061604224145412445, -0.10225313156843185, -0.03956693038344383, -0.08359541743993759, -0.04830683767795563, 0.020758740603923798, 0.03759283199906349, 0.013169920071959496, 0.11381040513515472, -0.007170981727540493, -0.02403077483177185, 0.026275308802723885, 0.009271240793168545, -0.0363163948059082, 0.03070903941988945, 0.0641961619257927, -0.018394529819488525, 0.09136151522397995, -0.00389591371640563, 0.0794638991355896, 0.06475373357534409, 0.028787115588784218, -0.047330521047115326, 0.03656868636608124, 0.024885259568691254, -0.04494915157556534, -0.0027280766516923904, 0.055232953280210495, 0.0067724972032010555, 0.04438326880335808, 0.09392178058624268, -0.06080136448144913, 0.05350147560238838, 0.0011823229724541306, 0.0195608027279377, -0.03712907060980797, 0.07796386629343033, -0.021338827908039093, 0.06886870414018631, 0.07708851993083954, -0.0027901146095246077, -0.012668052688241005, -0.0006506506470032036, -0.012530581094324589, -0.04482082277536392, 0.03700588643550873, -0.056553661823272705, -0.1388288140296936, 0.023431679233908653, -0.01016477681696415, 0.041372839361429214, -0.07591769844293594, -0.03275051712989807, 0.04186839237809181, -0.011629143729805946, 0.02781485766172409, 0.08072856068611145, 0.03918362036347389, 0.016992002725601196, -0.018031146377325058, -0.020561853423714638, 0.015685902908444405, 0.10523135215044022, -0.03408075124025345, -0.06258353590965271 ]
null
null
keras
# [Deep Chimpact](https://www.drivendata.org/competitions/82/competition-wildlife-video-depth-estimation/page/390/) > Depth Estimation for Wildlife Conservation (1st place solution) <div align=center> <img src="https://user-images.githubusercontent.com/36858976/138281204-c3cbcb77-11ca-448b-a693-cb3cfa3c5181.png" width=800> ## Overview Healthy natural ecosystems have wide-ranging benefits from public health to the economy to agriculture. In order to protect the Earth's natural resources, conservationists need to be able to monitor species population sizes and population change. Camera traps are widely used in conservation research to capture images and videos of wildlife without human interference. Using statistical models for distance sampling, the frequency of animal sightings can be combined with the distance of each animal from the camera to estimate a species' full population size. However, getting distances from camera trap footage currently entails an extremely manual, time-intensive process. It takes a researcher more than **10 minutes** on average to label distance for every **1 minute** of video - thatโ€™s a lot of time when you have a million videos! This also creates a bottleneck for critical information that conservationists can use to **monitor wildlife populations**. > Your goal in this challenge is to use machine learning to automatically estimate the distance between a camera trap and an animal in a series of camera trap videos. You will be given a series of timestamps indicating when animals are visible in each camera trap video. To complete the challenge, you will predict the distance between the animal and the camera at each point in time. Along the way, keep an eye out for some sneaky leopards hunting at night, baby chimpanzees getting piggy-back rides, and diva elephants that can't get enough of the limelight. By contributing to this challenge, you can help advance cutting-edge methods for keeping these animal populations (and humans) healthy and safe!
{}
null
awsaf49/deep-chimpact
[ "keras", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #keras #region-us
# Deep Chimpact > Depth Estimation for Wildlife Conservation (1st place solution) <div align=center> <img src="URL width=800> ## Overview Healthy natural ecosystems have wide-ranging benefits from public health to the economy to agriculture. In order to protect the Earth's natural resources, conservationists need to be able to monitor species population sizes and population change. Camera traps are widely used in conservation research to capture images and videos of wildlife without human interference. Using statistical models for distance sampling, the frequency of animal sightings can be combined with the distance of each animal from the camera to estimate a species' full population size. However, getting distances from camera trap footage currently entails an extremely manual, time-intensive process. It takes a researcher more than 10 minutes on average to label distance for every 1 minute of video - thatโ€™s a lot of time when you have a million videos! This also creates a bottleneck for critical information that conservationists can use to monitor wildlife populations. > Your goal in this challenge is to use machine learning to automatically estimate the distance between a camera trap and an animal in a series of camera trap videos. You will be given a series of timestamps indicating when animals are visible in each camera trap video. To complete the challenge, you will predict the distance between the animal and the camera at each point in time. Along the way, keep an eye out for some sneaky leopards hunting at night, baby chimpanzees getting piggy-back rides, and diva elephants that can't get enough of the limelight. By contributing to this challenge, you can help advance cutting-edge methods for keeping these animal populations (and humans) healthy and safe!
[ "# Deep Chimpact\n> Depth Estimation for Wildlife Conservation (1st place solution)\n\n<div align=center> <img src=\"URL width=800>", "## Overview\n\nHealthy natural ecosystems have wide-ranging benefits from public health to the economy to agriculture. In order to protect the Earth's natural resources, conservationists need to be able to monitor species population sizes and population change. Camera traps are widely used in conservation research to capture images and videos of wildlife without human interference. Using statistical models for distance sampling, the frequency of animal sightings can be combined with the distance of each animal from the camera to estimate a species' full population size.\n\nHowever, getting distances from camera trap footage currently entails an extremely manual, time-intensive process. It takes a researcher more than 10 minutes on average to label distance for every 1 minute of video - thatโ€™s a lot of time when you have a million videos! This also creates a bottleneck for critical information that conservationists can use to monitor wildlife populations.\n\n> Your goal in this challenge is to use machine learning to automatically estimate the distance between a camera trap and an animal in a series of camera trap videos. You will be given a series of timestamps indicating when animals are visible in each camera trap video. To complete the challenge, you will predict the distance between the animal and the camera at each point in time.\n\nAlong the way, keep an eye out for some sneaky leopards hunting at night, baby chimpanzees getting piggy-back rides, and diva elephants that can't get enough of the limelight. By contributing to this challenge, you can help advance cutting-edge methods for keeping these animal populations (and humans) healthy and safe!" ]
[ "TAGS\n#keras #region-us \n", "# Deep Chimpact\n> Depth Estimation for Wildlife Conservation (1st place solution)\n\n<div align=center> <img src=\"URL width=800>", "## Overview\n\nHealthy natural ecosystems have wide-ranging benefits from public health to the economy to agriculture. In order to protect the Earth's natural resources, conservationists need to be able to monitor species population sizes and population change. Camera traps are widely used in conservation research to capture images and videos of wildlife without human interference. Using statistical models for distance sampling, the frequency of animal sightings can be combined with the distance of each animal from the camera to estimate a species' full population size.\n\nHowever, getting distances from camera trap footage currently entails an extremely manual, time-intensive process. It takes a researcher more than 10 minutes on average to label distance for every 1 minute of video - thatโ€™s a lot of time when you have a million videos! This also creates a bottleneck for critical information that conservationists can use to monitor wildlife populations.\n\n> Your goal in this challenge is to use machine learning to automatically estimate the distance between a camera trap and an animal in a series of camera trap videos. You will be given a series of timestamps indicating when animals are visible in each camera trap video. To complete the challenge, you will predict the distance between the animal and the camera at each point in time.\n\nAlong the way, keep an eye out for some sneaky leopards hunting at night, baby chimpanzees getting piggy-back rides, and diva elephants that can't get enough of the limelight. By contributing to this challenge, you can help advance cutting-edge methods for keeping these animal populations (and humans) healthy and safe!" ]
[ 9, 40, 360 ]
[ "passage: TAGS\n#keras #region-us \n# Deep Chimpact\n> Depth Estimation for Wildlife Conservation (1st place solution)\n\n<div align=center> <img src=\"URL width=800>## Overview\n\nHealthy natural ecosystems have wide-ranging benefits from public health to the economy to agriculture. In order to protect the Earth's natural resources, conservationists need to be able to monitor species population sizes and population change. Camera traps are widely used in conservation research to capture images and videos of wildlife without human interference. Using statistical models for distance sampling, the frequency of animal sightings can be combined with the distance of each animal from the camera to estimate a species' full population size.\n\nHowever, getting distances from camera trap footage currently entails an extremely manual, time-intensive process. It takes a researcher more than 10 minutes on average to label distance for every 1 minute of video - thatโ€™s a lot of time when you have a million videos! This also creates a bottleneck for critical information that conservationists can use to monitor wildlife populations.\n\n> Your goal in this challenge is to use machine learning to automatically estimate the distance between a camera trap and an animal in a series of camera trap videos. You will be given a series of timestamps indicating when animals are visible in each camera trap video. To complete the challenge, you will predict the distance between the animal and the camera at each point in time.\n\nAlong the way, keep an eye out for some sneaky leopards hunting at night, baby chimpanzees getting piggy-back rides, and diva elephants that can't get enough of the limelight. By contributing to this challenge, you can help advance cutting-edge methods for keeping these animal populations (and humans) healthy and safe!" ]
[ -0.016787000000476837, 0.06410630792379379, -0.00021983768965583295, 0.015950320288538933, 0.061920613050460815, -0.08633604645729065, -0.12440649420022964, 0.0878363698720932, 0.000058415862440597266, 0.07305929064750671, -0.006569799035787582, -0.0006446929182857275, 0.1260526180267334, 0.10245458036661148, 0.0798720195889473, -0.24286556243896484, 0.0632309541106224, -0.02801370434463024, 0.038139380514621735, 0.07311864197254181, 0.059961218386888504, -0.13233241438865662, 0.09182637184858322, 0.022851303219795227, -0.12588991224765778, -0.05699533224105835, -0.08954137563705444, -0.015820061787962914, 0.11693160235881805, -0.10445405542850494, -0.07274587452411652, -0.062149468809366226, 0.11709266901016235, -0.04429467022418976, -0.014020344242453575, 0.09782461822032928, 0.019928807392716408, -0.027295416221022606, 0.14468571543693542, 0.12884481251239777, 0.021631529554724693, -0.15177759528160095, 0.05714630335569382, 0.08377884328365326, -0.06696340441703796, -0.03387840837240219, -0.1004193052649498, -0.2259019911289215, 0.15921370685100555, 0.11557422578334808, -0.037173256278038025, 0.10044649988412857, -0.06181979551911354, 0.015065085142850876, 0.09885343909263611, -0.1832817792892456, 0.00381169724278152, 0.03574100136756897, 0.042740218341350555, 0.13338790833950043, -0.1166243925690651, 0.06439194828271866, 0.0438704639673233, -0.024539697915315628, 0.08573917299509048, -0.027987617999315262, 0.02590114064514637, -0.13873329758644104, -0.17598770558834076, -0.06746688485145569, 0.023033063858747482, 0.07954126596450806, -0.11751259863376617, -0.0833716168999672, -0.04567825794219971, 0.011490559205412865, -0.04051362723112106, 0.05856529250741005, 0.10232733935117722, -0.023047056049108505, 0.033335261046886444, 0.048031218349933624, -0.0697004422545433, -0.025894302874803543, -0.054232336580753326, -0.11977705359458923, 0.048226237297058105, 0.04207610711455345, -0.08219721913337708, 0.10050584375858307, -0.0014835979091003537, -0.008339617401361465, -0.07524382323026657, 0.04097340628504753, -0.1623433381319046, -0.03911761939525604, -0.0418570339679718, -0.08320001512765884, -0.013879218138754368, 0.06300926953554153, -0.01236521452665329, 0.06625001132488251, -0.09923280775547028, 0.01835460215806961, 0.11057266592979431, 0.11045731604099274, -0.07418099790811539, -0.12116790562868118, 0.026750870048999786, -0.07980991899967194, 0.02025592513382435, -0.06571384519338608, -0.007167643401771784, -0.05079343914985657, 0.016676582396030426, 0.03378013148903847, 0.07862892001867294, -0.013922829180955887, 0.02984921634197235, -0.03877773508429527, -0.007042186800390482, -0.017226118594408035, -0.007488158531486988, -0.07362031936645508, -0.05513329058885574, 0.11606165021657944, -0.052637819200754166, 0.04194556176662445, 0.007870581932365894, 0.034075777977705, -0.07056955993175507, -0.03282373398542404, -0.10910636931657791, -0.03525210916996002, 0.020006345584988594, -0.04694199189543724, -0.057067982852458954, -0.1478998064994812, -0.11880836635828018, 0.006078771315515041, -0.048574112355709076, -0.03726096451282501, -0.03380855545401573, 0.00859332736581564, 0.042993638664484024, 0.004034509416669607, 0.03203756362199783, 0.13586074113845825, -0.011408830992877483, -0.01971196010708809, -0.11036358028650284, 0.14511190354824066, 0.032932378351688385, 0.03792773187160492, -0.07128234207630157, 0.06519435346126556, -0.08431485295295715, 0.14201687276363373, -0.0031606443226337433, 0.09477735310792923, -0.03316042199730873, 0.02697714790701866, -0.032275933772325516, 0.012067122384905815, -0.0466330386698246, 0.12631362676620483, -0.20199371874332428, 0.028766367584466934, 0.07763069123029709, -0.07378660142421722, -0.04138197377324104, 0.13625143468379974, -0.023746702820062637, 0.06207694485783577, 0.10136085748672485, 0.16483226418495178, 0.11063241958618164, -0.07433899492025375, 0.001229805639013648, -0.032678063958883286, -0.23419329524040222, 0.006410798989236355, 0.060236021876335144, 0.03329798951745033, 0.028472790494561195, -0.026872962713241577, 0.04164975881576538, 0.053010474890470505, 0.007060710806399584, 0.0031366641633212566, 0.02628196030855179, -0.03674560412764549, -0.047164954245090485, -0.04589381068944931, 0.028943199664354324, -0.05188843235373497, 0.04921535402536392, -0.015426223166286945, 0.10499580949544907, 0.03807569295167923, 0.08976544439792633, 0.028108803555369377, 0.05469069629907608, -0.08692111819982529, -0.03141780197620392, -0.11579703539609909, 0.08991201221942902, 0.06369023025035858, -0.15991216897964478, 0.11176085472106934, -0.027194341644644737, 0.04343998432159424, 0.04842653125524521, 0.03302386775612831, 0.020081473514437675, 0.00831533782184124, 0.01250525750219822, -0.09344378858804703, 0.0189199335873127, 0.08879205584526062, -0.04273190721869469, 0.1647806167602539, 0.06766985356807709, -0.047532662749290466, 0.08896318078041077, 0.05872343108057976, 0.04845121502876282, -0.08706863969564438, 0.08122900873422623, -0.02290121465921402, -0.021060720086097717, -0.046325813978910446, 0.002490315353497863, -0.0004207350721117109, 0.009842507541179657, 0.0801641121506691, -0.09192211180925369, -0.26467016339302063, 0.09448938071727753, -0.17495425045490265, -0.09403398633003235, 0.049929503351449966, 0.07227787375450134, -0.03487277030944824, -0.09323123097419739, -0.11031977087259293, 0.11001361906528473, -0.003833709517493844, 0.05512656643986702, -0.06595270335674286, -0.009452343918383121, 0.046186912804841995, 0.005658657290041447, 0.023648612201213837, -0.004993143957108259, 0.05910209193825722, -0.2224418669939041, 0.06018451601266861, 0.15433920919895172, 0.13055069744586945, 0.0561538003385067, 0.049489833414554596, -0.15230919420719147, -0.013709429651498795, 0.060026396065950394, 0.024802198633551598, 0.019324984401464462, 0.033898063004016876, 0.025814613327383995, 0.010585376992821693, 0.05825011432170868, -0.008546426892280579, -0.06903094798326492, 0.09676065295934677, 0.11267821490764618, 0.07038040459156036, -0.06788495928049088, -0.0614456981420517, -0.013699710369110107, 0.07291893661022186, 0.1087154820561409, -0.013163433410227299, 0.01680910587310791, -0.01134771853685379, -0.07693859189748764, 0.13572005927562714, -0.08412009477615356, -0.3705933392047882, -0.04848161339759827, -0.03259433060884476, 0.019854409620165825, -0.03705887496471405, 0.015575582161545753, -0.07555048167705536, -0.004751719068735838, -0.026929814368486404, -0.07912399619817734, -0.09760083258152008, 0.05158419907093048, 0.006514710374176502, 0.034525878727436066, -0.017988257110118866, -0.05642491206526756, 0.007525134831666946, -0.03416397422552109, -0.10271254181861877, 0.05301661789417267, 0.04048118367791176, 0.12028951197862625, 0.05417856201529503, -0.04337228834629059, -0.042793650180101395, 0.01135313231498003, 0.12125340849161148, -0.10429571568965912, 0.10991068929433823, 0.1276099979877472, 0.015454511158168316, 0.07701979577541351, 0.044813644140958786, -0.02714947611093521, -0.05467090383172035, 0.007932428270578384, 0.08575838059186935, -0.10548560321331024, -0.018494706600904465, -0.047835271805524826, -0.07596737146377563, 0.03568216785788536, 0.07142973691225052, -0.013192580081522465, 0.042252689599990845, 0.034713730216026306, -0.011967688798904419, -0.10013847798109055, -0.007423069793730974, 0.1290549784898758, -0.09525365382432938, -0.03306582197546959, 0.04046553373336792, -0.06213845685124397, 0.022271418944001198, 0.08751724660396576, 0.007580769248306751, 0.2109326422214508, 0.017422901466488838, 0.04853098466992378, 0.019216269254684448, -0.017949292436242104, 0.052059195935726166, 0.05076434463262558, -0.14038632810115814, 0.0050984034314751625, -0.05946355685591698, 0.004043989814817905, 0.06583233922719955, 0.08417393267154694, 0.12038430571556091, -0.022770456969738007, -0.04462987557053566, 0.0238784272223711, 0.0610896497964859, 0.11654313653707504, 0.1041286289691925, -0.04594089835882187, 0.08154125511646271, 0.011576435528695583, -0.022824415937066078, -0.08254215121269226, 0.07954970002174377, 0.043608881533145905, -0.14182335138320923, 0.0193554125726223, -0.0008303720387630165, 0.07106439024209976, -0.1266564130783081, -0.023678984493017197, -0.07133230566978455, 0.02646244876086712, -0.05053500086069107, 0.0340249128639698, -0.030480481684207916, 0.0019492936553433537, -0.02175770327448845, -0.03657595068216324, -0.0245286263525486, -0.007388778496533632, 0.04428986832499504, -0.061753444373607635, 0.11217771470546722, 0.050702016800642014, -0.19392642378807068, -0.056972529739141464, -0.10042940825223923, -0.0807766541838646, 0.07043222337961197, -0.02608722262084484, 0.1172153651714325, 0.02532292529940605, -0.013305265456438065, -0.036182891577482224, -0.10088203847408295, -0.02583516389131546, -0.11883997172117233, 0.011950572021305561, -0.018071593716740608, -0.012326141819357872, 0.0028539381455630064, 0.047108493745326996, -0.08628764003515244, -0.05792137607932091, -0.10690440237522125, -0.07843940705060959, -0.05128386989235878, -0.09273364394903183, 0.005987364333122969, -0.019105009734630585, 0.19794054329395294, -0.0047017657198011875, 0.16047941148281097, -0.03261277824640274, -0.09354978054761887, 0.04365833103656769, -0.04863203689455986, -0.058452632278203964, -0.06945929676294327, 0.09274206310510635, 0.048090510070323944, 0.0850176066160202, -0.006901112385094166, 0.09048565477132797, -0.0033096112310886383, -0.044803038239479065, 0.05132294073700905, -0.03286341577768326, -0.040452294051647186, 0.03436210751533508, -0.009095126762986183, -0.048374176025390625, -0.09837305545806885, -0.023892052471637726, 0.23808450996875763, 0.16124194860458374, -0.11469195038080215, 0.17228740453720093, 0.14560794830322266, -0.0072222063317894936, -0.18985219299793243, 0.1186058521270752, 0.01550318580120802, -0.07289654016494751, 0.05836019665002823, -0.11729107797145844, 0.05857963487505913, 0.10264275223016739, 0.023070231080055237, 0.0880012959241867, -0.29691827297210693, -0.08480674773454666, -0.07748598605394363, 0.010936329141259193, 0.012033218517899513, -0.02840098738670349, 0.0666307583451271, -0.008053060621023178, -0.009751096367835999, 0.08238551765680313, -0.04529985412955284, 0.08385231345891953, 0.004437808878719807, 0.08734278380870819, 0.03198440372943878, -0.03164196386933327, 0.1054089143872261, -0.026283835992217064, 0.04504373297095299, -0.06337551772594452, -0.14914491772651672, 0.03086957521736622, -0.09110372513532639, 0.02543746493756771, 0.026526862755417824, 0.004524235147982836, -0.13095977902412415, 0.017254270613193512, -0.023813337087631226, -0.057146403938531876, -0.015612982213497162, -0.0769643485546112, -0.15838481485843658, 0.1309143304824829, 0.11606007069349289, -0.03765544295310974, -0.031535375863313675, -0.027732351794838905, -0.2247806191444397, 0.22376756370067596, -0.051146890968084335, 0.046835288405418396, -0.18544168770313263, -0.11183110624551773, -0.00957189965993166, 0.0702032744884491, -0.07305286079645157, 0.012579609639942646, 0.055562134832143784, 0.044598594307899475, -0.0003255150804761797, 0.0169784277677536, -0.10881517827510834, 0.14448203146457672, 0.013075540773570538, 0.052844610065221786, -0.15115104615688324, 0.05234783887863159, 0.16460785269737244, -0.0007904216763563454, -0.04354620352387428, 0.023969804868102074, 0.012263532727956772, -0.0399601086974144, -0.022884920239448547, 0.09808702766895294, 0.007303181570023298, 0.01831330917775631, -0.07022475451231003, 0.0038425244856625795, -0.09716441482305527, 0.16561409831047058, 0.028933046385645866, -0.00572759797796607, -0.027510104700922966, 0.11217143386602402, -0.10282476991415024, 0.016472479328513145, 0.06978795677423477, -0.02246871404349804, -0.12471403926610947, 0.03709571808576584, 0.1031922996044159, -0.1313416063785553, 0.024595964699983597, 0.1308576613664627, -0.008137119933962822, -0.04525119811296463, 0.016969645395874977, 0.00366624491289258, -0.12618182599544525, 0.019149042665958405, -0.008476453833281994, 0.0369897335767746, -0.15622343122959137, 0.06446155905723572, 0.04250466451048851, -0.036986093968153, -0.01359221339225769, -0.14203545451164246, -0.048760391771793365, 0.01552245207130909, -0.08303084969520569, -0.018197577446699142, -0.024121178314089775, 0.02014831453561783, -0.008146997541189194, 0.06254911422729492, 0.018102630972862244, 0.029006684198975563, -0.028682375326752663, -0.044761769473552704, -0.02020031027495861, 0.017225094139575958, -0.00021865345479454845, -0.1012645959854126, 0.05537166818976402, -0.03334648162126541, -0.011222507804632187, 0.03914036974310875, 0.029143109917640686, -0.06117965281009674, 0.039441756904125214, -0.11907406151294708, 0.030364276841282845, 0.023441892117261887, 0.01586906984448433, -0.12494999170303345, 0.006922118365764618, -0.07618183642625809, -0.12980914115905762, -0.03842480108141899, 0.07716212421655655, -0.0792304202914238, 0.11455279588699341, -0.01459332462400198, -0.09607851505279541, -0.0768795907497406, -0.02525140717625618, 0.055169302970170975, 0.05845875293016434, 0.07286433130502701, -0.0017919137608259916, 0.0732431635260582, -0.0671989694237709, -0.03672154247760773, -0.016999928280711174, -0.022643791511654854, -0.05593247711658478, -0.057852037250995636, 0.033476971089839935, -0.005335123743861914, 0.017040690407156944, -0.058691106736660004, -0.03212999552488327, 0.007522877771407366, 0.019202690571546555, 0.01129789836704731, 0.023415565490722656, 0.05000279098749161, -0.033132344484329224, -0.0018273730529472232, -0.031086737290024757, 0.04931800067424774, 0.08607069402933121, -0.04795338958501816, 0.035315439105033875, 0.10545838624238968, 0.04630962759256363, -0.01110150758177042, -0.03013004921376705, 0.055833254009485245, -0.19457392394542694, 0.00472917128354311, -0.03681681677699089, -0.04270489141345024, -0.01323341391980648, 0.061020754277706146, 0.15735332667827606, -0.11214794963598251, 0.10316956788301468, 0.05823060870170593, 0.026841720566153526, 0.027056174352765083, 0.048426706343889236, 0.010735553689301014, 0.010862319730222225, -0.018639683723449707, -0.023750318214297295, 0.09762035310268402, 0.09133223444223404, 0.08627956360578537, -0.007459786720573902, 0.23380990326404572, -0.21230313181877136, -0.06307428330183029, 0.056544676423072815, 0.036853816360235214, -0.004429623018950224, 0.09784812480211258, -0.014769429340958595, 0.05134075880050659, -0.004999649245291948, 0.019981782883405685, 0.028932947665452957, 0.032092489302158356, 0.028010331094264984, 0.005597804673016071, -0.05239646136760712, 0.04630530998110771, -0.06581759452819824, 0.03227567300200462, 0.09712249040603638, 0.09160342067480087, -0.111119844019413, -0.030906977131962776, 0.037119779735803604, -0.07258937507867813, 0.027280380949378014, -0.03328874707221985, 0.25297489762306213, -0.0843178778886795, 0.016688689589500427, -0.04690026491880417, -0.10564955323934555, -0.029938485473394394, 0.21392682194709778, -0.020851390436291695, -0.09095650166273117, -0.0723983645439148, -0.09746799618005753, -0.008068620227277279, -0.036477431654930115, 0.13020549714565277, -0.014563542790710926, 0.2546294331550598, -0.07332326471805573, 0.14419139921665192, -0.039864663034677505, -0.03291698172688484, -0.13655917346477509, 0.03514297679066658, 0.014701503328979015, 0.049873728305101395, -0.11643572896718979, 0.08591870963573456, 0.040825214236974716, -0.11419375240802765, 0.02042197808623314, -0.04055430367588997, -0.045261453837156296, 0.026403332129120827, -0.08297815173864365, 0.02011696994304657, 0.05210403725504875, 0.012400192208588123, 0.060313496738672256, -0.14621219038963318, 0.08028019219636917, -0.011030790396034718, -0.1368531584739685, 0.04029347375035286, -0.03107001818716526, 0.28481629490852356, -0.03916705399751663, -0.04223507642745972, 0.09130118787288666, -0.030129779130220413, -0.1580190658569336, -0.011470372788608074, -0.01500412542372942, -0.07619579136371613, -0.026538856327533722, 0.19843535125255585, -0.051759764552116394, 0.0993339791893959, 0.07308448851108551, 0.08785372972488403, 0.08713028579950333, -0.17283667623996735, 0.04673052951693535, -0.034548718482255936, 0.10557801276445389, -0.028818121179938316, 0.1454341560602188, 0.040924735367298126, -0.018426815047860146, 0.027366328984498978, -0.0349990613758564, -0.08906248956918716, 0.039363447576761246, -0.049302272498607635, -0.012518866918981075, -0.050437964498996735, 0.026650935411453247, -0.05698582902550697, 0.05083151534199715, -0.07009845972061157, -0.1319681853055954, -0.004585068207234144, -0.027194103226065636, 0.05040360242128372, 0.04883236438035965, 0.04564709961414337, -0.06707701832056046, -0.0066094910725951195, -0.028295332565903664, -0.006054517347365618, 0.1894230991601944, 0.00723657151684165, -0.010830615647137165 ]
null
null
transformers
# My Awesome Model
{"tags": ["conversational"]}
text-generation
awvik360/DialoGPT-medium-plemons
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# My Awesome Model
[ "# My Awesome Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# My Awesome Model" ]
[ 51, 4 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# My Awesome Model" ]
[ -0.05259015038609505, 0.05521034821867943, -0.005910294596105814, 0.017722278833389282, 0.15250112116336823, 0.02286236733198166, 0.07657632976770401, 0.09513414651155472, -0.025391526520252228, -0.047348517924547195, 0.15119488537311554, 0.19781284034252167, -0.020334534347057343, 0.101333387196064, -0.04688440263271332, -0.3143521845340729, 0.06439975649118423, 0.05463787540793419, -0.015605635941028595, 0.12023304402828217, 0.09468326717615128, -0.0530015267431736, 0.08742043375968933, -0.012155864387750626, -0.1293085366487503, -0.0027921805158257484, -0.002384399762377143, -0.10180269181728363, 0.11194873601198196, 0.033712033182382584, 0.05166437849402428, 0.0182647667825222, -0.05843055993318558, -0.139859139919281, 0.03845210000872612, -0.015005595050752163, -0.05602653697133064, 0.05648263916373253, 0.059830192476511, -0.07164353132247925, 0.1669619083404541, 0.13275989890098572, -0.04237370565533638, 0.056127581745386124, -0.17620700597763062, 0.017941240221261978, 0.01800798624753952, 0.019184142351150513, 0.05306641012430191, 0.10830496996641159, -0.03932326287031174, 0.09217294305562973, -0.11410652846097946, 0.08313368260860443, 0.07800983637571335, -0.29151955246925354, -0.025312699377536774, 0.10440942645072937, 0.06437138468027115, 0.048375632613897324, -0.013386772945523262, 0.0621674507856369, 0.02149512618780136, 0.008602659218013287, 0.02225899137556553, -0.06727100163698196, -0.05789240449666977, 0.032748885452747345, -0.0967593789100647, -0.03634428232908249, 0.19753605127334595, -0.024647634476423264, 0.053590498864650726, -0.06265407055616379, -0.11300963163375854, -0.039751436561346054, -0.050429005175828934, -0.029761891812086105, -0.05090925097465515, 0.09489558637142181, 0.004352911841124296, -0.09534718841314316, -0.13405443727970123, -0.01370926946401596, -0.1618979275226593, 0.15892250835895538, 0.012579603120684624, 0.046201955527067184, -0.19210097193717957, 0.11465331166982651, -0.03857925534248352, -0.08259090781211853, 0.030513519421219826, -0.12010065466165543, 0.03160654753446579, -0.008132083341479301, -0.019599268212914467, -0.049325279891490936, 0.061037879437208176, 0.08101806789636612, 0.018783701583743095, 0.005755073390901089, 0.018167443573474884, 0.05343452841043472, 0.05891622602939606, 0.10033947974443436, -0.02891627699136734, -0.0625043511390686, 0.0025436533614993095, -0.12051084637641907, -0.01122665498405695, -0.05357983708381653, -0.18095199763774872, 0.002246231772005558, 0.02455340512096882, 0.05192234739661217, 0.011778532527387142, 0.09955989569425583, -0.028496338054537773, -0.026898741722106934, 0.06898727267980576, 0.002862759632989764, -0.015707949176430702, -0.005368964280933142, -0.010934269987046719, 0.11485416442155838, -0.023099146783351898, 0.04774846136569977, -0.12022071331739426, 0.020393015816807747, -0.07851235568523407, -0.0019349842332303524, -0.06214260309934616, -0.04864754155278206, -0.0019346009939908981, -0.06985589861869812, 0.021118074655532837, -0.14833110570907593, -0.17990200221538544, -0.005064866971224546, 0.021302316337823868, -0.052403319627046585, -0.09162671118974686, -0.0982397273182869, -0.02586611732840538, 0.03574685752391815, -0.05873546749353409, 0.013170980848371983, -0.06884536147117615, 0.06542801111936569, 0.0029820678755640984, 0.05682007595896721, -0.14085575938224792, 0.08719147741794586, -0.12582023441791534, -0.023288866505026817, -0.061977192759513855, 0.1109607070684433, 0.024780582636594772, 0.1267160177230835, 0.004311583004891872, -0.0033308975398540497, -0.08729329705238342, 0.08271238207817078, -0.04243258014321327, 0.22770646214485168, -0.10479787737131119, -0.08809807151556015, 0.2632525563240051, -0.05423165112733841, -0.16432519257068634, 0.10179096460342407, -0.014350244775414467, 0.12198644131422043, 0.13850919902324677, 0.16080057621002197, 0.007628654129803181, 0.03313867375254631, 0.10115300863981247, 0.08631709218025208, -0.08573295921087265, -0.0611947737634182, 0.023627014830708504, -0.011463395319879055, -0.10670105367898941, 0.046802595257759094, 0.04794782027602196, 0.08188598603010178, -0.04982871189713478, -0.028600862249732018, -0.01972118206322193, -0.044152840971946716, 0.05264130234718323, 0.007675500120967627, 0.13217447698116302, -0.03674980252981186, -0.03692879155278206, -0.023745311424136162, 0.01699630729854107, -0.03115241602063179, 0.007061392068862915, -0.05687357112765312, 0.11091547459363937, -0.03406180441379547, 0.051789235323667526, -0.16953988373279572, -0.04873261600732803, -0.02087729424238205, 0.1402055323123932, 0.04973345249891281, 0.1329866498708725, 0.06287940591573715, -0.010758201591670513, 0.00859389640390873, 0.007998145185410976, 0.13181665539741516, 0.007865442894399166, -0.07660657912492752, -0.047718439251184464, 0.09176599979400635, -0.05973208695650101, 0.06147782504558563, -0.098741315305233, -0.004747362341731787, -0.01433002483099699, 0.08674649894237518, 0.006352655589580536, 0.029382232576608658, -0.006192679051309824, 0.003654100699350238, -0.06161240115761757, 0.017873648554086685, 0.12492607533931732, -0.01421504095196724, -0.07439801841974258, 0.22084392607212067, -0.15798072516918182, 0.18006981909275055, 0.18165533244609833, -0.3081994652748108, 0.024602634832262993, -0.08860466629266739, -0.036338552832603455, 0.03426366671919823, 0.0491504967212677, -0.034147560596466064, 0.16587987542152405, -0.016766328364610672, 0.201018825173378, -0.03547777235507965, -0.01287798210978508, -0.010399105958640575, -0.03656993433833122, -0.010632630437612534, 0.09065473079681396, 0.15122920274734497, -0.1677125245332718, 0.18270380795001984, 0.1660280078649521, 0.06873020529747009, 0.17776396870613098, 0.034313347190618515, -0.006856906693428755, 0.07112615555524826, -0.022670727223157883, -0.07675548642873764, -0.049287427216768265, -0.26302891969680786, -0.027947327122092247, 0.06471601128578186, 0.04510856419801712, 0.11924877762794495, -0.10971947014331818, -0.037208184599876404, 0.010892451740801334, -0.013165894895792007, 0.02132410928606987, 0.09682225435972214, 0.01171150617301464, 0.11804302036762238, -0.021027036011219025, -0.05209195241332054, 0.0898953229188919, 0.02727191150188446, -0.0787680521607399, 0.19168277084827423, -0.10074768215417862, -0.3233809769153595, -0.11354339867830276, -0.18166927993297577, -0.017843691632151604, 0.05878754332661629, 0.08049646019935608, -0.09228580445051193, -0.02625267766416073, -0.01639235019683838, 0.0758359357714653, -0.09145816415548325, -0.015880629420280457, -0.09367848187685013, 0.034986745566129684, -0.10827737301588058, -0.07011983543634415, -0.05141967162489891, -0.03368452936410904, -0.04457031562924385, 0.13157756626605988, -0.12242637574672699, 0.06396433711051941, 0.2076517641544342, 0.06227295100688934, 0.05622440204024315, -0.0229496993124485, 0.23288212716579437, -0.10842552781105042, 0.02383521944284439, 0.1717897206544876, -0.03566030040383339, 0.0727933868765831, 0.13435456156730652, 0.006721907295286655, -0.08144525438547134, 0.03465581312775612, -0.04592517390847206, -0.08630958944559097, -0.20441576838493347, -0.14156180620193481, -0.12814727425575256, 0.07913564145565033, 0.03285396471619606, 0.05478321388363838, 0.15024253726005554, 0.11386489123106003, 0.007987297140061855, 0.00976672861725092, -0.006888182368129492, 0.05438044294714928, 0.17482298612594604, -0.05838097631931305, 0.10041683167219162, -0.037591226398944855, -0.1924494504928589, 0.08022978901863098, 0.04309763014316559, 0.08280511945486069, 0.07474655658006668, 0.0856199786067009, 0.013537914492189884, 0.03723837807774544, 0.10897084325551987, 0.1165735274553299, 0.031679023057222366, -0.038079675287008286, -0.04882059991359711, -0.026300756260752678, -0.03285675123333931, 0.05745977535843849, 0.07790146768093109, -0.1608346849679947, -0.06348084658384323, -0.06350091099739075, 0.07662643492221832, 0.09017108380794525, 0.11811108142137527, -0.21219493448734283, 0.01579318381845951, 0.092556893825531, -0.0494147390127182, -0.1304239183664322, 0.07402537018060684, -0.00466050673276186, -0.1397053301334381, 0.037663187831640244, -0.014095795340836048, 0.1359514445066452, -0.0778401643037796, 0.10336452722549438, -0.08307972550392151, -0.06147889420390129, 0.03632286190986633, 0.1355396956205368, -0.30774354934692383, 0.2137020230293274, -0.022472934797406197, -0.05296783149242401, -0.10508129745721817, -0.011727629229426384, 0.020913105458021164, 0.09079049527645111, 0.10090240091085434, -0.0025442070327699184, 0.0061299679800868034, -0.0345483273267746, -0.053218815475702286, 0.024456629529595375, 0.07957815378904343, -0.08542889356613159, 0.0017540202243253589, -0.02361489273607731, -0.004407065454870462, -0.032844748347997665, -0.01189463958144188, -0.011617658659815788, -0.16786961257457733, 0.06556065380573273, -0.002625665394589305, 0.11129079759120941, 0.03491498529911041, 0.0024013579823076725, -0.1009332686662674, 0.19977013766765594, 0.01796281896531582, -0.08052749931812286, -0.08830537647008896, -0.03254766762256622, 0.03660419583320618, -0.06121435388922691, 0.027481911703944206, -0.06916457414627075, 0.033381566405296326, -0.06441576033830643, -0.18325145542621613, 0.1268530637025833, -0.10945470631122589, -0.03609596937894821, -0.04321056231856346, 0.18323224782943726, -0.00929707009345293, -0.0011623724130913615, 0.05866571143269539, 0.0032208464108407497, -0.1347510665655136, -0.10740556567907333, 0.020214511081576347, -0.015275230631232262, 0.009142245166003704, 0.05559912323951721, -0.009665844030678272, 0.00045268211397342384, -0.039558928459882736, -0.023234419524669647, 0.32348164916038513, 0.10732097923755646, -0.04944206401705742, 0.17007054388523102, 0.13087597489356995, -0.0827672928571701, -0.30699312686920166, -0.10971353948116302, -0.10529600828886032, -0.026918673887848854, -0.037983208894729614, -0.19617970287799835, 0.09504909813404083, -0.03528566658496857, -0.022136637941002846, 0.11253651231527328, -0.2759084105491638, -0.0770430713891983, 0.1826775223016739, 0.003314757253974676, 0.3998824954032898, -0.10265109688043594, -0.08777514100074768, -0.06741699576377869, -0.1120782196521759, 0.2033512443304062, -0.05560711398720741, 0.08663415163755417, -0.00517998356372118, 0.15513743460178375, 0.055607251822948456, -0.02176513522863388, 0.08932057023048401, -0.005811662413179874, -0.0546204075217247, -0.1219351515173912, -0.03444604203104973, -0.009159418754279613, 0.007239421829581261, 0.03589896112680435, -0.04242607578635216, 0.01279151439666748, -0.1399589478969574, -0.045490626245737076, -0.0764620453119278, 0.024699507281184196, 0.021008269861340523, -0.0652410089969635, -0.01643640361726284, -0.03945036977529526, -0.012804778292775154, 0.03164318576455116, 0.15236099064350128, -0.06478006392717361, 0.1476556956768036, 0.04904455319046974, 0.15412139892578125, -0.14745712280273438, -0.02258288487792015, -0.06896031647920609, -0.05498642474412918, 0.04900865629315376, -0.10053684562444687, 0.050061121582984924, 0.1202658861875534, -0.0742902010679245, 0.0987328365445137, 0.0922594666481018, -0.01938629150390625, 0.0012483424507081509, 0.1226617842912674, -0.2489612102508545, -0.07742628455162048, -0.10509459674358368, 0.013337249867618084, 0.10138551890850067, 0.06995654851198196, 0.17304721474647522, -0.0037713919300585985, -0.036284226924180984, -0.0064643872901797295, 0.025414984673261642, -0.03540204465389252, 0.05724727362394333, -0.002706433180719614, 0.016663886606693268, -0.15213344991207123, 0.060368724167346954, -0.00024176653823815286, -0.1438901126384735, -0.013603870756924152, 0.16073721647262573, -0.11208858340978622, -0.15145981311798096, -0.007263668347150087, 0.13685113191604614, -0.13171035051345825, -0.03302847594022751, -0.03708777576684952, -0.170182466506958, 0.07439173012971878, 0.1024777740240097, 0.08549231290817261, 0.08025266975164413, -0.06620611250400543, -0.00807863101363182, -0.011656313203275204, -0.026087598875164986, 0.031810320913791656, -0.023377234116196632, -0.09044221043586731, 0.03872343525290489, -0.026654237881302834, 0.13591371476650238, -0.09607382118701935, -0.09331836551427841, -0.135749951004982, 0.039314381778240204, -0.12405620515346527, -0.08138058334589005, -0.12200927734375, -0.0591500885784626, 0.00224387738853693, -0.0001289021165575832, -0.035674065351486206, -0.06687422841787338, -0.13582271337509155, 0.04366770386695862, -0.04484611004590988, 0.0013091047294437885, -0.040241483598947525, 0.04561002552509308, 0.06766383349895477, -0.03493715822696686, 0.13722217082977295, 0.11722734570503235, -0.07864081114530563, 0.08946478366851807, -0.16657429933547974, -0.0683990865945816, 0.08854512125253677, 0.008173754438757896, 0.06165994703769684, 0.06743349134922028, 0.033807408064603806, 0.06109451875090599, 0.04151686280965805, 0.03488299250602722, 0.01739438995718956, -0.09271225333213806, 0.015541021712124348, 0.022296719253063202, -0.1294609159231186, -0.04801803454756737, -0.029226921498775482, 0.00939185917377472, 0.008117396384477615, 0.11003357172012329, -0.0426274873316288, 0.09439733624458313, -0.05888751894235611, 0.036728594452142715, 0.016222506761550903, -0.16461637616157532, -0.020102784037590027, -0.11915475130081177, 0.028684545308351517, -0.0033096212428063154, 0.25625869631767273, 0.06346847862005234, 0.020517030730843544, 0.01250078622251749, 0.08567021042108536, 0.07241600006818771, 0.02562166005373001, 0.1956365555524826, 0.10854171961545944, -0.05020022392272949, -0.12334850430488586, 0.09686340391635895, 0.034720368683338165, 0.06432123482227325, 0.13385434448719025, -0.026959087699651718, 0.002498799469321966, 0.11019360274076462, 0.011678861454129219, 0.04961980879306793, -0.09859088063240051, -0.16400282084941864, -0.00994415208697319, 0.061864156275987625, -0.04559077322483063, 0.12240655720233917, 0.11382720619440079, -0.020697353407740593, 0.03180128335952759, -0.010503606870770454, -0.05694027617573738, -0.16998925805091858, -0.1630837321281433, -0.08357038348913193, -0.11794789135456085, -0.0027763545513153076, -0.11386270076036453, 0.013879159465432167, 0.06452289968729019, 0.0604364387691021, -0.09019444137811661, 0.08891061693429947, 0.0687386617064476, -0.11843101680278778, 0.08828350901603699, -0.033263903111219406, 0.07249268144369125, 0.0015160300536081195, 0.003872724948450923, -0.13800905644893646, 0.032393742352724075, -0.008493867702782154, 0.04159298539161682, -0.09244006127119064, 0.022458361461758614, -0.11297028511762619, -0.07659684121608734, -0.07971972227096558, 0.05093973129987717, -0.03541257977485657, 0.1390930563211441, 0.001295371213927865, -0.035233911126852036, 0.024190181866288185, 0.22729112207889557, -0.06350252777338028, -0.030667411163449287, -0.0618741400539875, 0.21414142847061157, 0.024466563016176224, 0.10703565180301666, -0.016775688156485558, 0.019240234047174454, -0.0764411985874176, 0.3689337372779846, 0.344390869140625, -0.1225387305021286, -0.0015968306688591838, 0.031062176451086998, 0.036916591227054596, 0.11621878296136856, 0.12602226436138153, 0.057955991476774216, 0.2995031177997589, -0.08396036922931671, -0.002026971662417054, -0.02688612788915634, -0.03624163940548897, -0.04409930482506752, 0.10547586530447006, 0.06835740804672241, -0.03330419585108757, -0.027012333273887634, 0.1376710683107376, -0.2966996431350708, 0.12323499470949173, -0.15714547038078308, -0.1487535685300827, -0.06873904913663864, -0.005042468197643757, 0.08589684963226318, 0.04748665541410446, 0.1069009080529213, -0.019124338403344154, -0.08203735202550888, 0.05766449123620987, 0.0320524163544178, -0.22844897210597992, 0.011852608993649483, 0.08361081779003143, -0.06153005734086037, 0.011767351068556309, -0.017906347289681435, 0.038472190499305725, 0.07790610194206238, 0.025976579636335373, -0.032770540565252304, 0.06325861811637878, -0.005814229138195515, -0.05033424496650696, 0.04302205145359039, 0.05059972032904625, 0.017107632011175156, -0.1511564701795578, 0.07320158183574677, -0.1762860119342804, 0.0566408596932888, -0.005331212189048529, -0.04948166385293007, 0.000018263708625454456, 0.01998119056224823, -0.06808236241340637, 0.05880929157137871, 0.0952666699886322, -0.012173139490187168, -0.002317852806299925, -0.056667573750019073, 0.007662574760615826, -0.0679154172539711, -0.0747012197971344, -0.10497893393039703, -0.1338900774717331, -0.11392296850681305, 0.10846775025129318, -0.011928223073482513, -0.19833622872829437, 0.02906924858689308, -0.11258108913898468, 0.04933213070034981, -0.13360801339149475, 0.08599711954593658, 0.1282832771539688, 0.021543797105550766, -0.01265349704772234, 0.04020093381404877, 0.01591683179140091, 0.08550478518009186, -0.09200563281774521, -0.10515180230140686 ]
null
null
null
# My Awesome Model
{"tags": ["conversational"]}
text-generation
awvik360/DialoGPT-medium-plemons2
[ "conversational", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #conversational #region-us
# My Awesome Model
[ "# My Awesome Model" ]
[ "TAGS\n#conversational #region-us \n", "# My Awesome Model" ]
[ 10, 4 ]
[ "passage: TAGS\n#conversational #region-us \n# My Awesome Model" ]
[ -0.03546040877699852, 0.10198262333869934, -0.009167643263936043, -0.06873539090156555, 0.09720280766487122, 0.08721881359815598, 0.05205019563436508, -0.007754879537969828, 0.1307058483362198, -0.07418710738420486, 0.15552383661270142, 0.09356602281332016, -0.055710893124341965, 0.054602161049842834, 0.023362739011645317, -0.24776877462863922, 0.05416174605488777, -0.018358204513788223, 0.038026146590709686, 0.059432677924633026, -0.008917242288589478, -0.06420617550611496, 0.032797642052173615, -0.054550930857658386, -0.047955937683582306, 0.05154015123844147, -0.015393521636724472, 0.022211134433746338, 0.13093532621860504, -0.025035109370946884, 0.11168728768825531, 0.02722480148077011, -0.08139488101005554, -0.2497675120830536, 0.04726335033774376, -0.03638986125588417, -0.05304733291268349, 0.00051964569138363, 0.040234215557575226, -0.0361054390668869, 0.10917626321315765, 0.2034350037574768, -0.013044852763414383, 0.12713150680065155, -0.29209578037261963, -0.05327993258833885, 0.019334495067596436, -0.017972392961382866, 0.007713220082223415, 0.010176903568208218, -0.0199898611754179, 0.12909340858459473, -0.18584373593330383, -0.03348228707909584, -0.05190730839967728, -0.19455693662166595, 0.02133459411561489, 0.195334330201149, 0.03859388828277588, 0.13404928147792816, -0.03862352669239044, 0.11985649168491364, -0.001176132122054696, -0.02696487121284008, -0.09787412732839584, -0.083623968064785, 0.05212550610303879, 0.10331118106842041, -0.045747701078653336, -0.01321598794311285, 0.2609056234359741, 0.08284381031990051, 0.022602153941988945, 0.052660685032606125, -0.037852801382541656, -0.01829439401626587, 0.0014054370112717152, -0.11398514360189438, -0.012235384434461594, 0.15790092945098877, 0.0919213518500328, -0.11842559278011322, -0.11358761787414551, 0.05386804789304733, -0.19349870085716248, 0.11392602324485779, -0.05408310890197754, 0.09635519981384277, -0.2302265763282776, -0.03165312856435776, -0.23591788113117218, -0.03137180954217911, 0.008059423416852951, -0.1426098644733429, -0.019717667251825333, -0.05638215318322182, 0.06416983157396317, 0.10064838826656342, 0.026653507724404335, 0.05982319638133049, 0.0024214035365730524, 0.03094852901995182, 0.04355722293257713, 0.06075821444392204, 0.13682802021503448, 0.07382544130086899, 0.08105745166540146, -0.015595734119415283, -0.13757064938545227, -0.15028178691864014, -0.042663972824811935, -0.0603194423019886, -0.10847903043031693, 0.1175917387008667, -0.1348201483488083, 0.056088559329509735, -0.008282779715955257, -0.06160689890384674, -0.1302913874387741, 0.07904011011123657, -0.009780440479516983, 0.07031673192977905, -0.0500408336520195, -0.030323538929224014, -0.012846678495407104, 0.09941235184669495, -0.15968181192874908, 0.07661482691764832, 0.08417865633964539, -0.017821768298745155, -0.13900868594646454, -0.0261816568672657, -0.06725679337978363, 0.07161393761634827, 0.027083612978458405, -0.05911887437105179, 0.07335446029901505, -0.1062290146946907, -0.10800924897193909, -0.0462496280670166, 0.05760948359966278, -0.04071643948554993, 0.02777569182217121, -0.08487419039011002, 0.04323674738407135, -0.043959055095911026, -0.011524932458996773, -0.0954136773943901, -0.10528495907783508, 0.03591811656951904, 0.07539372146129608, 0.007705247029662132, -0.16573995351791382, 0.034548647701740265, -0.0835181400179863, 0.049681901931762695, -0.02308676578104496, 0.0353124737739563, 0.043504517525434494, 0.1896352767944336, 0.09132891893386841, 0.11749628186225891, -0.18855012953281403, 0.05541319027543068, -0.12649250030517578, 0.2574777901172638, -0.15162399411201477, -0.0263845045119524, 0.23871201276779175, -0.028618283569812775, -0.1612473875284195, 0.031156204640865326, -0.024838248267769814, 0.21029748022556305, 0.15949539840221405, 0.3231875002384186, -0.14657290279865265, -0.040706682950258255, 0.10120458155870438, 0.1641000658273697, -0.0947515219449997, -0.05097932368516922, 0.07063271850347519, -0.08443530648946762, -0.10856795310974121, -0.009461409412324429, 0.06791952252388, 0.08725127577781677, -0.08676911145448685, -0.057048194110393524, 0.04897143691778183, -0.05990632250905037, 0.0539734810590744, 0.09225133061408997, 0.019252249971032143, -0.06463979184627533, 0.03549259155988693, -0.03487003222107887, 0.06323052942752838, 0.1344345659017563, -0.07567445188760757, -0.04621171951293945, 0.05208157002925873, -0.008416261523962021, 0.011982166208326817, -0.07247529923915863, -0.10293032974004745, -0.09282280504703522, 0.15225498378276825, 0.12907147407531738, 0.24349625408649445, 0.07771562039852142, -0.08539113402366638, 0.023976441472768784, 0.05230777710676193, 0.02473422884941101, 0.08408135920763016, 0.006961719132959843, -0.04070304334163666, 0.1336214691400528, -0.06637702882289886, 0.0505339689552784, -0.15414337813854218, -0.06644152104854584, -0.003572809975594282, 0.03696983680129051, 0.08356471359729767, -0.04645514488220215, -0.0019455882720649242, 0.0381651446223259, 0.05654265359044075, 0.010789581574499607, 0.11545533686876297, -0.04579673334956169, -0.07804842293262482, 0.17771583795547485, -0.10181614011526108, 0.1486288160085678, 0.11471833288669586, -0.2789360582828522, 0.027798451483249664, -0.09091047197580338, -0.016147438436746597, 0.03836512193083763, 0.07616107910871506, -0.04532977193593979, 0.0876641571521759, 0.008135076612234116, 0.045742783695459366, 0.030224641785025597, 0.049904048442840576, -0.05050472542643547, -0.038475628942251205, -0.1468115597963333, 0.11378216743469238, 0.14237768948078156, -0.1664484292268753, 0.15665550529956818, 0.344457745552063, 0.2189280241727829, 0.24893705546855927, -0.02391449362039566, -0.0023995572701096535, -0.007929098792374134, -0.05631881207227707, -0.1310952752828598, 0.1292494386434555, -0.3013957738876343, -0.0045976797118783, 0.0019251068588346243, 0.020475171506404877, 0.10425154119729996, -0.11188157647848129, -0.11943034082651138, 0.01928076706826687, 0.013840734958648682, -0.04093893617391586, 0.05837646499276161, -0.08973170816898346, 0.06690972298383713, 0.044297512620687485, -0.09705004096031189, 0.12118203192949295, 0.032428398728370667, -0.027439109981060028, 0.06038731336593628, -0.13205169141292572, -0.161821186542511, -0.014722511172294617, -0.12305234372615814, 0.03822726011276245, 0.02258457988500595, -0.0011937115341424942, -0.11695839464664459, -0.017612792551517487, 0.08651383221149445, 0.06550532579421997, -0.21545930206775665, -0.08663901686668396, -0.05402546748518944, 0.06747249513864517, -0.14086408913135529, -0.005392237100750208, -0.0580231137573719, -0.04639870673418045, -0.015921805053949356, 0.03196299821138382, -0.1326124668121338, 0.05850570648908615, 0.1952214241027832, 0.11960950493812561, 0.08012472093105316, 0.00031379942083731294, 0.25854313373565674, -0.15044859051704407, -0.031102577224373817, 0.04013100266456604, -0.03211164101958275, 0.08570355176925659, 0.2004045695066452, 0.07713621109724045, -0.04407672584056854, -0.032609183341264725, -0.06265520304441452, -0.08382013440132141, -0.17984598875045776, -0.09944407641887665, -0.09241506457328796, 0.11593815684318542, -0.10455609858036041, 0.02279566414654255, 0.1293002963066101, 0.038562655448913574, 0.10591760277748108, -0.17956066131591797, -0.08083869516849518, -0.015997247770428658, 0.10070767998695374, -0.1476999819278717, -0.036619413644075394, -0.057095691561698914, -0.13042104244232178, 0.09085097908973694, 0.07351890206336975, -0.07598336786031723, 0.2753245234489441, 0.14356902241706848, 0.06460168957710266, 0.0372491329908371, 0.050815973430871964, 0.03296361491084099, 0.06032256409525871, -0.08821487426757812, -0.024852164089679718, 0.008612229488790035, -0.022380370646715164, 0.05025824159383774, 0.21010570228099823, -0.24013914167881012, -0.010044138878583908, -0.12094947695732117, 0.058911196887493134, -0.09226593375205994, 0.15273147821426392, -0.005419398192316294, 0.07938405126333237, 0.13775718212127686, 0.017697615548968315, -0.08790077269077301, 0.10226619243621826, 0.06094779446721077, -0.12483128160238266, -0.00920578371733427, 0.10987824946641922, 0.08385234326124191, -0.016504161059856415, 0.11643458902835846, -0.21195663511753082, -0.13761650025844574, 0.033488254994153976, 0.10529548674821854, -0.1958140879869461, 0.3039077818393707, 0.009235309436917305, -0.1351068764925003, -0.0639132410287857, -0.11496353149414062, -0.012014171108603477, 0.10743112862110138, 0.10711206495761871, 0.042469725012779236, -0.07393775135278702, -0.026096675544977188, 0.009214960969984531, -0.007742607034742832, 0.09298452734947205, -0.08414001762866974, -0.12013377249240875, 0.010150929912924767, 0.03940318152308464, -0.048703476786613464, 0.1009429320693016, -0.08256068825721741, -0.07715889811515808, -0.009262125939130783, -0.013167787343263626, 0.013363508507609367, 0.0613013356924057, 0.09955485910177231, -0.03248724341392517, -0.016322879120707512, 0.17398953437805176, 0.06143142655491829, -0.028618335723876953, -0.15928512811660767, -0.0019777673296630383, -0.04513169452548027, -0.039428479969501495, -0.06899980455636978, -0.07661525160074234, -0.11753853410482407, -0.09040885418653488, 0.12060512602329254, -0.1011100485920906, 0.08252820372581482, -0.07234728336334229, 0.1556718945503235, 0.0449947789311409, 0.027362238615751266, 0.0420842170715332, 0.010504845529794693, -0.0637444332242012, -0.06514158844947815, 0.13852131366729736, -0.17852531373500824, -0.03856880962848663, 0.10458311438560486, 0.07282499223947525, 0.011911713518202305, 0.01789284311234951, -0.12299755960702896, 0.19862310588359833, 0.21264135837554932, -0.00729813938960433, 0.16878525912761688, 0.24475444853305817, -0.06593596935272217, -0.205520361661911, -0.06422454118728638, -0.23529553413391113, -0.07331079244613647, 0.15935871005058289, -0.17814143002033234, 0.09661445021629333, -0.008739825338125229, -0.0568917840719223, 0.14693109691143036, -0.32532361149787903, -0.005068281665444374, 0.19984132051467896, -0.03098643198609352, 0.5961567759513855, -0.067476287484169, -0.10163161158561707, -0.020260926336050034, -0.08047869801521301, 0.2328169047832489, -0.10700056701898575, 0.022178582847118378, 0.05551070347428322, 0.10260957479476929, 0.06909338384866714, 0.020471658557653427, 0.1812010407447815, -0.05112580955028534, -0.06011022627353668, -0.11829929798841476, -0.20441202819347382, 0.01321274135261774, -0.002565407194197178, -0.09609947353601456, 0.06216618791222572, -0.06233842670917511, -0.17662794888019562, 0.015285306610167027, -0.08876541256904602, -0.03281906247138977, 0.012494787573814392, -0.04522540047764778, -0.03335950896143913, 0.03184078261256218, -0.1269286721944809, 0.02083008363842964, 0.15761429071426392, -0.09231042861938477, 0.21183836460113525, -0.09038146585226059, 0.11980026215314865, -0.1715039312839508, -0.07381453365087509, -0.0892910361289978, -0.0739288181066513, 0.022808637470006943, -0.05498562753200531, 0.039651405066251755, 0.1229073703289032, -0.06014255806803703, 0.13346922397613525, 0.04274857044219971, -0.07362601161003113, -0.009332284331321716, 0.14846959710121155, -0.19871611893177032, -0.2910851538181305, -0.11566967517137527, 0.05923045426607132, 0.2028619349002838, 0.007538134697824717, 0.0777674987912178, 0.11478970944881439, -0.016507035121321678, 0.015079355798661709, 0.010267493315041065, -0.11431148648262024, -0.10444751381874084, 0.06896253675222397, 0.0007569619920104742, -0.08558188378810883, 0.10230734944343567, 0.019022267311811447, -0.18386487662792206, -0.11988023668527603, 0.20762711763381958, -0.0200219564139843, -0.09439916163682938, -0.0994015783071518, 0.1823630928993225, -0.060141727328300476, -0.026388145983219147, 0.060486629605293274, -0.0971314087510109, -0.027692077681422234, 0.17358030378818512, 0.05215360224246979, 0.07331566512584686, 0.043052736669778824, -0.019896553829312325, 0.19111433625221252, -0.07610252499580383, -0.08666720986366272, -0.11411479860544205, -0.1113981381058693, -0.09765390306711197, -0.022090891376137733, 0.17170315980911255, -0.1043042540550232, -0.1465844213962555, -0.23367111384868622, 0.08566625416278839, -0.07618606090545654, -0.14835943281650543, -0.12351330369710922, -0.09960491955280304, 0.07022807002067566, -0.005836328491568565, -0.025358503684401512, -0.09784673154354095, -0.1479180008172989, 0.10302628576755524, 0.09353149682283401, 0.02215663343667984, 0.03276374191045761, 0.06490205228328705, 0.16425716876983643, 0.006264934781938791, 0.11560661345720291, 0.09335105121135712, 0.004334130324423313, 0.12737876176834106, -0.24313320219516754, -0.03612852096557617, 0.061543241143226624, -0.02008756995201111, 0.03869541361927986, 0.1556338667869568, -0.07101669907569885, -0.008599703200161457, 0.07346312701702118, 0.05884246155619621, -0.06158248707652092, -0.07029277831315994, -0.020444681867957115, 0.15146341919898987, -0.21854759752750397, -0.010464908555150032, -0.13543701171875, 0.08618341386318207, -0.06382738798856735, 0.026516791433095932, 0.07620655745267868, 0.08659784495830536, 0.003671627026051283, 0.051998503506183624, 0.02891702763736248, -0.10376659035682678, 0.11479650437831879, -0.1011483371257782, -0.010525095276534557, -0.04059837758541107, 0.3260350227355957, 0.008407027460634708, 0.01702333241701126, 0.04124368727207184, 0.1525338888168335, 0.036274950951337814, 0.002469088416546583, 0.11112259328365326, 0.1318541020154953, -0.05502143129706383, -0.1530759334564209, 0.1053222268819809, -0.03983991593122482, 0.017480194568634033, 0.12883120775222778, -0.017984678968787193, 0.05133776366710663, 0.0598396472632885, 0.03326093778014183, 0.06138930097222328, 0.08058228343725204, -0.2519816756248474, 0.05864633992314339, -0.008193781599402428, -0.10156036913394928, 0.14093659818172455, 0.12776172161102295, -0.04358195886015892, 0.03643115237355232, -0.08332061767578125, -0.017144199460744858, -0.13900910317897797, -0.08347687125205994, 0.011046548373997211, -0.0890955775976181, 0.024407757446169853, -0.02158970944583416, -0.01773250661790371, 0.20654316246509552, 0.00810841005295515, -0.08619078248739243, 0.024125345051288605, -0.03246486932039261, -0.1350407898426056, -0.028485149145126343, 0.004192930646240711, 0.07534855604171753, -0.10631464421749115, -0.009459893219172955, -0.1966288685798645, -0.03228876367211342, -0.10370618849992752, 0.031585320830345154, -0.13690978288650513, -0.056601304560899734, -0.1394949108362198, -0.046161238104104996, -0.07070588320493698, 0.0341620109975338, -0.10124900937080383, 0.14852306246757507, -0.02862531691789627, 0.04669342562556267, 0.001221607206389308, 0.21473883092403412, -0.0055917128920555115, 0.1111997663974762, -0.03376127779483795, 0.025253819301724434, -0.07697580754756927, 0.12172159552574158, -0.06678933650255203, -0.016789207234978676, -0.0435773991048336, 0.26328104734420776, 0.3666163384914398, -0.13708895444869995, -0.03541192784905434, -0.029750946909189224, 0.032959096133708954, 0.055593449622392654, 0.09368465840816498, -0.04153439402580261, 0.29192063212394714, -0.11245544999837875, 0.09509938210248947, 0.017580494284629822, 0.02136683464050293, 0.05382193997502327, 0.03408944979310036, 0.0992191731929779, 0.02480826899409294, -0.06201104819774628, 0.22126658260822296, -0.28489235043525696, 0.12929458916187286, -0.09419834613800049, -0.1977759450674057, -0.024422185495495796, -0.09278329461812973, 0.1075873076915741, 0.009720941074192524, 0.1459415853023529, -0.059647753834724426, -0.14828547835350037, -0.10750431567430496, 0.04621806740760803, -0.34202659130096436, -0.19561326503753662, 0.13985638320446014, 0.041433196514844894, 0.09628590941429138, -0.006398599129170179, 0.017934424802660942, 0.040109071880578995, 0.005008349195122719, 0.011764837428927422, 0.06299147754907608, 0.06230494752526283, -0.03965628892183304, -0.18101933598518372, 0.038948025554418564, 0.03522862121462822, -0.13118009269237518, 0.08088650554418564, -0.19435498118400574, 0.03574736788868904, 0.1191537082195282, -0.08429282158613205, 0.052640412002801895, 0.12771375477313995, -0.12840406596660614, 0.03900361806154251, 0.03826753795146942, 0.04374406486749649, -0.039075564593076706, 0.019395308569073677, 0.00933300144970417, -0.021677589043974876, -0.11719156801700592, -0.12927356362342834, 0.06891030073165894, -0.07255057990550995, 0.15803246200084686, -0.043637361377477646, -0.06318710744380951, 0.02716391533613205, -0.05997319892048836, 0.09305576235055923, -0.026282401755452156, 0.04433848708868027, 0.20267623662948608, 0.046981945633888245, 0.005474291741847992, -0.09724274277687073, 0.07344987988471985, 0.0035022953525185585, 0.002335904398933053, -0.08806760609149933 ]
null
null
transformers
# My Awesome Model
{"tags": ["conversational"]}
text-generation
awvik360/DialoGPT-small-plemons
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# My Awesome Model
[ "# My Awesome Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# My Awesome Model" ]
[ 51, 4 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# My Awesome Model" ]
[ -0.05259015038609505, 0.05521034821867943, -0.005910294596105814, 0.017722278833389282, 0.15250112116336823, 0.02286236733198166, 0.07657632976770401, 0.09513414651155472, -0.025391526520252228, -0.047348517924547195, 0.15119488537311554, 0.19781284034252167, -0.020334534347057343, 0.101333387196064, -0.04688440263271332, -0.3143521845340729, 0.06439975649118423, 0.05463787540793419, -0.015605635941028595, 0.12023304402828217, 0.09468326717615128, -0.0530015267431736, 0.08742043375968933, -0.012155864387750626, -0.1293085366487503, -0.0027921805158257484, -0.002384399762377143, -0.10180269181728363, 0.11194873601198196, 0.033712033182382584, 0.05166437849402428, 0.0182647667825222, -0.05843055993318558, -0.139859139919281, 0.03845210000872612, -0.015005595050752163, -0.05602653697133064, 0.05648263916373253, 0.059830192476511, -0.07164353132247925, 0.1669619083404541, 0.13275989890098572, -0.04237370565533638, 0.056127581745386124, -0.17620700597763062, 0.017941240221261978, 0.01800798624753952, 0.019184142351150513, 0.05306641012430191, 0.10830496996641159, -0.03932326287031174, 0.09217294305562973, -0.11410652846097946, 0.08313368260860443, 0.07800983637571335, -0.29151955246925354, -0.025312699377536774, 0.10440942645072937, 0.06437138468027115, 0.048375632613897324, -0.013386772945523262, 0.0621674507856369, 0.02149512618780136, 0.008602659218013287, 0.02225899137556553, -0.06727100163698196, -0.05789240449666977, 0.032748885452747345, -0.0967593789100647, -0.03634428232908249, 0.19753605127334595, -0.024647634476423264, 0.053590498864650726, -0.06265407055616379, -0.11300963163375854, -0.039751436561346054, -0.050429005175828934, -0.029761891812086105, -0.05090925097465515, 0.09489558637142181, 0.004352911841124296, -0.09534718841314316, -0.13405443727970123, -0.01370926946401596, -0.1618979275226593, 0.15892250835895538, 0.012579603120684624, 0.046201955527067184, -0.19210097193717957, 0.11465331166982651, -0.03857925534248352, -0.08259090781211853, 0.030513519421219826, -0.12010065466165543, 0.03160654753446579, -0.008132083341479301, -0.019599268212914467, -0.049325279891490936, 0.061037879437208176, 0.08101806789636612, 0.018783701583743095, 0.005755073390901089, 0.018167443573474884, 0.05343452841043472, 0.05891622602939606, 0.10033947974443436, -0.02891627699136734, -0.0625043511390686, 0.0025436533614993095, -0.12051084637641907, -0.01122665498405695, -0.05357983708381653, -0.18095199763774872, 0.002246231772005558, 0.02455340512096882, 0.05192234739661217, 0.011778532527387142, 0.09955989569425583, -0.028496338054537773, -0.026898741722106934, 0.06898727267980576, 0.002862759632989764, -0.015707949176430702, -0.005368964280933142, -0.010934269987046719, 0.11485416442155838, -0.023099146783351898, 0.04774846136569977, -0.12022071331739426, 0.020393015816807747, -0.07851235568523407, -0.0019349842332303524, -0.06214260309934616, -0.04864754155278206, -0.0019346009939908981, -0.06985589861869812, 0.021118074655532837, -0.14833110570907593, -0.17990200221538544, -0.005064866971224546, 0.021302316337823868, -0.052403319627046585, -0.09162671118974686, -0.0982397273182869, -0.02586611732840538, 0.03574685752391815, -0.05873546749353409, 0.013170980848371983, -0.06884536147117615, 0.06542801111936569, 0.0029820678755640984, 0.05682007595896721, -0.14085575938224792, 0.08719147741794586, -0.12582023441791534, -0.023288866505026817, -0.061977192759513855, 0.1109607070684433, 0.024780582636594772, 0.1267160177230835, 0.004311583004891872, -0.0033308975398540497, -0.08729329705238342, 0.08271238207817078, -0.04243258014321327, 0.22770646214485168, -0.10479787737131119, -0.08809807151556015, 0.2632525563240051, -0.05423165112733841, -0.16432519257068634, 0.10179096460342407, -0.014350244775414467, 0.12198644131422043, 0.13850919902324677, 0.16080057621002197, 0.007628654129803181, 0.03313867375254631, 0.10115300863981247, 0.08631709218025208, -0.08573295921087265, -0.0611947737634182, 0.023627014830708504, -0.011463395319879055, -0.10670105367898941, 0.046802595257759094, 0.04794782027602196, 0.08188598603010178, -0.04982871189713478, -0.028600862249732018, -0.01972118206322193, -0.044152840971946716, 0.05264130234718323, 0.007675500120967627, 0.13217447698116302, -0.03674980252981186, -0.03692879155278206, -0.023745311424136162, 0.01699630729854107, -0.03115241602063179, 0.007061392068862915, -0.05687357112765312, 0.11091547459363937, -0.03406180441379547, 0.051789235323667526, -0.16953988373279572, -0.04873261600732803, -0.02087729424238205, 0.1402055323123932, 0.04973345249891281, 0.1329866498708725, 0.06287940591573715, -0.010758201591670513, 0.00859389640390873, 0.007998145185410976, 0.13181665539741516, 0.007865442894399166, -0.07660657912492752, -0.047718439251184464, 0.09176599979400635, -0.05973208695650101, 0.06147782504558563, -0.098741315305233, -0.004747362341731787, -0.01433002483099699, 0.08674649894237518, 0.006352655589580536, 0.029382232576608658, -0.006192679051309824, 0.003654100699350238, -0.06161240115761757, 0.017873648554086685, 0.12492607533931732, -0.01421504095196724, -0.07439801841974258, 0.22084392607212067, -0.15798072516918182, 0.18006981909275055, 0.18165533244609833, -0.3081994652748108, 0.024602634832262993, -0.08860466629266739, -0.036338552832603455, 0.03426366671919823, 0.0491504967212677, -0.034147560596466064, 0.16587987542152405, -0.016766328364610672, 0.201018825173378, -0.03547777235507965, -0.01287798210978508, -0.010399105958640575, -0.03656993433833122, -0.010632630437612534, 0.09065473079681396, 0.15122920274734497, -0.1677125245332718, 0.18270380795001984, 0.1660280078649521, 0.06873020529747009, 0.17776396870613098, 0.034313347190618515, -0.006856906693428755, 0.07112615555524826, -0.022670727223157883, -0.07675548642873764, -0.049287427216768265, -0.26302891969680786, -0.027947327122092247, 0.06471601128578186, 0.04510856419801712, 0.11924877762794495, -0.10971947014331818, -0.037208184599876404, 0.010892451740801334, -0.013165894895792007, 0.02132410928606987, 0.09682225435972214, 0.01171150617301464, 0.11804302036762238, -0.021027036011219025, -0.05209195241332054, 0.0898953229188919, 0.02727191150188446, -0.0787680521607399, 0.19168277084827423, -0.10074768215417862, -0.3233809769153595, -0.11354339867830276, -0.18166927993297577, -0.017843691632151604, 0.05878754332661629, 0.08049646019935608, -0.09228580445051193, -0.02625267766416073, -0.01639235019683838, 0.0758359357714653, -0.09145816415548325, -0.015880629420280457, -0.09367848187685013, 0.034986745566129684, -0.10827737301588058, -0.07011983543634415, -0.05141967162489891, -0.03368452936410904, -0.04457031562924385, 0.13157756626605988, -0.12242637574672699, 0.06396433711051941, 0.2076517641544342, 0.06227295100688934, 0.05622440204024315, -0.0229496993124485, 0.23288212716579437, -0.10842552781105042, 0.02383521944284439, 0.1717897206544876, -0.03566030040383339, 0.0727933868765831, 0.13435456156730652, 0.006721907295286655, -0.08144525438547134, 0.03465581312775612, -0.04592517390847206, -0.08630958944559097, -0.20441576838493347, -0.14156180620193481, -0.12814727425575256, 0.07913564145565033, 0.03285396471619606, 0.05478321388363838, 0.15024253726005554, 0.11386489123106003, 0.007987297140061855, 0.00976672861725092, -0.006888182368129492, 0.05438044294714928, 0.17482298612594604, -0.05838097631931305, 0.10041683167219162, -0.037591226398944855, -0.1924494504928589, 0.08022978901863098, 0.04309763014316559, 0.08280511945486069, 0.07474655658006668, 0.0856199786067009, 0.013537914492189884, 0.03723837807774544, 0.10897084325551987, 0.1165735274553299, 0.031679023057222366, -0.038079675287008286, -0.04882059991359711, -0.026300756260752678, -0.03285675123333931, 0.05745977535843849, 0.07790146768093109, -0.1608346849679947, -0.06348084658384323, -0.06350091099739075, 0.07662643492221832, 0.09017108380794525, 0.11811108142137527, -0.21219493448734283, 0.01579318381845951, 0.092556893825531, -0.0494147390127182, -0.1304239183664322, 0.07402537018060684, -0.00466050673276186, -0.1397053301334381, 0.037663187831640244, -0.014095795340836048, 0.1359514445066452, -0.0778401643037796, 0.10336452722549438, -0.08307972550392151, -0.06147889420390129, 0.03632286190986633, 0.1355396956205368, -0.30774354934692383, 0.2137020230293274, -0.022472934797406197, -0.05296783149242401, -0.10508129745721817, -0.011727629229426384, 0.020913105458021164, 0.09079049527645111, 0.10090240091085434, -0.0025442070327699184, 0.0061299679800868034, -0.0345483273267746, -0.053218815475702286, 0.024456629529595375, 0.07957815378904343, -0.08542889356613159, 0.0017540202243253589, -0.02361489273607731, -0.004407065454870462, -0.032844748347997665, -0.01189463958144188, -0.011617658659815788, -0.16786961257457733, 0.06556065380573273, -0.002625665394589305, 0.11129079759120941, 0.03491498529911041, 0.0024013579823076725, -0.1009332686662674, 0.19977013766765594, 0.01796281896531582, -0.08052749931812286, -0.08830537647008896, -0.03254766762256622, 0.03660419583320618, -0.06121435388922691, 0.027481911703944206, -0.06916457414627075, 0.033381566405296326, -0.06441576033830643, -0.18325145542621613, 0.1268530637025833, -0.10945470631122589, -0.03609596937894821, -0.04321056231856346, 0.18323224782943726, -0.00929707009345293, -0.0011623724130913615, 0.05866571143269539, 0.0032208464108407497, -0.1347510665655136, -0.10740556567907333, 0.020214511081576347, -0.015275230631232262, 0.009142245166003704, 0.05559912323951721, -0.009665844030678272, 0.00045268211397342384, -0.039558928459882736, -0.023234419524669647, 0.32348164916038513, 0.10732097923755646, -0.04944206401705742, 0.17007054388523102, 0.13087597489356995, -0.0827672928571701, -0.30699312686920166, -0.10971353948116302, -0.10529600828886032, -0.026918673887848854, -0.037983208894729614, -0.19617970287799835, 0.09504909813404083, -0.03528566658496857, -0.022136637941002846, 0.11253651231527328, -0.2759084105491638, -0.0770430713891983, 0.1826775223016739, 0.003314757253974676, 0.3998824954032898, -0.10265109688043594, -0.08777514100074768, -0.06741699576377869, -0.1120782196521759, 0.2033512443304062, -0.05560711398720741, 0.08663415163755417, -0.00517998356372118, 0.15513743460178375, 0.055607251822948456, -0.02176513522863388, 0.08932057023048401, -0.005811662413179874, -0.0546204075217247, -0.1219351515173912, -0.03444604203104973, -0.009159418754279613, 0.007239421829581261, 0.03589896112680435, -0.04242607578635216, 0.01279151439666748, -0.1399589478969574, -0.045490626245737076, -0.0764620453119278, 0.024699507281184196, 0.021008269861340523, -0.0652410089969635, -0.01643640361726284, -0.03945036977529526, -0.012804778292775154, 0.03164318576455116, 0.15236099064350128, -0.06478006392717361, 0.1476556956768036, 0.04904455319046974, 0.15412139892578125, -0.14745712280273438, -0.02258288487792015, -0.06896031647920609, -0.05498642474412918, 0.04900865629315376, -0.10053684562444687, 0.050061121582984924, 0.1202658861875534, -0.0742902010679245, 0.0987328365445137, 0.0922594666481018, -0.01938629150390625, 0.0012483424507081509, 0.1226617842912674, -0.2489612102508545, -0.07742628455162048, -0.10509459674358368, 0.013337249867618084, 0.10138551890850067, 0.06995654851198196, 0.17304721474647522, -0.0037713919300585985, -0.036284226924180984, -0.0064643872901797295, 0.025414984673261642, -0.03540204465389252, 0.05724727362394333, -0.002706433180719614, 0.016663886606693268, -0.15213344991207123, 0.060368724167346954, -0.00024176653823815286, -0.1438901126384735, -0.013603870756924152, 0.16073721647262573, -0.11208858340978622, -0.15145981311798096, -0.007263668347150087, 0.13685113191604614, -0.13171035051345825, -0.03302847594022751, -0.03708777576684952, -0.170182466506958, 0.07439173012971878, 0.1024777740240097, 0.08549231290817261, 0.08025266975164413, -0.06620611250400543, -0.00807863101363182, -0.011656313203275204, -0.026087598875164986, 0.031810320913791656, -0.023377234116196632, -0.09044221043586731, 0.03872343525290489, -0.026654237881302834, 0.13591371476650238, -0.09607382118701935, -0.09331836551427841, -0.135749951004982, 0.039314381778240204, -0.12405620515346527, -0.08138058334589005, -0.12200927734375, -0.0591500885784626, 0.00224387738853693, -0.0001289021165575832, -0.035674065351486206, -0.06687422841787338, -0.13582271337509155, 0.04366770386695862, -0.04484611004590988, 0.0013091047294437885, -0.040241483598947525, 0.04561002552509308, 0.06766383349895477, -0.03493715822696686, 0.13722217082977295, 0.11722734570503235, -0.07864081114530563, 0.08946478366851807, -0.16657429933547974, -0.0683990865945816, 0.08854512125253677, 0.008173754438757896, 0.06165994703769684, 0.06743349134922028, 0.033807408064603806, 0.06109451875090599, 0.04151686280965805, 0.03488299250602722, 0.01739438995718956, -0.09271225333213806, 0.015541021712124348, 0.022296719253063202, -0.1294609159231186, -0.04801803454756737, -0.029226921498775482, 0.00939185917377472, 0.008117396384477615, 0.11003357172012329, -0.0426274873316288, 0.09439733624458313, -0.05888751894235611, 0.036728594452142715, 0.016222506761550903, -0.16461637616157532, -0.020102784037590027, -0.11915475130081177, 0.028684545308351517, -0.0033096212428063154, 0.25625869631767273, 0.06346847862005234, 0.020517030730843544, 0.01250078622251749, 0.08567021042108536, 0.07241600006818771, 0.02562166005373001, 0.1956365555524826, 0.10854171961545944, -0.05020022392272949, -0.12334850430488586, 0.09686340391635895, 0.034720368683338165, 0.06432123482227325, 0.13385434448719025, -0.026959087699651718, 0.002498799469321966, 0.11019360274076462, 0.011678861454129219, 0.04961980879306793, -0.09859088063240051, -0.16400282084941864, -0.00994415208697319, 0.061864156275987625, -0.04559077322483063, 0.12240655720233917, 0.11382720619440079, -0.020697353407740593, 0.03180128335952759, -0.010503606870770454, -0.05694027617573738, -0.16998925805091858, -0.1630837321281433, -0.08357038348913193, -0.11794789135456085, -0.0027763545513153076, -0.11386270076036453, 0.013879159465432167, 0.06452289968729019, 0.0604364387691021, -0.09019444137811661, 0.08891061693429947, 0.0687386617064476, -0.11843101680278778, 0.08828350901603699, -0.033263903111219406, 0.07249268144369125, 0.0015160300536081195, 0.003872724948450923, -0.13800905644893646, 0.032393742352724075, -0.008493867702782154, 0.04159298539161682, -0.09244006127119064, 0.022458361461758614, -0.11297028511762619, -0.07659684121608734, -0.07971972227096558, 0.05093973129987717, -0.03541257977485657, 0.1390930563211441, 0.001295371213927865, -0.035233911126852036, 0.024190181866288185, 0.22729112207889557, -0.06350252777338028, -0.030667411163449287, -0.0618741400539875, 0.21414142847061157, 0.024466563016176224, 0.10703565180301666, -0.016775688156485558, 0.019240234047174454, -0.0764411985874176, 0.3689337372779846, 0.344390869140625, -0.1225387305021286, -0.0015968306688591838, 0.031062176451086998, 0.036916591227054596, 0.11621878296136856, 0.12602226436138153, 0.057955991476774216, 0.2995031177997589, -0.08396036922931671, -0.002026971662417054, -0.02688612788915634, -0.03624163940548897, -0.04409930482506752, 0.10547586530447006, 0.06835740804672241, -0.03330419585108757, -0.027012333273887634, 0.1376710683107376, -0.2966996431350708, 0.12323499470949173, -0.15714547038078308, -0.1487535685300827, -0.06873904913663864, -0.005042468197643757, 0.08589684963226318, 0.04748665541410446, 0.1069009080529213, -0.019124338403344154, -0.08203735202550888, 0.05766449123620987, 0.0320524163544178, -0.22844897210597992, 0.011852608993649483, 0.08361081779003143, -0.06153005734086037, 0.011767351068556309, -0.017906347289681435, 0.038472190499305725, 0.07790610194206238, 0.025976579636335373, -0.032770540565252304, 0.06325861811637878, -0.005814229138195515, -0.05033424496650696, 0.04302205145359039, 0.05059972032904625, 0.017107632011175156, -0.1511564701795578, 0.07320158183574677, -0.1762860119342804, 0.0566408596932888, -0.005331212189048529, -0.04948166385293007, 0.000018263708625454456, 0.01998119056224823, -0.06808236241340637, 0.05880929157137871, 0.0952666699886322, -0.012173139490187168, -0.002317852806299925, -0.056667573750019073, 0.007662574760615826, -0.0679154172539711, -0.0747012197971344, -0.10497893393039703, -0.1338900774717331, -0.11392296850681305, 0.10846775025129318, -0.011928223073482513, -0.19833622872829437, 0.02906924858689308, -0.11258108913898468, 0.04933213070034981, -0.13360801339149475, 0.08599711954593658, 0.1282832771539688, 0.021543797105550766, -0.01265349704772234, 0.04020093381404877, 0.01591683179140091, 0.08550478518009186, -0.09200563281774521, -0.10515180230140686 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-indonesian-1.5G-finetuned-sentiment-analysis-smsa This model is a fine-tuned version of [cahya/bert-base-indonesian-1.5G](https://huggingface.co/cahya/bert-base-indonesian-1.5G) on the indonlu dataset. It achieves the following results on the evaluation set: - Loss: 0.3390 - Accuracy: 0.9373 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.2864 | 1.0 | 688 | 0.2154 | 0.9286 | | 0.1648 | 2.0 | 1376 | 0.2238 | 0.9357 | | 0.0759 | 3.0 | 2064 | 0.3351 | 0.9365 | | 0.044 | 4.0 | 2752 | 0.3390 | 0.9373 | | 0.0308 | 5.0 | 3440 | 0.4346 | 0.9365 | | 0.0113 | 6.0 | 4128 | 0.4708 | 0.9365 | | 0.006 | 7.0 | 4816 | 0.5533 | 0.9325 | | 0.0047 | 8.0 | 5504 | 0.5888 | 0.9310 | | 0.0001 | 9.0 | 6192 | 0.5961 | 0.9333 | | 0.0 | 10.0 | 6880 | 0.5992 | 0.9357 | ### Framework versions - Transformers 4.14.1 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"language": "id", "license": "mit", "tags": ["generated_from_trainer"], "datasets": ["indonlu"], "metrics": ["accuracy"], "widget": [{"text": "Saya mengapresiasi usaha anda"}], "model-index": [{"name": "bert-base-indonesian-1.5G-finetuned-sentiment-analysis-smsa", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "indonlu", "type": "indonlu", "args": "smsa"}, "metrics": [{"type": "accuracy", "value": 0.9373015873015873, "name": "Accuracy"}]}]}]}
text-classification
ayameRushia/bert-base-indonesian-1.5G-sentiment-analysis-smsa
[ "transformers", "pytorch", "bert", "text-classification", "generated_from_trainer", "id", "dataset:indonlu", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
bert-base-indonesian-1.5G-finetuned-sentiment-analysis-smsa =========================================================== This model is a fine-tuned version of cahya/bert-base-indonesian-1.5G on the indonlu dataset. It achieves the following results on the evaluation set: * Loss: 0.3390 * Accuracy: 0.9373 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.14.1 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 61, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.11437919735908508, 0.06797140091657639, -0.0017803250811994076, 0.1313408762216568, 0.18969780206680298, 0.045225195586681366, 0.1362580507993698, 0.10920530557632446, -0.07014871388673782, 0.009777862578630447, 0.12498773634433746, 0.17389145493507385, 0.007851325906813145, 0.09938061982393265, -0.0523936003446579, -0.2635502815246582, -0.01372514944523573, 0.04697141796350479, -0.05414706468582153, 0.13845610618591309, 0.09473682940006256, -0.1464952826499939, 0.09721025824546814, 0.0019515258027240634, -0.2133648544549942, 0.004483429249376059, 0.017443262040615082, -0.054060954600572586, 0.15972307324409485, 0.023213006556034088, 0.1337159276008606, 0.01012072991579771, 0.0966791957616806, -0.17856422066688538, 0.009487507864832878, 0.032163508236408234, 0.005944873206317425, 0.08956585824489594, 0.044238943606615067, -0.01384176965802908, 0.11574035882949829, -0.069818876683712, 0.05668064206838608, 0.019715778529644012, -0.14098875224590302, -0.20680473744869232, -0.06997211277484894, 0.046056486666202545, 0.05135658383369446, 0.1009083017706871, -0.0062061031349003315, 0.13046133518218994, -0.11695049703121185, 0.0912955179810524, 0.2130535989999771, -0.2713475227355957, -0.05953285098075867, 0.037731003016233444, -0.003766547655686736, 0.06628616154193878, -0.11394321918487549, -0.02361641824245453, 0.05416123941540718, 0.044448141008615494, 0.12632432579994202, -0.029906075447797775, -0.09579635411500931, 0.03203408420085907, -0.14053350687026978, -0.03153502941131592, 0.14652016758918762, 0.03934447467327118, -0.034379325807094574, -0.029420984908938408, -0.042246051132678986, -0.15309515595436096, -0.038650039583444595, -0.011788635514676571, 0.041852161288261414, -0.04019995033740997, -0.08007272332906723, 0.014839150942862034, -0.09945571422576904, -0.07184206694364548, -0.07790681719779968, 0.16117912530899048, 0.03955147787928581, 0.004469555802643299, -0.027012227103114128, 0.10645553469657898, 0.007590318564325571, -0.12227090448141098, 0.018495630472898483, 0.02446342632174492, 0.008937285281717777, -0.06265106797218323, -0.058459196239709854, -0.018479488790035248, 0.0015655782772228122, 0.11186569184064865, -0.05391763523221016, 0.027538815513253212, 0.0556572861969471, 0.03670341521501541, -0.09041165560483932, 0.2013523429632187, -0.039106808602809906, -0.00828644260764122, -0.009878856129944324, 0.04607759788632393, -0.011856374330818653, -0.009770538657903671, -0.12824289500713348, -0.007137895096093416, 0.09858786314725876, 0.00864447932690382, -0.08777341991662979, 0.08115801960229874, -0.06303024291992188, -0.03349719196557999, 0.010060069151222706, -0.08325374126434326, 0.03709491342306137, 0.00549157103523612, -0.08772461116313934, -0.02013300359249115, 0.015502581372857094, 0.014687765389680862, -0.009707312099635601, 0.11869394034147263, -0.0980023741722107, 0.035536862909793854, -0.09079638123512268, -0.1261352151632309, 0.004782302305102348, -0.09640131145715714, 0.03684055432677269, -0.10217949002981186, -0.17225047945976257, -0.008222455158829689, 0.03983008489012718, -0.025162089616060257, -0.04717505723237991, -0.06960278004407883, -0.06399761885404587, 0.0053876470774412155, -0.0060966238379478455, 0.09896785765886307, -0.06331566721200943, 0.11226058006286621, 0.0404488742351532, 0.06446892023086548, -0.04511190205812454, 0.06393086165189743, -0.10963074117898941, -0.004476584494113922, -0.17303729057312012, 0.05364120751619339, -0.04174020513892174, 0.07028072327375412, -0.06971371173858643, -0.11766982078552246, 0.02162964642047882, 0.008141394704580307, 0.0725998654961586, 0.12351477891206741, -0.16281791031360626, -0.09427018463611603, 0.15761107206344604, -0.062138184905052185, -0.12747804820537567, 0.12281015515327454, -0.07579197734594345, 0.05483599379658699, 0.08318115025758743, 0.1704866886138916, 0.07279513776302338, -0.07612241059541702, 0.017427721992135048, 0.0054123797453939915, 0.05258834362030029, -0.07175204902887344, 0.07210035622119904, 0.009502907283604145, -0.00978475995361805, 0.02593676745891571, -0.022812720388174057, 0.06912249326705933, -0.12138855457305908, -0.09114952385425568, -0.023071164265275, -0.09814003109931946, 0.08262485265731812, 0.077518992125988, 0.07549212872982025, -0.10492252558469772, -0.06454932689666748, 0.0873878002166748, 0.08278781175613403, -0.04612951725721359, 0.012188167311251163, -0.06716324388980865, 0.06870689243078232, -0.02577146328985691, -0.03283550217747688, -0.17400328814983368, -0.014570881612598896, 0.0008729764958843589, 0.05241499841213226, 0.026254849508404732, 0.02588050067424774, 0.06747499108314514, 0.07134616374969482, -0.05056526139378548, -0.01738881692290306, -0.031511787325143814, 0.0013978749047964811, -0.1406737118959427, -0.20439034700393677, -0.032795581966638565, -0.007505808956921101, 0.14837124943733215, -0.23582583665847778, 0.0435955710709095, -0.03620235249400139, 0.06415752321481705, 0.003235745010897517, -0.00999328400939703, -0.05153840035200119, 0.08924543857574463, -0.03638185188174248, -0.0383467935025692, 0.07667195051908493, -0.004085196182131767, -0.08828277140855789, -0.04510101303458214, -0.08434154838323593, 0.1937151998281479, 0.13703680038452148, -0.1508449763059616, -0.08536606281995773, -0.015962619334459305, -0.05082867294549942, -0.01864619180560112, -0.04997849836945534, 0.03774787485599518, 0.1860833466053009, -0.015470478683710098, 0.15167467296123505, -0.06367035210132599, -0.03229161724448204, 0.021149562671780586, -0.03757014498114586, 0.032637178897857666, 0.14168043434619904, 0.12959280610084534, -0.09564905613660812, 0.14557486772537231, 0.1378016471862793, -0.06594858318567276, 0.1338530033826828, -0.032235242426395416, -0.06397013366222382, -0.015309627167880535, -0.04601798579096794, -0.011823784559965134, 0.10401257872581482, -0.1689993292093277, -0.012620065361261368, 0.01988374814391136, 0.01868641935288906, 0.008482794277369976, -0.21990631520748138, -0.05960991233587265, 0.050299499183893204, -0.03254597261548042, -0.060893137007951736, -0.016634073108434677, 0.016161229461431503, 0.11159331351518631, 0.005525727290660143, -0.09758256375789642, 0.033144496381282806, 0.011851971037685871, -0.08623563498258591, 0.21900807321071625, -0.07539169490337372, -0.13614267110824585, -0.122889444231987, -0.10176458954811096, -0.04259994626045227, 0.018373334780335426, 0.059686992317438126, -0.10381487756967545, -0.012919158674776554, -0.050306446850299835, 0.04099142551422119, -0.01758710853755474, 0.03199102729558945, -0.0037160806823521852, -0.009691337123513222, 0.048715803772211075, -0.10885617882013321, -0.017292004078626633, -0.06688785552978516, -0.06686023622751236, 0.057994574308395386, 0.009421265684068203, 0.11795321106910706, 0.15487511456012726, -0.030839409679174423, 0.012592026963829994, -0.03962014243006706, 0.2579967975616455, -0.06274687498807907, -0.03786233067512512, 0.1325155645608902, -0.0003670032601803541, 0.04599896818399429, 0.10898227244615555, 0.08291681855916977, -0.08982774615287781, -0.000023188653358374722, 0.02886175364255905, -0.03259105607867241, -0.2154148370027542, -0.049625515937805176, -0.05548667907714844, -0.012736564502120018, 0.0960804745554924, 0.009846468456089497, 0.04382161423563957, 0.07883822172880173, 0.03737499564886093, 0.09121455252170563, -0.047569338232278824, 0.05148761346936226, 0.12966680526733398, 0.04205210134387016, 0.12664872407913208, -0.03736360743641853, -0.07643676549196243, 0.0374041385948658, -0.04851878434419632, 0.2190418839454651, 0.011631463654339314, 0.11346438527107239, 0.06080015376210213, 0.17857249081134796, 0.0021701729856431484, 0.09047948569059372, -0.0008652216056361794, -0.047843270003795624, -0.015631118789315224, -0.03311421349644661, -0.04737458750605583, 0.006895691622048616, -0.05443640053272247, 0.0632738396525383, -0.15400953590869904, -0.002512563718482852, 0.04810174182057381, 0.21536017954349518, 0.048138827085494995, -0.34049081802368164, -0.1075809970498085, -0.009681102819740772, -0.018246328458189964, -0.019830802455544472, 0.018529126420617104, 0.08869026601314545, -0.10065660625696182, 0.016272960230708122, -0.04614153876900673, 0.09955698251724243, -0.05199653282761574, 0.05726909264922142, 0.0569981224834919, 0.08934405446052551, -0.014802565798163414, 0.09071394056081772, -0.31196656823158264, 0.2878953516483307, 0.0027864291332662106, 0.06337723135948181, -0.0759933665394783, -0.0264337956905365, 0.038675546646118164, 0.10367836803197861, 0.05083133652806282, 0.00018551136599853635, 0.033929940313100815, -0.2437017560005188, -0.03664225712418556, 0.029058368876576424, 0.08902566879987717, -0.04888519272208214, 0.09593656659126282, -0.028538376092910767, 0.01645279861986637, 0.07946937531232834, 0.005865037441253662, -0.05324817821383476, -0.08393705636262894, -0.02727578952908516, 0.022728141397237778, -0.0229099839925766, -0.050871483981609344, -0.12119316309690475, -0.13616147637367249, 0.12163642048835754, -0.004485507030040026, -0.03254146873950958, -0.11134501546621323, 0.09046109765768051, 0.06570203602313995, -0.09761437773704529, 0.033831432461738586, 0.013019134290516376, 0.055223122239112854, 0.031218141317367554, -0.05959930643439293, 0.11262389272451401, -0.06534737348556519, -0.1563425213098526, -0.06035241857171059, 0.09273605793714523, 0.05156998708844185, 0.07364534586668015, -0.002657379722222686, 0.01059792935848236, -0.04048383608460426, -0.08961201459169388, 0.02552170306444168, -0.024207059293985367, 0.0669078528881073, 0.04197079315781593, -0.060009781271219254, 0.028836680576205254, -0.07350560277700424, -0.031463079154491425, 0.1929205358028412, 0.23363317549228668, -0.09766595810651779, 0.010596156120300293, 0.03088982403278351, -0.06832654029130936, -0.19523797929286957, 0.05184559524059296, 0.045763324946165085, 0.011673952452838421, 0.04283665120601654, -0.19065089523792267, 0.12816642224788666, 0.09520486742258072, 0.002248651348054409, 0.09201256930828094, -0.31056952476501465, -0.12421751767396927, 0.13684846460819244, 0.14597852528095245, 0.15046624839305878, -0.12215202301740646, -0.013906549662351608, -0.01768452674150467, -0.1362585723400116, 0.10213522613048553, -0.06513334065675735, 0.12057960033416748, -0.0367535762488842, 0.09689285606145859, 0.014307254925370216, -0.045102525502443314, 0.11099769175052643, 0.04207918792963028, 0.10818712413311005, -0.06287328898906708, -0.06754717975854874, 0.024683916941285133, -0.026354368776082993, 0.011652042157948017, -0.07454992830753326, 0.030605843290686607, -0.1253519207239151, -0.02466638945043087, -0.08063746988773346, 0.037717945873737335, -0.039393629878759384, -0.07050444930791855, -0.033052414655685425, 0.028820941224694252, 0.04128763824701309, -0.01435921061784029, 0.11406280845403671, 0.011489692144095898, 0.15164221823215485, 0.05656006187200546, 0.08507762104272842, -0.08005490899085999, -0.04572364687919617, -0.02240862511098385, -0.01409920398145914, 0.052393268793821335, -0.12641815841197968, 0.02157055027782917, 0.14501500129699707, 0.03703506290912628, 0.14755408465862274, 0.0963335782289505, -0.01135854423046112, 0.0064443438313901424, 0.063010573387146, -0.1489442139863968, -0.07728391885757446, -0.02080993540585041, -0.08766984194517136, -0.10797877609729767, 0.05770936980843544, 0.08041982352733612, -0.07797475159168243, -0.011834751814603806, -0.02344178408384323, -0.007122493814677, -0.06108131632208824, 0.18630456924438477, 0.07378822565078735, 0.052718162536621094, -0.11026734113693237, 0.05036572366952896, 0.05231248587369919, -0.042986176908016205, -0.003214402124285698, 0.05910845845937729, -0.08782307803630829, -0.04496663808822632, 0.07602625340223312, 0.21408221125602722, -0.09034015238285065, -0.032398976385593414, -0.13737286627292633, -0.13181474804878235, 0.08348427712917328, 0.1673954874277115, 0.12217919528484344, 0.00910865142941475, -0.0651354193687439, 0.0037545980885624886, -0.13038142025470734, 0.07903517782688141, 0.04094073176383972, 0.060051079839468, -0.13987499475479126, 0.18674486875534058, 0.0011728409444913268, 0.04697924852371216, -0.022547829896211624, 0.011044749058783054, -0.1174079030752182, 0.021644815802574158, -0.1203419491648674, -0.039493415504693985, -0.02979770302772522, 0.017792263999581337, -0.011545209214091301, -0.0739303007721901, -0.04778070002794266, 0.0007082734373398125, -0.12072592228651047, -0.009807789698243141, 0.048098791390657425, 0.07193911075592041, -0.12081784754991531, -0.04192783683538437, 0.02754855901002884, -0.05710435286164284, 0.07047338783740997, 0.06314237415790558, 0.016778001561760902, 0.06268180906772614, -0.14807529747486115, 0.010686679743230343, 0.05492106452584267, 0.018160324543714523, 0.08201800286769867, -0.07284790277481079, 0.0019503332441672683, -0.009292243979871273, 0.05958718806505203, 0.03528088331222534, 0.06370877474546432, -0.12449836730957031, 0.005118131171911955, -0.016286738216876984, -0.08725032210350037, -0.06338071078062057, 0.045326292514801025, 0.0832536593079567, 0.005644085351377726, 0.19372235238552094, -0.07698048651218414, 0.04677975922822952, -0.2195456176996231, -0.004884114023298025, -0.012928634881973267, -0.10798226296901703, -0.13133691251277924, -0.0919283926486969, 0.06871449202299118, -0.046729762107133865, 0.14352165162563324, 0.046664729714393616, 0.03908780589699745, 0.016572263091802597, -0.0005770448478870094, 0.01434333436191082, 0.005876111797988415, 0.21506145596504211, 0.04757721722126007, -0.04735938832163811, 0.06457147002220154, 0.06505545973777771, 0.10754632204771042, 0.13100139796733856, 0.19877751171588898, 0.15235662460327148, -0.02371305227279663, 0.07316675037145615, 0.04033895209431648, -0.05379163473844528, -0.1483495831489563, -0.00609747227281332, -0.027810312807559967, 0.07472680509090424, -0.022769445553421974, 0.22673547267913818, 0.048211753368377686, -0.16909784078598022, 0.04918651282787323, -0.06254316866397858, -0.09058781713247299, -0.11070458590984344, -0.025342391803860664, -0.07730193436145782, -0.1466948539018631, 0.0023892477620393038, -0.11774937063455582, 0.008173082023859024, 0.12027861922979355, 0.008319029584527016, -0.027023538947105408, 0.1418205052614212, 0.017337385565042496, 0.033317867666482925, 0.05840848758816719, -0.007387218531221151, -0.024886196479201317, -0.13279017806053162, -0.06516636908054352, -0.01309604849666357, -0.02154042012989521, 0.03153352066874504, -0.06922858953475952, -0.06628604233264923, 0.02798774279654026, -0.02401510812342167, -0.11169781535863876, 0.015263965353369713, 0.013073612004518509, 0.062369249761104584, 0.04004222899675369, 0.004008244723081589, 0.024463996291160583, -0.013355739414691925, 0.22222955524921417, -0.06678709387779236, -0.08183388411998749, -0.08658191561698914, 0.2779443860054016, 0.05783776193857193, -0.0055407313629984856, 0.0377630852162838, -0.06747760623693466, 0.015426243655383587, 0.26220712065696716, 0.21565859019756317, -0.103908471763134, -0.0021155523136258125, 0.0009401289862580597, -0.008501976728439331, 0.000012327022886893246, 0.1112707108259201, 0.1221834123134613, 0.03412386029958725, -0.09731612354516983, -0.026654286310076714, -0.06842509657144547, -0.006417467724531889, -0.013991745188832283, 0.06852120906114578, 0.06635019183158875, 0.002810099860653281, -0.05085223540663719, 0.06176890805363655, -0.08055490255355835, -0.0889926329255104, 0.0736009031534195, -0.21164865791797638, -0.15031825006008148, -0.03306252136826515, 0.07298312336206436, 0.025541961193084717, 0.07597694545984268, -0.02980637550354004, -0.00043848546920344234, 0.05886496976017952, -0.010680390521883965, -0.10357999056577682, -0.08545628935098648, 0.10888604074716568, -0.07532694190740585, 0.19351983070373535, -0.04841740429401398, 0.07438857853412628, 0.12316121906042099, 0.07540282607078552, -0.05986727774143219, 0.059350576251745224, 0.03667879104614258, -0.05116529017686844, 0.022697653621435165, 0.05694343522191048, -0.030449923127889633, 0.04862082004547119, 0.03949059545993805, -0.15150243043899536, 0.03753200173377991, -0.062265582382678986, -0.07406900078058243, -0.04127918556332588, -0.04494939744472504, -0.044574011117219925, 0.12174176424741745, 0.22991731762886047, -0.024940090253949165, 0.0005543474107980728, -0.06803807616233826, 0.010574884712696075, 0.06735159456729889, 0.004793656524270773, -0.07920749485492706, -0.2178969383239746, -0.001983930356800556, 0.0670996904373169, -0.021952306851744652, -0.2617975175380707, -0.07957474142313004, -0.006695582531392574, -0.07343441992998123, -0.08582944422960281, 0.08186116814613342, 0.08090026676654816, 0.06119376793503761, -0.053941525518894196, -0.06447077542543411, -0.07613306492567062, 0.17025065422058105, -0.13869865238666534, -0.08842679113149643 ]
null
null
transformers
# Indonesian GPT-2-medium finetuned on Indonesian poems This is the [Indonesian gpt2-medium model](https://huggingface.co/flax-community/gpt2-medium-indonesian) fine-tuned to Indonesian poems. The dataset can be found in [here](https://huggingface.co/datasets/id_puisi) All training was done on Google Colab Jupyter Notebook (soon). The dataset is splitted into two subset with details belows: | split | count (examples) | percentage | | ---------- | ---------- | -------------- | | train | 7,358 | 80% | | validation | 1,890 | 20% | ### Evaluation results The model evaluation results after 10 epochs are as follows: | dataset | train/loss | eval/loss | eval perplexity | | ---------- | ---------- | -------------- | ---------- | | [id puisi](https://huggingface.co/datasets/id_puisi) | 3.104 | 3.384 | 29.4884 | The logs can be found in [wandb page here](https://wandb.ai/ayamerushia/gpt-2_poem/runs/3jsu1orj/overview?workspace=user-ayamerushia)
{"language": "id", "widget": [{"text": "Wahai rembulan yang tertutup awan hujan"}]}
text-generation
ayameRushia/gpt2-medium-fine-tuning-indonesia-poem
[ "transformers", "pytorch", "gpt2", "text-generation", "id", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Indonesian GPT-2-medium finetuned on Indonesian poems ===================================================== This is the Indonesian gpt2-medium model fine-tuned to Indonesian poems. The dataset can be found in here All training was done on Google Colab Jupyter Notebook (soon). The dataset is splitted into two subset with details belows: split: train, count (examples): 7,358, percentage: 80% split: validation, count (examples): 1,890, percentage: 20% ### Evaluation results The model evaluation results after 10 epochs are as follows: The logs can be found in wandb page here
[ "### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here" ]
[ 49, 31 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here" ]
[ -0.10117746889591217, 0.06658052653074265, -0.0006968708476051688, -0.00990425143390894, 0.13880082964897156, 0.023944832384586334, 0.12929512560367584, 0.09361468255519867, 0.013800856657326221, -0.038545507937669754, 0.14305365085601807, 0.12181886285543442, 0.006422610487788916, 0.12987573444843292, -0.09492801129817963, -0.2591094374656677, 0.0633668377995491, 0.0642278790473938, -0.0007437833119183779, 0.15105338394641876, 0.09026111662387848, -0.09822893142700195, 0.08229352533817291, 0.03585956618189812, -0.16685311496257782, -0.009849673137068748, -0.02075161412358284, -0.0877496749162674, 0.12647491693496704, 0.03605896607041359, 0.05977118760347366, 0.05191203951835632, -0.003026457503437996, -0.08407635241746902, 0.026113109663128853, -0.011937840841710567, 0.016784612089395523, 0.10102831572294235, 0.07623087614774704, 0.03851056471467018, 0.2029859870672226, 0.167047381401062, -0.017159054055809975, 0.04097798466682434, -0.14191307127475739, -0.0884615033864975, -0.05533887818455696, 0.030267590656876564, 0.13360603153705597, 0.04859006404876709, -0.03796888142824173, 0.18521931767463684, -0.12090126425027847, 0.0649976134300232, 0.1414162963628769, -0.2750057578086853, 0.003814300987869501, 0.180914968252182, 0.11513890326023102, -0.02464173175394535, -0.04498458281159401, 0.08436647802591324, 0.10393265634775162, 0.006568870972841978, 0.013045941479504108, -0.04888943210244179, 0.018643401563167572, 0.0353349931538105, -0.11813944578170776, -0.05867897719144821, 0.31875279545783997, 0.00041750905802473426, 0.0118205351755023, -0.07050982862710953, -0.06587549299001694, -0.10567886382341385, -0.012125266715884209, -0.07022324949502945, -0.029737690463662148, 0.025221774354577065, 0.07047294080257416, -0.025679951533675194, -0.10400240868330002, -0.07993440330028534, -0.13888664543628693, 0.14969205856323242, 0.007762464694678783, 0.03540826588869095, -0.21331891417503357, 0.15774023532867432, -0.04606609791517258, -0.10461236536502838, -0.02441106177866459, -0.10153506696224213, 0.0020106930751353502, -0.040599942207336426, -0.08150015771389008, 0.04059302434325218, 0.05092308670282364, 0.2050885409116745, -0.02217562310397625, -0.029029961675405502, 0.007766971364617348, 0.03972174599766731, 0.07164749503135681, 0.17671184241771698, -0.06444306671619415, -0.11610571295022964, 0.03156314417719841, -0.06738843023777008, -0.07506851106882095, -0.010739751160144806, -0.09745548665523529, -0.07287026941776276, 0.10825683176517487, 0.0934220403432846, -0.03262916952371597, 0.06670533120632172, -0.01617361046373844, -0.04305463656783104, 0.1448899209499359, -0.047524891793727875, -0.057885754853487015, -0.02048163115978241, -0.10172079503536224, 0.07983868569135666, -0.008909471333026886, 0.022566121071577072, -0.07261315733194351, 0.059183962643146515, -0.11147965490818024, -0.06386574357748032, -0.03255428746342659, -0.08884741365909576, 0.006263389252126217, -0.12875720858573914, 0.06727217137813568, -0.09899076819419861, -0.284720242023468, -0.022310815751552582, -0.024493839591741562, -0.03833182528614998, -0.03291996195912361, -0.033115606755018234, -0.04153544828295708, 0.0019502327777445316, -0.06089223548769951, 0.07057593762874603, -0.061660006642341614, 0.04694318026304245, 0.014975105412304401, 0.05306479334831238, -0.06913848966360092, 0.057897426187992096, -0.1490732878446579, -0.0013039074838161469, -0.04241720214486122, 0.046667974442243576, 0.03596062213182449, 0.10676790773868561, -0.05374163016676903, -0.0510082021355629, -0.0937020555138588, 0.04184476286172867, 0.003954668529331684, 0.24002236127853394, -0.12821996212005615, -0.06348571181297302, 0.19644400477409363, -0.08886779844760895, -0.20819252729415894, 0.09798750281333923, -0.0018901156727224588, 0.07600445300340652, 0.07936692982912064, 0.12672355771064758, 0.09356187283992767, 0.014426819048821926, 0.05047857016324997, 0.01784408465027809, -0.044779181480407715, -0.0977216362953186, 0.04184098541736603, 0.030086591839790344, -0.02805102989077568, 0.05509172007441521, -0.013473953120410442, 0.038518328219652176, -0.07752963155508041, -0.03576646372675896, -0.03552759811282158, -0.027158064767718315, 0.0006187504041008651, -0.0014888672158122063, 0.13632209599018097, -0.04352723807096481, -0.014324077405035496, -0.02099798619747162, 0.04521763324737549, -0.0029517170041799545, -0.015865508466959, -0.03668242692947388, 0.13945476710796356, -0.10947823524475098, 0.039738211780786514, -0.20645979046821594, -0.0379643514752388, 0.005374816246330738, 0.08955951780080795, -0.0459044985473156, 0.20914395153522491, 0.04825013130903244, -0.01960727944970131, 0.025242013856768608, -0.03468387573957443, 0.12358435988426208, -0.04819800332188606, -0.1335344761610031, -0.06258854269981384, 0.0568634457886219, -0.06530842930078506, 0.0000578274812141899, -0.12237841635942459, -0.0036704647354781628, -0.07654963433742523, 0.10311765968799591, -0.0500338152050972, 0.0067079514265060425, 0.023678351193666458, -0.027028003707528114, -0.1117594838142395, -0.0013786561321467161, 0.05480147525668144, -0.02217325195670128, -0.0971996933221817, 0.2288416475057602, -0.16760492324829102, 0.21906843781471252, 0.167263463139534, -0.1050654798746109, 0.019455881789326668, -0.06650857627391815, -0.043185554444789886, -0.016108771786093712, 0.002309751696884632, -0.03156778961420059, 0.20275236666202545, 0.01281729806214571, 0.16152936220169067, -0.0767650157213211, -0.08553551137447357, 0.00424224603921175, -0.030197717249393463, -0.006729097571223974, 0.09975327551364899, 0.09975441545248032, -0.16397884488105774, 0.09016340225934982, 0.16687479615211487, 0.06092560291290283, 0.1676945835351944, -0.013922308571636677, -0.08393779397010803, 0.040819115936756134, -0.05528300255537033, -0.08468308299779892, 0.04351567476987839, -0.1744040697813034, -0.020805159583687782, 0.09204426407814026, 0.0756625086069107, 0.11526977270841599, -0.13378798961639404, -0.023630565032362938, 0.008533366024494171, -0.0026002207305282354, -0.020226242020726204, 0.12786255776882172, 0.014619001187384129, 0.1269453763961792, -0.01798894815146923, -0.04943783953785896, -0.015802554786205292, 0.016020016744732857, -0.091328464448452, 0.23190180957317352, -0.046758245676755905, -0.30284491181373596, -0.13683182001113892, -0.051294684410095215, -0.04962817206978798, 0.04683510586619377, 0.05818545073270798, -0.10018707811832428, -0.05607219040393829, -0.005629779305309057, 0.15449243783950806, -0.06411130726337433, 0.04071367532014847, 0.09214673936367035, -0.0020647665951400995, -0.07154970616102219, -0.09834811836481094, -0.07836292684078217, -0.05720072612166405, 0.014825498685240746, 0.07156507670879364, -0.10212763398885727, 0.020574718713760376, 0.15106967091560364, 0.003979784436523914, 0.06650418043136597, 0.005846017971634865, 0.23508305847644806, -0.08970443904399872, -0.0021536084823310375, 0.24494569003582, -0.013752647675573826, 0.009764487855136395, 0.052434615790843964, -0.003906352911144495, -0.11415251344442368, 0.028471415862441063, -0.020295802503824234, -0.10285459458827972, -0.24100138247013092, -0.11660696566104889, -0.05432457476854324, 0.05155021697282791, 0.04319833591580391, 0.0072307586669921875, 0.12600375711917877, 0.10545798391103745, -0.04531377926468849, 0.01585575006902218, -0.03881137818098068, 0.06548004597425461, 0.11566641181707382, -0.014910237863659859, 0.1311163753271103, -0.0354795977473259, -0.04883814603090286, 0.10201425850391388, -0.03141346201300621, 0.11496774107217789, 0.06500464677810669, -0.03490401804447174, 0.03355806693434715, -0.014461874961853027, 0.08803165704011917, 0.15855272114276886, 0.0584968738257885, -0.05976267531514168, -0.030508020892739296, -0.007324691396206617, -0.04731176421046257, 0.05538352206349373, -0.06170390173792839, -0.10255605727434158, -0.10667788237333298, -0.012409461662173271, 0.050989262759685516, 0.03274912014603615, 0.15214964747428894, -0.3425038754940033, -0.07720824331045151, 0.04989299178123474, -0.055967967957258224, -0.11584047228097916, 0.09483972936868668, -0.0882764533162117, -0.13096769154071808, 0.04431888833642006, -0.047386135905981064, 0.15923656523227692, -0.07790036499500275, 0.07181702554225922, -0.06284245103597641, -0.04137672856450081, -0.013689344748854637, 0.07006474584341049, -0.2517396807670593, 0.2310003787279129, -0.025593746453523636, -0.037349577993154526, -0.07787414640188217, -0.005519252736121416, 0.0357191264629364, 0.16015508770942688, 0.08785779029130936, -0.0027648755349218845, 0.021298930048942566, 0.012836879119277, -0.06625358015298843, 0.04436555132269859, 0.049855124205350876, -0.1148897334933281, 0.06617220491170883, -0.0359729565680027, 0.04052511230111122, -0.019632840529084206, -0.0506005696952343, -0.03199927508831024, 0.010901196859776974, 0.047859009355306625, 0.04210246354341507, 0.06465647369623184, 0.008650539442896843, -0.0775052085518837, -0.1500309258699417, 0.2282600700855255, 0.02111007086932659, -0.09709690511226654, -0.10798384994268417, 0.07034188508987427, 0.18204128742218018, -0.05959732085466385, -0.004899180959910154, -0.06643997877836227, 0.06911429017782211, 0.010433422401547432, -0.14846161007881165, 0.07485929131507874, -0.13966229557991028, -0.08808138966560364, -0.0176012571901083, 0.1391814947128296, -0.046737607568502426, -0.010388923808932304, 0.03768837824463844, -0.00901192519813776, -0.13244697451591492, -0.1207796260714531, -0.03015981614589691, -0.02955801784992218, 0.05121982842683792, 0.022511180490255356, -0.03188280388712883, 0.0182805098593235, -0.026263708248734474, -0.0209032092243433, 0.16977925598621368, 0.17478568851947784, -0.02806093916296959, 0.06995762884616852, 0.1317472904920578, -0.06250153481960297, -0.24197682738304138, -0.031040918081998825, -0.06445535272359848, 0.05697048828005791, -0.03842119500041008, -0.08287402242422104, 0.12596499919891357, 0.04060421511530876, -0.005244446452707052, 0.10578694939613342, -0.23512674868106842, -0.11256472021341324, 0.21700234711170197, -0.00658103683963418, 0.3589533865451813, -0.08951298147439957, -0.05515652522444725, -0.020631296560168266, -0.2088642418384552, 0.07408253103494644, 0.0554887056350708, 0.038441747426986694, -0.06341546773910522, 0.14151862263679504, 0.05460713058710098, -0.07159021496772766, 0.12770412862300873, 0.008151719346642494, 0.015052689239382744, -0.07224267721176147, 0.0024161275941878557, 0.027813324704766273, -0.04164654389023781, 0.11617902666330338, -0.01575568877160549, 0.0779576301574707, -0.16623196005821228, -0.04707046225667, -0.06064601242542267, 0.04478248581290245, -0.005994599312543869, -0.09174470603466034, -0.0037363842129707336, -0.008554019033908844, -0.045534513890743256, -0.015788007527589798, 0.1324125975370407, -0.032529499381780624, 0.07473015785217285, -0.04550681263208389, 0.1273908019065857, -0.07928488403558731, 0.0175539031624794, -0.018639102578163147, 0.010532675310969353, 0.06447190046310425, -0.20159964263439178, 0.02474903129041195, 0.15855683386325836, -0.015826741233468056, 0.08741562813520432, 0.08072296530008316, -0.0025540124624967575, -0.023250728845596313, 0.11840866506099701, -0.2227233499288559, 0.04785531386733055, -0.07905712723731995, -0.15039746463298798, 0.03929724171757698, 0.1443280279636383, 0.16447493433952332, -0.054005157202482224, -0.016617469489574432, -0.0037663031835108995, 0.007878807373344898, -0.018056031316518784, 0.15614697337150574, -0.038635678589344025, 0.05443904921412468, -0.15556743741035461, 0.08525416254997253, -0.05703563988208771, -0.01986076682806015, -0.008269975893199444, 0.05858264118432999, -0.14850009977817535, -0.08695594221353531, -0.07318396866321564, 0.07801607996225357, -0.21807964146137238, -0.06538234651088715, -0.1083301454782486, -0.057681504637002945, 0.08535277843475342, 0.0827661082148552, 0.10320407897233963, 0.08722022920846939, -0.0644097700715065, -0.0909326896071434, -0.08357229083776474, 0.01868111453950405, 0.07331138104200363, 0.013896740972995758, -0.11008626222610474, 0.05168377235531807, -0.021695610135793686, 0.08792609721422195, -0.0884651392698288, -0.03874271735548973, -0.09555903822183609, 0.019384780898690224, -0.2009699046611786, -0.11222466081380844, -0.11526064574718475, -0.045698948204517365, -0.024651121348142624, -0.046466320753097534, -0.11386008560657501, -0.006551435217261314, -0.10446085035800934, 0.025746598839759827, -0.03698783367872238, 0.04759921133518219, -0.029678285121917725, 0.030978284776210785, 0.042501144111156464, -0.012119097635149956, 0.1258915662765503, 0.12120131403207779, -0.04449880123138428, 0.1150897890329361, -0.05734752491116524, 0.06626634299755096, 0.09922866523265839, 0.009089399129152298, 0.05541854351758957, -0.019261708483099937, 0.0357142873108387, 0.05969098582863808, 0.02019187994301319, 0.0741015374660492, -0.06999848783016205, -0.09110715985298157, -0.06464942544698715, -0.044061511754989624, -0.024068692699074745, -0.041249699890613556, -0.037105493247509, 0.048150647431612015, 0.09737461060285568, 0.10261420905590057, 0.03583329916000366, 0.05237238109111786, -0.11012952774763107, 0.017957070842385292, -0.004732741974294186, -0.14904899895191193, -0.017419105395674706, -0.05508509650826454, 0.052794843912124634, 0.005201998632401228, 0.31490084528923035, 0.02176336757838726, 0.1405998319387436, 0.022502684965729713, 0.15880247950553894, -0.0017650907393544912, -0.008119070902466774, 0.21215513348579407, 0.0995592400431633, -0.02570328861474991, -0.08174252510070801, 0.13250663876533508, 0.04097207263112068, -0.07743696123361588, 0.14401330053806305, 0.0011504899011924863, -0.07711702585220337, 0.10698825865983963, -0.0670156329870224, -0.032450344413518906, -0.1183587908744812, -0.12810996174812317, -0.0650467723608017, 0.0932103767991066, 0.04109949991106987, 0.09439710527658463, 0.17812326550483704, -0.05182620510458946, 0.018644897267222404, -0.0008857537759467959, -0.02624271810054779, -0.15081378817558289, -0.11458257585763931, -0.07988099008798599, -0.1053173765540123, 0.03780157491564751, -0.15208642184734344, 0.021717935800552368, 0.040034372359514236, 0.09640809148550034, -0.11609801650047302, 0.12106145173311234, 0.03399624675512314, -0.13052499294281006, 0.09925059974193573, -0.004779668990522623, 0.03037102147936821, -0.02703074924647808, 0.012305451557040215, -0.05690222233533859, 0.023211410269141197, 0.005926783196628094, -0.043494340032339096, -0.08898250013589859, 0.013142580166459084, -0.10128218680620193, -0.07574417442083359, -0.06893382966518402, 0.025001071393489838, -0.007528201676905155, 0.038484152406454086, 0.015691658481955528, -0.012470253743231297, -0.0008482423145323992, 0.2541476786136627, -0.06183423846960068, -0.032095931470394135, -0.11925660073757172, 0.18919837474822998, 0.07629066705703735, 0.036161743104457855, 0.008819255977869034, 0.007207984570413828, -0.04886733740568161, 0.321429967880249, 0.24520136415958405, -0.010388831607997417, -0.02168165147304535, 0.012219013646245003, 0.019038017839193344, 0.048636749386787415, 0.08543959259986877, 0.06996545940637589, 0.19005094468593597, -0.027756843715906143, -0.04032528027892113, -0.05862877517938614, -0.06069738417863846, 0.05084499344229698, 0.07877222448587418, 0.07597146183252335, -0.08244122564792633, -0.044275809079408646, 0.12951762974262238, -0.15420062839984894, 0.009439961053431034, -0.13752247393131256, -0.11661671102046967, -0.11739531904459, -0.01860811933875084, 0.016503725200891495, -0.011192478239536285, 0.0661657452583313, -0.0328223779797554, -0.06615559011697769, 0.03971310332417488, -0.02043589949607849, -0.1467374861240387, -0.06310657411813736, 0.0324423722922802, 0.04026283323764801, 0.056835249066352844, -0.0012724193511530757, 0.02422289364039898, 0.062120988965034485, 0.016724636778235435, -0.06201795116066933, 0.05386253073811531, 0.024124957621097565, -0.022493839263916016, 0.0778457522392273, 0.0782400518655777, 0.031992413103580475, -0.11575285345315933, 0.07372725754976273, -0.1664799451828003, 0.010533789172768593, -0.050348859280347824, -0.04439185559749603, -0.044911883771419525, 0.06682757288217545, -0.030965950340032578, 0.11900997161865234, 0.14261749386787415, -0.035237621515989304, -0.009487141855061054, -0.07901853322982788, 0.09218795597553253, -0.030826453119516373, -0.0040186685509979725, -0.031137261539697647, -0.15669485926628113, -0.05059518292546272, 0.1232447475194931, -0.06599677354097366, -0.23428651690483093, 0.021559659391641617, -0.07197689265012741, -0.02570679597556591, -0.06829077005386353, 0.10841508209705353, 0.061260320246219635, 0.027293477207422256, 0.004709625616669655, -0.005091623403131962, 0.019780190661549568, 0.08033320307731628, -0.1520712673664093, -0.13176406919956207 ]
null
null
transformers
# Indonesian GPT-2 finetuned on Indonesian poems This is the [Indonesian gpt2-small model](https://huggingface.co/flax-community/gpt2-small-indonesian) fine-tuned to Indonesian poems. The dataset can be found in [here](https://huggingface.co/datasets/id_puisi) All training was done on Google Colab Jupyter Notebook (soon). The dataset is splitted into two subset with details belows: | split | count (examples) | percentage | | ---------- | ---------- | -------------- | | train | 7,358 | 80% | | validation | 1,890 | 20% | ### Evaluation results The model evaluation results after 10 epochs are as follows: | dataset | train/loss | eval/loss | eval perplexity | | ---------- | ---------- | -------------- | ---------- | | [id puisi](https://huggingface.co/datasets/id_puisi) | 3.324700 | 3.502665 | 33.20 | The logs can be found in [wandb page here](https://wandb.ai/ayamerushia/gpt-2_poem/runs/36ymudz9/overview?workspace=user-ayamerushia) or tensorboard [here](https://huggingface.co/ayameRushia/gpt2-small-indonesia-fine-tuning-poem/tensorboard)
{"language": "id", "widget": [{"text": "Wahai rembulan yang tertutup awan hujan"}]}
text-generation
ayameRushia/gpt2-small-indonesia-fine-tuning-poem
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "id", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #tensorboard #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Indonesian GPT-2 finetuned on Indonesian poems ============================================== This is the Indonesian gpt2-small model fine-tuned to Indonesian poems. The dataset can be found in here All training was done on Google Colab Jupyter Notebook (soon). The dataset is splitted into two subset with details belows: split: train, count (examples): 7,358, percentage: 80% split: validation, count (examples): 1,890, percentage: 20% ### Evaluation results The model evaluation results after 10 epochs are as follows: The logs can be found in wandb page here or tensorboard here
[ "### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here or tensorboard here" ]
[ "TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here or tensorboard here" ]
[ 53, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #id #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Evaluation results\n\n\nThe model evaluation results after 10 epochs are as follows:\n\n\n\nThe logs can be found in wandb page here or tensorboard here" ]
[ -0.11473763734102249, 0.08347636461257935, -0.0009562394698150456, 0.015419983305037022, 0.13231243193149567, 0.011825288645923138, 0.1853383332490921, 0.11607883870601654, 0.025238260626792908, -0.0005190070951357484, 0.14725059270858765, 0.13851012289524078, 0.01770622283220291, 0.024345161393284798, -0.074005626142025, -0.26009440422058105, -0.0018005919409915805, 0.0835559070110321, -0.02480950951576233, 0.12923933565616608, 0.063873291015625, -0.11781938374042511, 0.06971319019794464, 0.05647486075758934, -0.18709667026996613, 0.003834386356174946, 0.0031435759738087654, -0.08123831450939178, 0.11660989373922348, 0.055093683302402496, 0.09720900654792786, 0.07001786679029465, 0.0037217866629362106, -0.06008148565888405, 0.03942427411675453, 0.02151079662144184, 0.002953430637717247, 0.13061802089214325, 0.11674090474843979, 0.011740196496248245, 0.15063633024692535, 0.11640693247318268, -0.020729495212435722, 0.023822089657187462, -0.10952439159154892, -0.07633395493030548, -0.06341145187616348, 0.02578490599989891, 0.11266198009252548, 0.03385147452354431, -0.044328782707452774, 0.1896507889032364, -0.06723655760288239, 0.07222101092338562, 0.1547517031431198, -0.33169639110565186, -0.0209721177816391, 0.1904374659061432, 0.10452251136302948, 0.005751187913119793, -0.058712031692266464, 0.10026086866855621, 0.09045348316431046, 0.005138513166457415, 0.06403603404760361, -0.04650561884045601, 0.02108796499669552, 0.045961908996105194, -0.13520710170269012, -0.0038940540980547667, 0.22884340584278107, -0.016199132427573204, 0.03504838049411774, -0.07994569838047028, -0.07140279561281204, -0.12034520506858826, -0.008150692097842693, -0.08683276176452637, -0.024891158565878868, 0.00873105600476265, 0.022469744086265564, -0.0967564508318901, -0.13787949085235596, -0.05440715700387955, -0.1614106297492981, 0.11722182482481003, 0.01295964140444994, 0.026888832449913025, -0.21333172917366028, 0.14725568890571594, -0.018857968971133232, -0.10556212067604065, 0.019199367612600327, -0.0662844106554985, -0.026296518743038177, -0.03959399461746216, -0.06255430728197098, -0.13463431596755981, 0.037118952721357346, 0.0923033356666565, 0.002380505669862032, -0.0029800073243677616, -0.02068217657506466, 0.08503391593694687, 0.054042842239141464, 0.1888899803161621, -0.08041843771934509, -0.07693363726139069, 0.03139333799481392, -0.05870741233229637, -0.03142638877034187, -0.019890600815415382, -0.12475626915693283, -0.04758540540933609, 0.09789063036441803, 0.06613894551992416, -0.03264576941728592, 0.06072168052196503, -0.016068514436483383, -0.05865204706788063, 0.10293995589017868, -0.050367146730422974, -0.021954238414764404, -0.013810867443680763, -0.09314084053039551, 0.09602192044258118, 0.024081815034151077, 0.002727547660470009, -0.07609182596206665, 0.057966720312833786, -0.11778595298528671, -0.024181442335247993, -0.04507105425000191, -0.10529216378927231, 0.022630495950579643, -0.16858667135238647, 0.038159020245075226, -0.09244076162576675, -0.21522264182567596, -0.013110486790537834, -0.012711583636701107, -0.05166369676589966, 0.0034195585176348686, -0.02176910638809204, -0.0308687686920166, 0.009815562516450882, -0.046781621873378754, 0.07713385671377182, -0.03544139489531517, 0.03184462711215019, -0.0028634376358240843, 0.07035844773054123, -0.05458888038992882, 0.06299144774675369, -0.10334508121013641, 0.00237692566588521, -0.0594053789973259, 0.05827135592699051, -0.005306868813931942, 0.06182298809289932, -0.07061655074357986, -0.04754599556326866, -0.12375908344984055, 0.03831906244158745, 0.040971145033836365, 0.20752586424350739, -0.16583849489688873, -0.0872182622551918, 0.20142893493175507, -0.07672593742609024, -0.1678701639175415, 0.12598682940006256, -0.022781725972890854, 0.03586313873529434, 0.06170662119984627, 0.14964964985847473, 0.11150078475475311, -0.02548941969871521, 0.04901692271232605, 0.04869769886136055, -0.03580671176314354, -0.1231951043009758, 0.026122257113456726, 0.023259039968252182, 0.01887667365372181, 0.04401320964097977, 0.01616680808365345, 0.0527302511036396, -0.06746216863393784, -0.03818992152810097, -0.038029663264751434, -0.02409045398235321, 0.03388112783432007, 0.034599658101797104, 0.17108206450939178, -0.03295368701219559, -0.032520391047000885, 0.03914603590965271, 0.011560522951185703, -0.04281884804368019, -0.013739261776208878, -0.04466469958424568, 0.1429903656244278, -0.15834708511829376, 0.01724294200539589, -0.2146533876657486, -0.10044204443693161, -0.00205442076548934, 0.05881670489907265, -0.013384230434894562, 0.24239425361156464, 0.07753386348485947, -0.0076889172196388245, 0.019878406077623367, 0.005959383212029934, 0.1344548761844635, -0.05289970338344574, -0.16750675439834595, -0.09474214911460876, 0.053067006170749664, -0.08380461484193802, -0.009564664214849472, -0.1640695482492447, 0.013918053358793259, -0.03671611100435257, 0.09732357412576675, -0.008337993174791336, -0.005582467652857304, 0.035300515592098236, -0.01724814809858799, -0.1183510422706604, -0.017528709024190903, 0.06077078357338905, -0.022408369928598404, -0.1209196001291275, 0.22878782451152802, -0.15548017621040344, 0.18276850879192352, 0.1781492680311203, -0.08699197322130203, 0.0007841655751690269, -0.07041291147470474, -0.04453404247760773, -0.028355194255709648, -0.02484951540827751, -0.020725779235363007, 0.21262159943580627, 0.020799199119210243, 0.16149334609508514, -0.07118167728185654, -0.09581716358661652, 0.0067743705585598946, -0.03355420380830765, -0.0349276103079319, 0.1148703470826149, 0.13563188910484314, -0.1480695605278015, 0.09649838507175446, 0.13694816827774048, -0.02223355881869793, 0.20546066761016846, -0.012604237534105778, -0.10465282201766968, 0.030544409528374672, -0.034288838505744934, -0.04644555598497391, 0.06536617130041122, -0.2036159485578537, -0.04957650229334831, 0.0718756765127182, 0.015663666650652885, 0.11930326372385025, -0.16638870537281036, -0.006676220800727606, 0.001525097293779254, 0.001681964029558003, 0.062142517417669296, 0.11528151482343674, 0.013003956526517868, 0.12133572995662689, -0.03443797677755356, -0.0539020411670208, -0.013307953253388405, 0.009912218898534775, -0.08639662712812424, 0.2238728255033493, -0.02946876361966133, -0.27648377418518066, -0.14987769722938538, -0.020310716703534126, -0.05568944290280342, 0.02760392427444458, 0.030622517690062523, -0.09436463564634323, -0.03789185360074043, -0.021075475960969925, 0.1407509595155716, -0.04429206624627113, 0.05293417349457741, 0.049099598079919815, -0.005358968861401081, -0.06324037909507751, -0.10093974322080612, -0.05527808144688606, -0.08115363121032715, 0.010002178139984608, 0.05914728716015816, -0.06291715800762177, 0.019081449136137962, 0.20412199199199677, 0.0033311969600617886, 0.0554119236767292, -0.006761668715626001, 0.1640099436044693, -0.08864955604076385, 0.033296748995780945, 0.18274877965450287, -0.030267640948295593, 0.02390133962035179, 0.01762339286506176, 0.03278757631778717, -0.08862888067960739, 0.014892547391355038, 0.029995188117027283, -0.11339245736598969, -0.289681613445282, -0.08375092595815659, -0.059449728578329086, 0.03708585724234581, 0.04395202547311783, 0.021688763052225113, 0.0987209901213646, 0.12577234208583832, 0.012119117192924023, 0.03698807209730148, -0.05619025602936745, 0.04368966445326805, 0.09724786877632141, -0.017058921977877617, 0.14598046243190765, -0.051113471388816833, -0.07032549381256104, 0.10300350934267044, 0.01314124371856451, 0.14810822904109955, 0.04016796872019768, 0.01640014909207821, 0.051991380751132965, 0.027465160936117172, 0.07997521758079529, 0.09629271179437637, 0.039239462465047836, -0.0704386755824089, -0.016128400340676308, -0.02210104651749134, -0.006052808836102486, 0.048294708132743835, -0.0023700175806879997, -0.10973481088876724, -0.05502691492438316, 0.017384270206093788, 0.045047271996736526, 0.06268216669559479, 0.13549460470676422, -0.3489506244659424, -0.07352988421916962, 0.052336435765028, -0.03242864832282066, -0.07619840651750565, 0.06986922025680542, -0.011497752740979195, -0.08555059880018234, 0.026570387184619904, -0.1115630567073822, 0.12075411528348923, -0.11784514784812927, 0.03488096594810486, -0.04080212116241455, 0.00015741842798888683, -0.015115424059331417, 0.040514811873435974, -0.21305912733078003, 0.23990382254123688, -0.0063803293742239475, -0.035145360976457596, -0.0808933824300766, 0.010061261244118214, 0.05354377254843712, 0.12696506083011627, 0.10921071469783783, -0.024118416011333466, 0.012280873954296112, 0.036716528236866, -0.07808053493499756, 0.02564050629734993, 0.056990183889865875, -0.08871016651391983, 0.0418991856276989, -0.003985595423728228, 0.027410412207245827, -0.001984293106943369, -0.05347028002142906, -0.015491298399865627, -0.024374129250645638, 0.04645734280347824, 0.03669818118214607, -0.09699360281229019, 0.013771386817097664, -0.11647540330886841, -0.19854719936847687, 0.28432604670524597, -0.018086638301610947, -0.075429767370224, -0.10556349903345108, 0.09776820987462997, 0.14055520296096802, -0.07474096864461899, -0.018789606168866158, -0.04768983647227287, 0.08902277052402496, -0.010831157676875591, -0.1736944317817688, 0.0975857600569725, -0.11972753703594208, -0.05645493417978287, -0.040053367614746094, 0.1432032287120819, -0.019295092672109604, -0.005579779390245676, 0.027777710929512978, -0.024442529305815697, -0.08807820081710815, -0.1146964579820633, 0.027113718912005424, -0.023617390543222427, 0.035260144621133804, -0.029900018125772476, -0.04375217854976654, -0.025222593918442726, -0.011350198648869991, 0.014031311497092247, 0.19251064956188202, 0.1157698705792427, -0.060003507882356644, 0.05297457426786423, 0.1139884889125824, -0.07292436063289642, -0.2682141959667206, 0.03944597393274307, -0.030327988788485527, 0.05729223042726517, 0.015722136944532394, -0.11296986043453217, 0.14410188794136047, 0.05509579926729202, -0.024778319522738457, 0.1260213851928711, -0.2914702594280243, -0.1293436884880066, 0.21833842992782593, 0.062312182039022446, 0.2836434245109558, -0.12400611490011215, -0.07586991786956787, 0.015847207978367805, -0.10183092206716537, 0.10871751606464386, -0.06311415880918503, 0.06262818723917007, -0.0356149785220623, 0.13329899311065674, 0.05534530431032181, -0.09046284109354019, 0.10944715142250061, 0.00044721236918121576, 0.016697004437446594, -0.05938730016350746, -0.01562361791729927, 0.05971069261431694, -0.04378388449549675, 0.09913011640310287, -0.01592315174639225, 0.0439961738884449, -0.1315499246120453, -0.04797155782580376, -0.04664883390069008, 0.07085953652858734, 0.005759327672421932, -0.08992526680231094, -0.030085889622569084, 0.006370361894369125, -0.0547250360250473, -0.018014881759881973, 0.19290673732757568, -0.016040777787566185, 0.07233378291130066, 0.03246074542403221, 0.12086059898138046, -0.024785013869404793, -0.027620069682598114, -0.023836730048060417, 0.009618186391890049, 0.06771928071975708, -0.21087254583835602, 0.012383745983242989, 0.14499898254871368, -0.0037646384444087744, 0.08440294861793518, 0.06883169710636139, -0.045296937227249146, 0.013128349557518959, 0.10451635718345642, -0.2084948569536209, -0.031701453030109406, -0.06710511445999146, -0.18889158964157104, 0.033620353788137436, 0.17069673538208008, 0.18047894537448883, -0.04602695256471634, 0.006856892723590136, 0.029704146087169647, 0.003980305511504412, 0.0011914814822375774, 0.16324114799499512, 0.0030774702318012714, 0.040300264954566956, -0.1153789833188057, 0.09801189601421356, -0.020616352558135986, -0.06445114314556122, 0.0010478749172762036, 0.07205764949321747, -0.1083408072590828, -0.09930532425642014, -0.0777624174952507, 0.10235802084207535, -0.15845152735710144, -0.05178725719451904, -0.11533257365226746, -0.040701303631067276, 0.05644351243972778, 0.09100161492824554, 0.09572999179363251, 0.0611250214278698, -0.09286805987358093, -0.07106713950634003, -0.08610314130783081, 0.04823409020900726, 0.06646951287984848, 0.029027746990323067, -0.09947752952575684, 0.06581910699605942, -0.013331321999430656, 0.09008307009935379, -0.09755148738622665, -0.02139393799006939, -0.10788876563310623, 0.003297733375802636, -0.17944325506687164, -0.07760921120643616, -0.10995063930749893, -0.05862390249967575, -0.023645510897040367, -0.040373288094997406, -0.10839162766933441, 0.0018815773073583841, -0.09987448900938034, 0.02748218923807144, -0.019283313304185867, 0.03745348006486893, -0.05924588441848755, 0.016257798299193382, 0.005621479358524084, -0.0017023792024701834, 0.13539999723434448, 0.0820019394159317, -0.027124589309096336, 0.10980897396802902, -0.12354373186826706, 0.09036692976951599, 0.11531993001699448, 0.00802763644605875, 0.06770531088113785, 0.004833292681723833, 0.03658154606819153, 0.03247476741671562, 0.05141814053058624, 0.0671190619468689, -0.04938667640089989, -0.08578832447528839, -0.045821599662303925, -0.06886725127696991, -0.015515036880970001, -0.0549684539437294, 0.019696494564414024, 0.02071366086602211, 0.09770134836435318, 0.08668169379234314, -0.011563029140233994, 0.052236370742321014, -0.12837152183055878, 0.029008008539676666, 0.009450329467654228, -0.13405567407608032, 0.0301483403891325, -0.04202243313193321, 0.04968748241662979, -0.03859352320432663, 0.27751579880714417, 0.05679735541343689, 0.08986175805330276, 0.0276227667927742, 0.12588463723659515, -0.004077521618455648, 0.005195667035877705, 0.16167974472045898, 0.06395865231752396, -0.021548807621002197, -0.10551615804433823, 0.13322190940380096, 0.054251328110694885, -0.027897262945771217, 0.1860329806804657, -0.004177907481789589, -0.1434275060892105, 0.12953388690948486, -0.010091927833855152, -0.02062745951116085, -0.1370360106229782, -0.029092445969581604, -0.10173632204532623, 0.11233825981616974, 0.02800975739955902, 0.09341743588447571, 0.17391391098499298, -0.04847736284136772, 0.020987998694181442, 0.014212582260370255, -0.05372250825166702, -0.15992577373981476, -0.17685267329216003, -0.07958993315696716, -0.07019839435815811, 0.028971567749977112, -0.12403275072574615, -0.014644021168351173, 0.004975670017302036, 0.11394805461168289, -0.12172193080186844, 0.1679459661245346, 0.06787151843309402, -0.10319259762763977, 0.11319077759981155, 0.031167276203632355, 0.022200440987944603, -0.0698479413986206, 0.020722463726997375, -0.07161379605531693, 0.05142522230744362, -0.005876749288290739, -0.05777589604258537, -0.04335243999958038, 0.028734533116221428, -0.07866581529378891, -0.07663046568632126, -0.06475424766540527, 0.027262045070528984, 0.015580869279801846, 0.021426355466246605, 0.027208544313907623, -0.02264404296875, -0.02731146104633808, 0.22099676728248596, -0.06467397511005402, 0.0488581620156765, -0.10712907463312149, 0.14227290451526642, 0.04022226855158806, 0.03282289579510689, -0.00125005969312042, -0.01564532145857811, -0.03733254596590996, 0.3163157105445862, 0.28317028284072876, -0.02803904004395008, -0.0071405754424631596, 0.029801322147250175, 0.011906430125236511, 0.014932806603610516, 0.08594637364149094, 0.03824135661125183, 0.2132243663072586, -0.03659302368760109, -0.06992164999246597, -0.03658803552389145, -0.04868223890662193, 0.03563283383846283, 0.10042934119701385, 0.08599155396223068, -0.04871150106191635, -0.04558972269296646, 0.12336263060569763, -0.16286759078502655, -0.0687471404671669, -0.09763829410076141, -0.13022026419639587, -0.1380651891231537, -0.003191143274307251, 0.031540870666503906, -0.0401642769575119, 0.0702163353562355, -0.03389789164066315, -0.05650300532579422, 0.012690185569226742, -0.016998078674077988, -0.16411042213439941, 0.01389013696461916, 0.009984356351196766, 0.006780180148780346, 0.08372968435287476, -0.01350899413228035, -0.0005033733905293047, 0.07922549545764923, 0.02800803631544113, -0.04629982262849808, 0.030209949240088463, 0.019835732877254486, -0.04719768092036247, 0.0747232437133789, 0.0696539506316185, 0.03389411419630051, -0.11214417964220047, 0.08118103444576263, -0.10398104041814804, 0.01009329129010439, -0.14444495737552643, -0.045543454587459564, -0.06741238385438919, 0.06053530424833298, -0.05359449237585068, 0.1104620024561882, 0.14376576244831085, -0.004573410376906395, -0.01786770671606064, -0.06614304333925247, 0.0628657266497612, -0.016188334673643112, 0.028553295880556107, -0.02301710657775402, -0.14730381965637207, -0.02161852829158306, 0.08017408847808838, -0.044095661491155624, -0.29006582498550415, 0.005712594371289015, -0.043838899582624435, -0.034592367708683014, -0.0806223452091217, 0.0650295615196228, 0.11501219868659973, 0.03777404874563217, -0.0120090888813138, -0.019361818209290504, 0.010873000137507915, 0.04637501761317253, -0.17532676458358765, -0.1083746999502182 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # indobert-base-uncased-finetuned-indonlu-smsa This model is a fine-tuned version of [indolem/indobert-base-uncased](https://huggingface.co/indolem/indobert-base-uncased) on the indonlu dataset. It achieves the following results on the evaluation set: - Loss: 0.2277 - Accuracy: 0.9302 - F1: 0.9066 - Precision: 0.8992 - Recall: 0.9147 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1500 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:| | No log | 1.0 | 344 | 0.3831 | 0.8476 | 0.7715 | 0.7817 | 0.7627 | | 0.4167 | 2.0 | 688 | 0.2809 | 0.8905 | 0.8406 | 0.8699 | 0.8185 | | 0.2624 | 3.0 | 1032 | 0.2254 | 0.9230 | 0.8842 | 0.9004 | 0.8714 | | 0.2624 | 4.0 | 1376 | 0.2378 | 0.9238 | 0.8797 | 0.9180 | 0.8594 | | 0.1865 | 5.0 | 1720 | 0.2277 | 0.9302 | 0.9066 | 0.8992 | 0.9147 | | 0.1217 | 6.0 | 2064 | 0.2444 | 0.9262 | 0.8981 | 0.9013 | 0.8957 | | 0.1217 | 7.0 | 2408 | 0.2985 | 0.9286 | 0.8999 | 0.9035 | 0.8971 | | 0.0847 | 8.0 | 2752 | 0.3397 | 0.9278 | 0.8969 | 0.9090 | 0.8871 | | 0.0551 | 9.0 | 3096 | 0.3542 | 0.9270 | 0.8961 | 0.9010 | 0.8924 | | 0.0551 | 10.0 | 3440 | 0.3862 | 0.9222 | 0.8895 | 0.8970 | 0.8846 | ### Framework versions - Transformers 4.14.1 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"language": "id", "license": "mit", "tags": ["generated_from_trainer"], "datasets": ["indonlu"], "metrics": ["accuracy", "f1", "precision", "recall"], "widget": [{"text": "Entah mengapa saya merasakan ada sesuatu yang janggal di produk ini"}], "model-index": [{"name": "indobert-base-uncased-finetuned-indonlu-smsa", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "indonlu", "type": "indonlu", "args": "smsa"}, "metrics": [{"type": "accuracy", "value": 0.9301587301587302, "name": "Accuracy"}, {"type": "f1", "value": 0.9066105299178986, "name": "F1"}, {"type": "precision", "value": 0.8992078788375845, "name": "Precision"}, {"type": "recall", "value": 0.9147307323234121, "name": "Recall"}]}]}]}
text-classification
ayameRushia/indobert-base-uncased-finetuned-indonlu-smsa
[ "transformers", "pytorch", "safetensors", "bert", "text-classification", "generated_from_trainer", "id", "dataset:indonlu", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #safetensors #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
indobert-base-uncased-finetuned-indonlu-smsa ============================================ This model is a fine-tuned version of indolem/indobert-base-uncased on the indonlu dataset. It achieves the following results on the evaluation set: * Loss: 0.2277 * Accuracy: 0.9302 * F1: 0.9066 * Precision: 0.8992 * Recall: 0.9147 ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1500 * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.14.1 * Pytorch 1.10.0+cu111 * Datasets 1.17.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #safetensors #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ 66, 116, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #bert #text-classification #generated_from_trainer #id #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ -0.1200881227850914, 0.0961245745420456, -0.0029513067565858364, 0.1186232790350914, 0.14128491282463074, 0.018396200612187386, 0.13863709568977356, 0.13996286690235138, -0.06858107447624207, 0.04150970280170441, 0.12430985271930695, 0.14602310955524445, 0.018594758585095406, 0.14098167419433594, -0.06697011739015579, -0.2752443552017212, 0.006323852110654116, 0.02066919021308422, -0.06385838240385056, 0.13124071061611176, 0.07773225009441376, -0.13538268208503723, 0.10055918246507645, -0.001933935796841979, -0.11049588024616241, 0.0028990223072469234, 0.017445705831050873, -0.07084286212921143, 0.14204426109790802, 0.00857257004827261, 0.11167526245117188, 0.03239155188202858, 0.10772334039211273, -0.1920042634010315, 0.009063380770385265, 0.055428341031074524, 0.008200155571103096, 0.08905636519193649, 0.05817205458879471, -0.01705443114042282, 0.13852663338184357, -0.1059165969491005, 0.07219638675451279, 0.019532514736056328, -0.1269238442182541, -0.22071562707424164, -0.07520505040884018, 0.058496393263339996, 0.07952356338500977, 0.09370304644107819, -0.006545391399413347, 0.128098264336586, -0.07972515374422073, 0.11447499692440033, 0.25032636523246765, -0.2836129367351532, -0.05970053747296333, 0.011531062424182892, 0.0208351518958807, 0.053226716816425323, -0.12306482344865799, -0.01641625352203846, 0.03732027858495712, 0.02877774089574814, 0.13713417947292328, -0.025088323280215263, -0.060693468898534775, 0.002694078953936696, -0.1370788812637329, -0.043468184769153595, 0.1207619234919548, 0.02375609055161476, -0.048030219972133636, -0.07423889636993408, -0.0672798901796341, -0.16656889021396637, -0.04655975103378296, -0.015309009701013565, 0.036474887281656265, -0.04674401134252548, -0.0646720826625824, 0.0016999252839013934, -0.09463228285312653, -0.08545534312725067, -0.026163650676608086, 0.17833992838859558, 0.04976019635796547, 0.004383373539894819, -0.02382836677134037, 0.1164378896355629, 0.017824415117502213, -0.14960260689258575, 0.0017984467558562756, 0.016364291310310364, 0.006197217386215925, -0.03953023999929428, -0.022244738414883614, -0.011634492315351963, 0.011709232814610004, 0.1470222771167755, -0.11606443673372269, 0.041916754096746445, 0.037890493869781494, 0.010533319786190987, -0.09934496134519577, 0.16479362547397614, -0.032841965556144714, -0.0216533150523901, 0.006291295401751995, 0.09212053567171097, 0.03220900148153305, -0.015298030339181423, -0.084809809923172, -0.0025803581811487675, 0.10957092046737671, 0.043790608644485474, -0.06538967043161392, 0.07955904304981232, -0.039777275174856186, -0.008205235004425049, 0.04984552040696144, -0.10419539362192154, 0.03753817453980446, 0.022886687889695168, -0.07250271737575531, -0.05535409599542618, 0.030862491577863693, -0.004383249208331108, -0.010248666629195213, 0.1465972512960434, -0.08841101825237274, 0.01391763798892498, -0.07956268638372421, -0.12355997413396835, 0.009550190530717373, -0.0967545211315155, 0.010768165811896324, -0.08683021366596222, -0.1983434408903122, -0.010487991385161877, 0.03955114260315895, -0.026737777516245842, -0.019471121951937675, -0.05547337606549263, -0.09163598716259003, 0.02638455480337143, -0.012896905653178692, 0.07954917103052139, -0.07094326615333557, 0.11666154116392136, 0.031847380101680756, 0.06628525257110596, 0.00007868393004173413, 0.045486800372600555, -0.1169268935918808, 0.02054859884083271, -0.16857865452766418, 0.026623209938406944, -0.05925735458731651, 0.04264357313513756, -0.07287126779556274, -0.1147908866405487, 0.006962451618164778, 0.010530517436563969, 0.07200144231319427, 0.14734232425689697, -0.1671770215034485, -0.07955610752105713, 0.17709122598171234, -0.11615408211946487, -0.11840969324111938, 0.11505693197250366, -0.05308663100004196, 0.025951962918043137, 0.06580637395381927, 0.16399578750133514, 0.09290028363466263, -0.11619704961776733, -0.018380330875515938, -0.006875927094370127, 0.042312026023864746, -0.02315244823694229, 0.07052159309387207, 0.005981933791190386, 0.016468821093440056, 0.013531502336263657, -0.03163626044988632, 0.027422554790973663, -0.09855130314826965, -0.09208282828330994, -0.03600995987653732, -0.07802075147628784, 0.06751394271850586, 0.07315078377723694, 0.04542132094502449, -0.1168108806014061, -0.09904536604881287, 0.03275800868868828, 0.09402062743902206, -0.06551365554332733, 0.0188148096203804, -0.06809218972921371, 0.06836343556642532, -0.013750748708844185, -0.024304667487740517, -0.16905014216899872, -0.049321696162223816, 0.01539409626275301, 0.0044977739453315735, -0.000505139643792063, -0.005517864599823952, 0.08705934882164001, 0.09483884274959564, -0.06308559328317642, -0.05662232264876366, -0.02854427509009838, 0.0016251524211838841, -0.11449018865823746, -0.21247777342796326, -0.029992379248142242, -0.031879790127277374, 0.10010161995887756, -0.21794594824314117, 0.04705381020903587, 0.004064752720296383, 0.0818091407418251, 0.039894949644804, -0.015423405915498734, -0.020473770797252655, 0.06826514005661011, -0.03788373991847038, -0.05682367831468582, 0.06558150053024292, -0.01181809138506651, -0.07524451613426208, -0.03756420686841011, -0.1287449151277542, 0.18194147944450378, 0.11856047809123993, -0.046160563826560974, -0.09344407916069031, -0.011804845184087753, -0.060612574219703674, -0.026049496605992317, -0.06126713752746582, 0.022467536851763725, 0.15534807741641998, 0.008717970922589302, 0.14230525493621826, -0.07865000516176224, -0.037761084735393524, 0.030442211776971817, -0.03222817927598953, 0.012427562847733498, 0.13757136464118958, 0.0903281718492508, -0.09111737459897995, 0.14437952637672424, 0.14055941998958588, -0.05314887315034866, 0.15850867331027985, -0.04197967052459717, -0.06444136053323746, -0.018045490607619286, -0.0051957606337964535, 0.0052610901184380054, 0.11718910187482834, -0.11104705929756165, 0.0006341564003378153, 0.014743175357580185, 0.01598455384373665, -0.005074511282145977, -0.21733669936656952, -0.035966623574495316, 0.034334905445575714, -0.06446702778339386, -0.039890218526124954, 0.008885501883924007, 0.01410557422786951, 0.11668138951063156, -0.011254185810685158, -0.08805849403142929, 0.018479526042938232, -0.0009308147127740085, -0.07539539784193039, 0.2152889519929886, -0.07575613260269165, -0.16220153868198395, -0.1211482509970665, -0.034215833991765976, -0.05207909271121025, 0.014348424971103668, 0.06745597720146179, -0.0909380167722702, -0.03300023451447487, -0.08991675823926926, 0.010547864250838757, 0.02109517715871334, 0.03499699383974075, -0.020366055890917778, 0.002029211027547717, 0.05787516012787819, -0.1153695285320282, -0.026652641594409943, -0.05594852194190025, -0.06200363114476204, 0.050756365060806274, 0.04994131624698639, 0.10595361143350601, 0.11742593348026276, -0.027416884899139404, 0.015055556781589985, -0.03984468802809715, 0.21444670855998993, -0.06885252892971039, -0.015199046581983566, 0.15292732417583466, -0.007463345304131508, 0.05574847385287285, 0.14031197130680084, 0.06493271887302399, -0.08630766719579697, 0.007082928437739611, 0.039590153843164444, -0.023846952244639397, -0.20857203006744385, -0.042833708226680756, -0.030157556757330894, 0.030621159821748734, 0.10347861796617508, 0.031402941793203354, 0.01956048235297203, 0.05799220874905586, 0.0025953517761081457, 0.020281637087464333, -0.011937253177165985, 0.07171925902366638, 0.10066910833120346, 0.03765708953142166, 0.12785474956035614, -0.03775175288319588, -0.07145386934280396, 0.03703702986240387, -0.04479493573307991, 0.21335601806640625, -0.013332019560039043, 0.10357813537120819, 0.045976027846336365, 0.1565771996974945, 0.008311372250318527, 0.10143143683671951, -0.0026508131995797157, -0.03768140822649002, -0.003682119771838188, -0.05624157562851906, -0.04736613854765892, 0.02537432499229908, -0.07906652987003326, 0.07170869410037994, -0.15356792509555817, 0.03837377950549126, 0.051336128264665604, 0.24614626169204712, 0.04556114599108696, -0.3208085596561432, -0.10258807241916656, 0.0039887926541268826, -0.035010650753974915, -0.03074098378419876, 0.03232240304350853, 0.12151600420475006, -0.08744892477989197, 0.04647669196128845, -0.08001425862312317, 0.07317093014717102, -0.05574490502476692, 0.04587585851550102, 0.06201624125242233, 0.08682915568351746, -0.025137588381767273, 0.07089339196681976, -0.26704874634742737, 0.2995362877845764, 0.012643743306398392, 0.08045829832553864, -0.06937799602746964, 0.00530997384339571, 0.04892921447753906, 0.0870932936668396, 0.10836411267518997, -0.02283044345676899, -0.06399737298488617, -0.2170969545841217, -0.0841924324631691, 0.020658602938055992, 0.1044309213757515, -0.04392249882221222, 0.11387371271848679, -0.041326798498630524, -0.01247213315218687, 0.06293756514787674, -0.04631645604968071, -0.07641035318374634, -0.07274132966995239, -0.006865328643471003, 0.02655133418738842, -0.018400438129901886, -0.07900426536798477, -0.12308050692081451, -0.08912345767021179, 0.12225552648305893, -0.036653000861406326, -0.05112098529934883, -0.12208062410354614, 0.052842188626527786, 0.09029079228639603, -0.10055923461914062, 0.03451203927397728, 0.008280985057353973, 0.07605776190757751, 0.022520247846841812, -0.05263330787420273, 0.10470898449420929, -0.08522259443998337, -0.21399472653865814, -0.05856506898999214, 0.13482604920864105, 0.027440842241048813, 0.05428382754325867, -0.011222300119698048, 0.028194108977913857, -0.01746186800301075, -0.0876983031630516, 0.024339359253644943, -0.0008011105819605291, 0.0759827271103859, 0.026335583999753, -0.07615433633327484, -0.004821607377380133, -0.059083156287670135, -0.01690012216567993, 0.15412917733192444, 0.26887694001197815, -0.09340042620897293, 0.046759240329265594, 0.05218026041984558, -0.056058987975120544, -0.21305707097053528, 0.011464506387710571, 0.03725268691778183, 0.0028966821264475584, 0.04899333044886589, -0.16769875586032867, 0.09089161455631256, 0.08139396458864212, -0.018853818997740746, 0.09909618645906448, -0.2877327799797058, -0.1298275738954544, 0.12566499412059784, 0.15162527561187744, 0.09805101156234741, -0.14678694307804108, -0.02252078428864479, 0.005594693589955568, -0.07417160272598267, 0.10221262276172638, -0.07124689221382141, 0.11611516773700714, -0.04244314879179001, 0.06553992629051208, 0.025401882827281952, -0.058069635182619095, 0.10079325735569, -0.00845467858016491, 0.1124451756477356, -0.05825298652052879, -0.040803708136081696, 0.020568203181028366, -0.0567370280623436, 0.05267975106835365, -0.10751798748970032, 0.0415772870182991, -0.07932768762111664, -0.023942451924085617, -0.07954303175210953, 0.049707554280757904, -0.04967070370912552, -0.0697934627532959, -0.025955989956855774, 0.04193241149187088, 0.04417995363473892, -0.023494156077504158, 0.15410611033439636, 0.004089062567800283, 0.15646174550056458, 0.09795050323009491, 0.10243185609579086, -0.0491362065076828, -0.04362724348902702, -0.0030335101764649153, -0.015128644183278084, 0.05468929558992386, -0.1277037262916565, 0.03578527271747589, 0.13450945913791656, 0.033567164093256, 0.13749302923679352, 0.07495694607496262, -0.017536969855427742, -0.00892203114926815, 0.07261145114898682, -0.17963357269763947, -0.07929029315710068, -0.019007401540875435, -0.06642203032970428, -0.12037066370248795, 0.07045061141252518, 0.10980647802352905, -0.07899391651153564, 0.0012517647119238973, -0.018679510802030563, 0.026150062680244446, -0.012272379361093044, 0.19290579855442047, 0.05558000132441521, 0.06026032567024231, -0.09796542674303055, 0.07222386449575424, 0.029686080291867256, -0.07653376460075378, 0.01641254872083664, 0.09128844738006592, -0.0903913825750351, -0.046477675437927246, 0.048238810151815414, 0.18792887032032013, -0.031503453850746155, -0.04757164046168327, -0.15275374054908752, -0.11486358195543289, 0.07098131626844406, 0.1834222972393036, 0.0870242714881897, 0.015074255876243114, -0.050032079219818115, 0.018350550904870033, -0.12767788767814636, 0.1318395733833313, 0.04848412424325943, 0.060904234647750854, -0.15814939141273499, 0.16018393635749817, 0.0012843506410717964, 0.02059748023748398, -0.018422868102788925, 0.03202700987458229, -0.11426392942667007, -0.0029819216579198837, -0.08708888292312622, -0.022647887468338013, -0.0500791072845459, 0.009042767807841301, -0.0077608744613826275, -0.05770231783390045, -0.049111541360616684, 0.01007028203457594, -0.09734369814395905, -0.014166836626827717, 0.02460932359099388, 0.06466986984014511, -0.1435670703649521, -0.04230911657214165, 0.019666308537125587, -0.07535195350646973, 0.07695148140192032, 0.042878903448581696, 0.015065068379044533, 0.05306475609540939, -0.13364779949188232, 0.026388827711343765, 0.05475074052810669, 0.0018089080695062876, 0.04920100420713425, -0.10966121405363083, 0.0023710494861006737, -0.018692204728722572, 0.026707496494054794, 0.02725105732679367, 0.06842666864395142, -0.12913115322589874, 0.03801447153091431, 0.0021512326784431934, -0.07010301202535629, -0.05892421677708626, 0.0388953797519207, 0.08790192008018494, -0.013405369594693184, 0.2059643566608429, -0.10036278516054153, 0.052316851913928986, -0.2194909006357193, 0.0011289168614894152, -0.0007618482341058552, -0.11897806078195572, -0.12982727587223053, -0.0612717866897583, 0.0761658102273941, -0.05434783548116684, 0.10836558789014816, 0.04329945519566536, 0.05399556830525398, 0.028398685157299042, -0.05553566291928291, 0.023229287937283516, 0.010138123296201229, 0.18554894626140594, 0.02721591480076313, -0.05135411024093628, 0.05877181515097618, 0.028557036072015762, 0.09312237799167633, 0.0790623351931572, 0.21928216516971588, 0.1426263451576233, -0.012092525139451027, 0.08288563787937164, 0.042799413204193115, -0.07657133042812347, -0.1667606681585312, -0.008571865037083626, -0.034956272691488266, 0.08771083503961563, -0.023334354162216187, 0.20442421734333038, 0.08881674706935883, -0.16594769060611725, 0.04511153697967529, -0.03734530881047249, -0.08031097054481506, -0.1390218734741211, -0.049340128898620605, -0.08600816875696182, -0.12742888927459717, 0.011468162760138512, -0.12501710653305054, 0.02865270897746086, 0.06515266746282578, 0.01173529401421547, -0.012059880420565605, 0.18067270517349243, 0.005755183286964893, 0.04052109643816948, 0.058838773518800735, 0.026415502652525902, -0.017315341159701347, -0.075166255235672, -0.07410430163145065, -0.013295508921146393, -0.04248835891485214, 0.015493319369852543, -0.05693773925304413, -0.06175679340958595, 0.04105548560619354, -0.005309928674250841, -0.10453496128320694, 0.013153641484677792, 0.01387002132833004, 0.07227213680744171, 0.07777121663093567, 0.02177216112613678, 0.010760411620140076, -0.009243305772542953, 0.2278234213590622, -0.08202682435512543, -0.06931336969137192, -0.08731075376272202, 0.2932443618774414, 0.04833854362368584, -0.0077840485610067844, 0.03702220693230629, -0.07271157950162888, 0.014246795326471329, 0.22072577476501465, 0.2014196366071701, -0.05393184348940849, 0.018761852756142616, -0.01826751045882702, -0.0085088349878788, -0.01873193122446537, 0.09162301570177078, 0.11093513667583466, 0.07671697437763214, -0.0726763904094696, -0.028502559289336205, -0.04349912703037262, -0.01560370996594429, -0.047641322016716, 0.07369273900985718, 0.03442271798849106, 0.00677453400567174, -0.04128335788846016, 0.0594770610332489, -0.04057290405035019, -0.12418173998594284, 0.0651422068476677, -0.2195121943950653, -0.14211907982826233, -0.017124993726611137, 0.09560637921094894, 0.020351344719529152, 0.061655137687921524, -0.012803071178495884, -0.01242618728429079, 0.06382567435503006, -0.00766892870888114, -0.07959261536598206, -0.08224068582057953, 0.08016173541545868, -0.09336980432271957, 0.234733447432518, -0.03532702103257179, 0.042003899812698364, 0.11691572517156601, 0.03563959151506424, -0.09144603461027145, 0.08190661668777466, 0.05459587275981903, -0.09969356656074524, 0.019023306667804718, 0.106458880007267, -0.037833768874406815, 0.08978643268346786, 0.04000606760382652, -0.14624133706092834, 0.018583860248327255, -0.04096028581261635, -0.09036216884851456, -0.04441047087311745, -0.011032969690859318, -0.03423737734556198, 0.13957619667053223, 0.21850477159023285, -0.026464316993951797, 0.007393653970211744, -0.06270609050989151, 0.02860952354967594, 0.06108503043651581, 0.011244667693972588, -0.033634886145591736, -0.23560871183872223, 0.024254385381937027, 0.0765516459941864, -0.005745375528931618, -0.26323604583740234, -0.09338458627462387, 0.014014307409524918, -0.05407965928316116, -0.11600174009799957, 0.0680532231926918, 0.07238972932100296, 0.05228922888636589, -0.03604632243514061, -0.0465664267539978, -0.05718184635043144, 0.17008110880851746, -0.15726648271083832, -0.07943059504032135 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-indonesian-1.5G-sentiment-analysis-smsa This model is a fine-tuned version of [cahya/roberta-base-indonesian-1.5G](https://huggingface.co/cahya/roberta-base-indonesian-1.5G) on the indonlu dataset. It achieves the following results on the evaluation set: - Loss: 0.4294 - Accuracy: 0.9262 ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1500 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.6461 | 1.0 | 688 | 0.2620 | 0.9087 | | 0.2627 | 2.0 | 1376 | 0.2291 | 0.9151 | | 0.1784 | 3.0 | 2064 | 0.2891 | 0.9167 | | 0.1099 | 4.0 | 2752 | 0.3317 | 0.9230 | | 0.0857 | 5.0 | 3440 | 0.4294 | 0.9262 | | 0.0346 | 6.0 | 4128 | 0.4759 | 0.9246 | | 0.0221 | 7.0 | 4816 | 0.4946 | 0.9206 | | 0.006 | 8.0 | 5504 | 0.5823 | 0.9175 | | 0.0047 | 9.0 | 6192 | 0.5777 | 0.9159 | | 0.004 | 10.0 | 6880 | 0.5800 | 0.9175 | ### How to use this model in Transformers Library ```python from transformers import pipeline pipe = pipeline( "text-classification", model="ayameRushia/roberta-base-indonesian-1.5G-sentiment-analysis-smsa" ) pipe("Terima kasih atas bantuannya ya!") ``` ### Framework versions - Transformers 4.14.1 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"language": ["id"], "tags": ["generated_from_trainer"], "datasets": ["indonlp/indonlu"], "metrics": ["accuracy"], "widget": [{"text": "Entah mengapa saya merasakan ada sesuatu yang janggal di produk ini"}], "model-index": [{"name": "roberta-base-indonesian-1.5G-sentiment-analysis-smsa", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "indonlu", "type": "indonlu", "args": "smsa"}, "metrics": [{"type": "accuracy", "value": 0.9261904761904762, "name": "Accuracy"}]}]}]}
text-classification
ayameRushia/roberta-base-indonesian-1.5G-sentiment-analysis-smsa
[ "transformers", "pytorch", "roberta", "text-classification", "generated_from_trainer", "id", "dataset:indonlp/indonlu", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #roberta #text-classification #generated_from_trainer #id #dataset-indonlp/indonlu #model-index #autotrain_compatible #endpoints_compatible #region-us
roberta-base-indonesian-1.5G-sentiment-analysis-smsa ==================================================== This model is a fine-tuned version of cahya/roberta-base-indonesian-1.5G on the indonlu dataset. It achieves the following results on the evaluation set: * Loss: 0.4294 * Accuracy: 0.9262 ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1500 * num\_epochs: 10 ### Training results ### How to use this model in Transformers Library ### Framework versions * Transformers 4.14.1 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10", "### Training results", "### How to use this model in Transformers Library", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #roberta #text-classification #generated_from_trainer #id #dataset-indonlp/indonlu #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10", "### Training results", "### How to use this model in Transformers Library", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 61, 116, 4, 12, 33 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #text-classification #generated_from_trainer #id #dataset-indonlp/indonlu #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1500\n* num\\_epochs: 10### Training results### How to use this model in Transformers Library### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.09014276415109634, 0.1129119023680687, -0.003086252138018608, 0.10822094976902008, 0.16209177672863007, 0.01805664412677288, 0.13475678861141205, 0.14819595217704773, -0.12067179381847382, 0.04961404576897621, 0.13248412311077118, 0.17654989659786224, 0.04191618412733078, 0.14752599596977234, -0.06949098408222198, -0.2752723693847656, 0.016668250784277916, 0.03131556510925293, -0.07950295507907867, 0.12688642740249634, 0.1009032130241394, -0.12762503325939178, 0.096707284450531, 0.007545106112957001, -0.18354403972625732, -0.007718630135059357, -0.018848849460482597, -0.07760955393314362, 0.11421026289463043, 0.026348674669861794, 0.09562154114246368, 0.023676911368966103, 0.09537379443645477, -0.15670526027679443, 0.006406845059245825, 0.042718976736068726, 0.01693219691514969, 0.11075097322463989, 0.05265270173549652, -0.03715869411826134, 0.1231749951839447, -0.0951254591345787, 0.0848311260342598, 0.03315814957022667, -0.1331394612789154, -0.22595633566379547, -0.0959603562951088, 0.0816572830080986, 0.0841585099697113, 0.08209117501974106, -0.0012371958000585437, 0.15430179238319397, -0.0510193333029747, 0.1119193583726883, 0.25812289118766785, -0.2596179246902466, -0.06632572412490845, 0.003926608245819807, 0.03519602119922638, 0.03968561068177223, -0.09086711704730988, -0.006237422116100788, 0.03841567412018776, 0.046091727912425995, 0.15744447708129883, -0.03102007880806923, -0.044753964990377426, 0.009479127824306488, -0.14435824751853943, -0.07243749499320984, 0.13759440183639526, 0.018303576856851578, -0.04861476272344589, -0.0669541135430336, -0.06896255910396576, -0.1464768499135971, -0.053021032363176346, -0.013055777177214622, 0.03786958381533623, -0.06709933280944824, -0.0895358994603157, -0.022509342059493065, -0.0551735982298851, -0.05265841633081436, -0.020304497331380844, 0.18239663541316986, 0.04606279358267784, 0.010588330216705799, -0.01022854819893837, 0.09348952770233154, -0.008119090460240841, -0.136814683675766, -0.0012598395114764571, 0.01623702235519886, -0.011518371291458607, -0.06030689924955368, -0.01700735278427601, -0.042429517954587936, -0.003412423189729452, 0.2073732316493988, -0.09856581687927246, 0.05218154191970825, 0.03851514682173729, 0.009127774275839329, -0.043384358286857605, 0.19302234053611755, -0.06865648180246353, -0.002349741756916046, 0.01842014119029045, 0.0693787932395935, 0.007082989439368248, -0.03620874509215355, -0.10890115052461624, 0.014461525715887547, 0.1076490506529808, 0.04505181312561035, -0.025000669062137604, 0.058473024517297745, -0.02121501788496971, -0.038492828607559204, 0.06512629240751266, -0.11248185485601425, 0.06004014611244202, 0.002532336162403226, -0.08288513869047165, -0.016494272276759148, 0.03814665228128433, -0.01316861528903246, -0.04438692703843117, 0.11105061322450638, -0.09186485409736633, 0.012766050174832344, -0.0957016795873642, -0.15214960277080536, 0.014188996516168118, -0.0921543538570404, 0.015460210852324963, -0.10082896053791046, -0.16262322664260864, -0.019993918016552925, 0.055409494787454605, -0.06073320657014847, -0.06292682141065598, -0.050630200654268265, -0.0973060205578804, 0.04322071373462677, 0.0044422391802072525, 0.0911039337515831, -0.04755324125289917, 0.08489871025085449, 0.05731850117444992, 0.07536690682172775, 0.015733860433101654, 0.05369489639997482, -0.07886708527803421, 0.04333892837166786, -0.21887221932411194, 0.0607004389166832, -0.060665883123874664, 0.07544422894716263, -0.11149214208126068, -0.11497991532087326, 0.07848593592643738, -0.010926139540970325, 0.09138309955596924, 0.13715820014476776, -0.1213102862238884, -0.07429618388414383, 0.14852005243301392, -0.06816674768924713, -0.12417857348918915, 0.1215372383594513, -0.05741540342569351, 0.013109938241541386, 0.038453951478004456, 0.14812757074832916, 0.08153549581766129, -0.09010058641433716, -0.031470105051994324, -0.02170695550739765, 0.030901938676834106, -0.022285213693976402, 0.06795628368854523, 0.026307526975870132, 0.024519551545381546, -0.00584345543757081, -0.007388891186565161, 0.026825299486517906, -0.10744310915470123, -0.0807289332151413, -0.041885409504175186, -0.08710181713104248, 0.054655808955430984, 0.050974249839782715, 0.0504942461848259, -0.09733836352825165, -0.10088072717189789, 0.09232766926288605, 0.09178733825683594, -0.06307824701070786, 0.042206473648548126, -0.0928073301911354, 0.09339595586061478, -0.05170705169439316, 0.004011832643300295, -0.19251026213169098, -0.014477994292974472, 0.013770993798971176, 0.026746775954961777, 0.028639495372772217, -0.07634339481592178, 0.08263411372900009, 0.07518381625413895, -0.04290584847331047, -0.044976718723773956, -0.029011467471718788, 0.007605137303471565, -0.11603861302137375, -0.21114876866340637, -0.07240182906389236, -0.03408534452319145, 0.1091349869966507, -0.18141324818134308, 0.0493292510509491, 0.0067582763731479645, 0.10519587993621826, 0.008991710841655731, -0.014918110333383083, -0.00860273465514183, 0.06443100422620773, -0.04963694140315056, -0.07890544086694717, 0.043005168437957764, 0.007438113912940025, -0.0701521635055542, -0.026568230241537094, -0.14194972813129425, 0.12898077070713043, 0.12656645476818085, -0.03636165335774422, -0.1065722107887268, 0.0027682699728757143, -0.052169445902109146, -0.026658926159143448, -0.03012918308377266, 0.018847640603780746, 0.15119139850139618, 0.013341357931494713, 0.15123699605464935, -0.06335866451263428, -0.05446095019578934, 0.038721926510334015, -0.046390749514102936, 0.010495561175048351, 0.14334885776042938, 0.06992587447166443, -0.09318824112415314, 0.13813818991184235, 0.08903130888938904, -0.0664111077785492, 0.1002429872751236, -0.04259251058101654, -0.0696166604757309, -0.03386177867650986, -0.02894183248281479, 0.03917427733540535, 0.08256838470697403, -0.09239031374454498, -0.030352886766195297, 0.043511033058166504, 0.0009890517685562372, -0.008213897235691547, -0.16915181279182434, 0.0026915892958641052, 0.058974489569664, -0.030238233506679535, -0.03409462049603462, -0.025955120101571083, 0.01818026974797249, 0.09635532647371292, 0.04368323087692261, -0.06089893728494644, 0.021964961662888527, 0.015598989091813564, -0.07608170062303543, 0.20851881802082062, -0.09314651042222977, -0.14821769297122955, -0.12665319442749023, -0.09736081957817078, -0.03377695381641388, 0.006719908677041531, 0.04770463705062866, -0.08163834363222122, -0.05236327275633812, -0.06571367383003235, 0.01560469251126051, -0.018989868462085724, 0.02078791707754135, 0.03135529160499573, -0.012344110757112503, 0.047738708555698395, -0.10760189592838287, -0.01334760244935751, -0.0031606622505933046, -0.020142177119851112, 0.05941319838166237, 0.03824376314878464, 0.12605039775371552, 0.13806873559951782, -0.04877076297998428, 0.04448278248310089, -0.022514784708619118, 0.24488815665245056, -0.08300559222698212, -0.00886385515332222, 0.13110606372356415, -0.009773404337465763, 0.07199990749359131, 0.11457603424787521, 0.062285203486680984, -0.07006588578224182, -0.0015251360600814223, 0.020874783396720886, -0.04324617609381676, -0.21572653949260712, -0.029788866639137268, -0.04021073132753372, 0.04291744530200958, 0.114588163793087, 0.01870507001876831, 0.06102781370282173, 0.06584127992391586, 0.01750294677913189, 0.0658804252743721, -0.014255985617637634, 0.11287001520395279, 0.11202942579984665, 0.05578096956014633, 0.11638820916414261, -0.04519955813884735, -0.05308223143219948, 0.04673388972878456, -0.030077602714300156, 0.19115613400936127, -0.008648612536489964, 0.15868915617465973, 0.021391401067376137, 0.12224651128053665, -0.004759808070957661, 0.1131674274802208, -0.01063819881528616, -0.01879587024450302, 0.002082858933135867, -0.04027751088142395, -0.03439430892467499, 0.028846224769949913, -0.03653477504849434, 0.0605047270655632, -0.12932653725147247, 0.04461962729692459, 0.05197972059249878, 0.27554526925086975, 0.04612381383776665, -0.3532198667526245, -0.11317561566829681, 0.006160260643810034, -0.049814946949481964, -0.061404719948768616, -0.00393168767914176, 0.047452010214328766, -0.12292704731225967, 0.04447955638170242, -0.07384203374385834, 0.113423652946949, -0.07733477652072906, 0.023787077516317368, 0.029062878340482712, 0.12944664061069489, -0.02309396117925644, 0.07945473492145538, -0.2835526764392853, 0.2394762486219406, 0.02536063827574253, 0.07691124826669693, -0.06904387474060059, 0.019903484731912613, 0.033365990966558456, 0.06498715281486511, 0.06775692850351334, -0.006840039510279894, -0.040776439011096954, -0.18622539937496185, -0.09602384269237518, -0.015568018890917301, 0.08748436719179153, -0.023910703137516975, 0.13362158834934235, -0.03043353743851185, -0.014510409906506538, 0.03860478103160858, -0.05222442001104355, -0.03623996675014496, -0.10337786376476288, 0.017312338575720787, -0.0188243817538023, -0.008660857565701008, -0.05016419291496277, -0.10467718541622162, -0.06324266642332077, 0.17646987736225128, 0.008490970358252525, -0.06935412436723709, -0.1568681299686432, 0.05513680353760719, 0.11056038737297058, -0.08960843086242676, 0.053071413189172745, -0.003235796932131052, 0.10982894152402878, -0.01000207383185625, -0.06694301217794418, 0.10275043547153473, -0.07603808492422104, -0.1884818822145462, -0.0242863018065691, 0.09569137543439865, 0.042298998683691025, 0.06639432162046432, -0.01082518044859171, 0.04479119926691055, -0.03365996852517128, -0.08993493020534515, 0.048571765422821045, -0.019476693123579025, 0.08439906686544418, 0.013458567671477795, -0.011310875415802002, 0.05720757320523262, -0.07076090574264526, -0.010645453818142414, 0.17285217344760895, 0.2533944845199585, -0.08948907256126404, 0.03563416376709938, 0.04963293671607971, -0.04742427542805672, -0.18469075858592987, 0.0005963817820884287, 0.04509074613451958, 0.005866559222340584, 0.009771986864507198, -0.19350796937942505, 0.049014050513505936, 0.05337357148528099, -0.00797485001385212, 0.07363195717334747, -0.2765624225139618, -0.11973359435796738, 0.13981051743030548, 0.11555483192205429, 0.02124384045600891, -0.14603696763515472, -0.03618010878562927, -0.026390451937913895, -0.12388510257005692, 0.09720829129219055, -0.08915214985609055, 0.10625655949115753, -0.0306991059333086, 0.11053011566400528, 0.019946489483118057, -0.05150150507688522, 0.1113906279206276, -0.005507696885615587, 0.0805211067199707, -0.058798059821128845, -0.04146164283156395, 0.05933508649468422, -0.08449076861143112, 0.07383248209953308, -0.06923439353704453, 0.0446130633354187, -0.128191277384758, -0.01685832440853119, -0.06843533366918564, 0.031944483518600464, -0.04720116779208183, -0.055946048349142075, -0.04391203820705414, 0.01648739166557789, 0.06207874044775963, -0.02534737065434456, 0.14415687322616577, 0.02909385785460472, 0.12675459682941437, 0.16800372302532196, 0.10446789115667343, -0.05329560488462448, -0.05151857063174248, 0.02077513188123703, -0.026283208280801773, 0.0464155487716198, -0.13902269303798676, 0.0340690016746521, 0.13830898702144623, 0.03897572308778763, 0.13597184419631958, 0.07076326757669449, -0.054001159965991974, -0.009730313904583454, 0.03790479525923729, -0.1621973216533661, -0.07925239950418472, -0.008651047013700008, -0.014419282786548138, -0.1378723531961441, 0.03749596327543259, 0.1056731641292572, -0.07844660431146622, -0.03332597017288208, 0.01092505268752575, 0.015389167703688145, -0.02167312614619732, 0.1852223128080368, 0.07911032438278198, 0.08417962491512299, -0.11974319815635681, 0.07451879233121872, 0.05636592209339142, -0.07391460239887238, 0.044670768082141876, 0.1427469104528427, -0.08173073828220367, -0.02754511870443821, 0.07660530507564545, 0.1614394634962082, -0.10864352434873581, -0.033472295850515366, -0.15045399963855743, -0.12691345810890198, 0.09582964330911636, 0.17020347714424133, 0.07332021743059158, 0.00011266659566899762, -0.04413291811943054, 0.029473906382918358, -0.171295166015625, 0.12187492102384567, 0.08275920897722244, 0.0768740251660347, -0.14637865126132965, 0.16079497337341309, -0.022769775241613388, 0.071198970079422, -0.02844764105975628, 0.01598403789103031, -0.13284102082252502, 0.006273089908063412, -0.13087411224842072, -0.023536309599876404, -0.06727295368909836, 0.008862558752298355, -0.01482514850795269, -0.03849821910262108, -0.04411003366112709, 0.01228518970310688, -0.10940396785736084, -0.049357373267412186, 0.021204007789492607, 0.020726632326841354, -0.1249946802854538, -0.038451846688985825, 0.023840799927711487, -0.08979973196983337, 0.0800136849284172, 0.06555908918380737, 0.006044989451766014, 0.008517644368112087, -0.02682764083147049, 0.00011215519771212712, 0.04240657389163971, -0.012118050828576088, 0.09462317824363708, -0.14742538332939148, -0.00663081556558609, -0.04226945713162422, 0.014850963838398457, 0.02235184609889984, 0.08857211470603943, -0.11881616711616516, 0.034938689321279526, -0.03446321561932564, -0.06601110845804214, -0.06437152624130249, 0.08986597508192062, 0.09456424415111542, -0.005926546175032854, 0.16001707315444946, -0.07987865805625916, 0.06273125857114792, -0.1980898231267929, -0.002830914221704006, -0.008761242032051086, -0.11228765547275543, -0.08103522658348083, -0.05300288274884224, 0.07513170689344406, -0.06768714636564255, 0.10849374532699585, 0.02383197657763958, 0.04959113150835037, 0.04211173579096794, -0.026776082813739777, -0.012086134403944016, 0.0242288988083601, 0.13684436678886414, 0.02762231044471264, -0.053360465914011, 0.07082805037498474, 0.053825825452804565, 0.09220446646213531, 0.07784523814916611, 0.1709410399198532, 0.08885866403579712, 0.009193923324346542, 0.09344291687011719, 0.033373232930898666, -0.07462179660797119, -0.1809685379266739, 0.00826943013817072, -0.018633633852005005, 0.102104052901268, -0.023436808958649635, 0.17176811397075653, 0.11818337440490723, -0.15184752643108368, 0.06473134458065033, -0.0332256555557251, -0.09551290422677994, -0.1205466240644455, -0.03438883274793625, -0.08693462610244751, -0.1399603635072708, -0.003087263321503997, -0.13074831664562225, 0.03986896947026253, 0.05145009234547615, 0.013737884350121021, -0.0025309922639280558, 0.13053090870380402, 0.038367561995983124, 0.033994950354099274, 0.054348379373550415, 0.0017542768036946654, -0.018112583085894585, -0.05472951382398605, -0.05866643413901329, 0.002547014970332384, -0.016809310764074326, 0.05571713298559189, -0.02548927813768387, -0.036224205046892166, 0.06641349196434021, -0.006192157510668039, -0.12320398539304733, 0.01902756467461586, 0.03199726715683937, 0.05813484638929367, 0.05698760226368904, -0.0029894420877099037, -0.0002847129653673619, -0.03438669815659523, 0.21077367663383484, -0.10647375136613846, -0.06898864358663559, -0.12515464425086975, 0.2852230370044708, 0.03116605058312416, -0.02492241933941841, 0.03289055451750755, -0.07614541053771973, -0.02909969352185726, 0.20914070308208466, 0.19171296060085297, -0.03483067825436592, -0.01311309915035963, -0.008064989931881428, -0.010076791048049927, -0.0029187160544097424, 0.07831000536680222, 0.11281070113182068, 0.03075864166021347, -0.09396521002054214, -0.029232800006866455, -0.06756382435560226, -0.0479758158326149, -0.04866369813680649, 0.09487248212099075, 0.03967811167240143, 0.011436049826443195, -0.026669947430491447, 0.04323539510369301, -0.08075820654630661, -0.11621475219726562, 0.03614136576652527, -0.18233373761177063, -0.14817923307418823, -0.021611876785755157, 0.0192182045429945, 0.0240618959069252, 0.07492081075906754, 0.007340261247009039, -0.022142311558127403, 0.10283692926168442, -0.0023238512221723795, -0.0851510614156723, -0.06999318301677704, 0.10592997819185257, -0.12152725458145142, 0.1898823231458664, -0.043700311332941055, 0.019986845552921295, 0.14456087350845337, 0.02937871217727661, -0.12142761796712875, 0.024524293839931488, 0.060675427317619324, -0.0663222074508667, 0.03048698790371418, 0.1479083150625229, -0.02958863601088524, 0.04730449244379997, 0.013120287097990513, -0.15376661717891693, -0.014981291256844997, -0.07007136195898056, -0.043396443128585815, -0.04255802184343338, -0.03552478179335594, -0.038036223500967026, 0.12554244697093964, 0.23184621334075928, -0.045535359531641006, 0.012299226596951485, -0.07076706737279892, 0.02123684249818325, 0.07066851109266281, 0.0020833266898989677, -0.05673832818865776, -0.2344500869512558, 0.008439742028713226, 0.05863828957080841, -0.0038543420378118753, -0.2489493489265442, -0.07367914915084839, -0.0048732394352555275, -0.05369434133172035, -0.0946018397808075, 0.09030519425868988, 0.08229038119316101, 0.0540124848484993, -0.05561197176575661, -0.005882371682673693, -0.08069183677434921, 0.17176488041877747, -0.1570986658334732, -0.05293416231870651 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-indonesian-sentiment-analysis-smsa This model is a fine-tuned version of [flax-community/indonesian-roberta-base](https://huggingface.co/flax-community/indonesian-roberta-base) on the indonlu dataset. It achieves the following results on the evaluation set: - Loss: 0.4252 - Accuracy: 0.9349 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7582 | 1.0 | 688 | 0.3280 | 0.8786 | | 0.3225 | 2.0 | 1376 | 0.2398 | 0.9206 | | 0.2057 | 3.0 | 2064 | 0.2574 | 0.9230 | | 0.1642 | 4.0 | 2752 | 0.2820 | 0.9302 | | 0.1266 | 5.0 | 3440 | 0.3344 | 0.9317 | | 0.0608 | 6.0 | 4128 | 0.3543 | 0.9341 | | 0.058 | 7.0 | 4816 | 0.4252 | 0.9349 | | 0.0315 | 8.0 | 5504 | 0.4736 | 0.9310 | | 0.0166 | 9.0 | 6192 | 0.4649 | 0.9349 | | 0.0143 | 10.0 | 6880 | 0.4648 | 0.9341 | ### Framework versions - Transformers 4.14.1 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["indonlu"], "metrics": ["accuracy"], "model-index": [{"name": "roberta-base-indonesian-sentiment-analysis-smsa", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "indonlu", "type": "indonlu", "args": "smsa"}, "metrics": [{"type": "accuracy", "value": 0.9349206349206349, "name": "Accuracy"}]}]}]}
text-classification
ayameRushia/roberta-base-indonesian-sentiment-analysis-smsa
[ "transformers", "pytorch", "roberta", "text-classification", "generated_from_trainer", "dataset:indonlu", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #roberta #text-classification #generated_from_trainer #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
roberta-base-indonesian-sentiment-analysis-smsa =============================================== This model is a fine-tuned version of flax-community/indonesian-roberta-base on the indonlu dataset. It achieves the following results on the evaluation set: * Loss: 0.4252 * Accuracy: 0.9349 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.14.1 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #roberta #text-classification #generated_from_trainer #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ 60, 116, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #text-classification #generated_from_trainer #dataset-indonlu #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.14.1\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ -0.09738113731145859, 0.08291201293468475, -0.002051000948995352, 0.12836791574954987, 0.17542676627635956, 0.036752812564373016, 0.12750540673732758, 0.13863885402679443, -0.078737773001194, 0.020604539662599564, 0.12331770360469818, 0.1622711569070816, 0.014937324449419975, 0.14365322887897491, -0.05351385101675987, -0.3064621090888977, -0.012232999317348003, 0.018413161858916283, -0.06637502461671829, 0.13838152587413788, 0.0921320915222168, -0.13162773847579956, 0.09971759468317032, -0.0035812726709991693, -0.17971310019493103, 0.0033037245739251375, 0.006059189792722464, -0.05584248900413513, 0.15433752536773682, 0.009814964607357979, 0.11040148138999939, 0.009810058400034904, 0.11563274264335632, -0.18411803245544434, 0.009839147329330444, 0.04285239428281784, 0.014751477167010307, 0.09340714663267136, 0.05616751313209534, -0.007088382262736559, 0.1406148225069046, -0.07996656745672226, 0.06311438977718353, 0.010443752631545067, -0.13705644011497498, -0.19802114367485046, -0.07454996556043625, 0.03410821408033371, 0.062492795288562775, 0.09621694684028625, -0.00891158264130354, 0.11991872638463974, -0.09174854308366776, 0.11030317842960358, 0.21355997025966644, -0.2638966143131256, -0.06670931726694107, 0.050014860928058624, 0.021006759256124496, 0.05665678158402443, -0.1244921013712883, -0.02566547319293022, 0.02987523563206196, 0.043109480291604996, 0.12822692096233368, -0.037209317088127136, -0.10892992466688156, 0.018081581220030785, -0.13890977203845978, -0.03176507353782654, 0.11606717854738235, 0.032312821596860886, -0.028159618377685547, -0.04418068379163742, -0.0552191436290741, -0.16686022281646729, -0.054668620228767395, -0.014889810234308243, 0.04709629341959953, -0.03482400253415108, -0.08360031247138977, -0.0027263755910098553, -0.09228961914777756, -0.07367963343858719, -0.06139051914215088, 0.15759581327438354, 0.03985758498311043, 0.0021673813462257385, -0.026414986699819565, 0.1106465607881546, 0.00807897001504898, -0.1377602219581604, 0.016493290662765503, 0.02195737697184086, -0.0274774469435215, -0.06128684803843498, -0.049776591360569, -0.01732841320335865, 0.003139170352369547, 0.141165629029274, -0.07202714681625366, 0.04468785971403122, 0.04956697300076485, 0.019642779603600502, -0.07690098136663437, 0.18399199843406677, -0.06169100105762482, -0.03448810055851936, -0.010691884905099869, 0.05708172544836998, -0.012514016591012478, -0.009458628483116627, -0.10489465296268463, -0.0028210701420903206, 0.10954399406909943, 0.015926072373986244, -0.06647210568189621, 0.0876355990767479, -0.039468128234148026, -0.030126824975013733, 0.004677072167396545, -0.0892440527677536, 0.0395204983651638, -0.0014529814943671227, -0.102033831179142, -0.041309233754873276, 0.016773656010627747, 0.01866484433412552, -0.0008309702971018851, 0.14184395968914032, -0.10033542662858963, 0.03548802435398102, -0.08936233818531036, -0.11959262192249298, 0.005590452346950769, -0.09104263037443161, 0.022896116599440575, -0.08300866186618805, -0.2098858505487442, -0.024344462901353836, 0.056553054600954056, -0.05160092934966087, -0.035772982984781265, -0.06333950161933899, -0.07163670659065247, 0.01929452456533909, -0.009252512827515602, 0.12904883921146393, -0.06639266759157181, 0.11558979004621506, 0.02253340370953083, 0.06961509585380554, -0.022274712100625038, 0.06248054653406143, -0.11427047103643417, -0.0025992554146796465, -0.18479464948177338, 0.05363859981298447, -0.044380590319633484, 0.06508947163820267, -0.08959711343050003, -0.11182206869125366, 0.008560150861740112, 0.009214314632117748, 0.07340151816606522, 0.1266263723373413, -0.17710484564304352, -0.07346270233392715, 0.16026350855827332, -0.06819549202919006, -0.10633692145347595, 0.10765785723924637, -0.07617193460464478, 0.07477322220802307, 0.06747088581323624, 0.17750252783298492, 0.06587192416191101, -0.08302798867225647, 0.01732485555112362, -0.01401920523494482, 0.029332518577575684, -0.03874066472053528, 0.06310994178056717, 0.019228512421250343, 0.02219265326857567, 0.0235338993370533, -0.012235468253493309, 0.060779549181461334, -0.11562901735305786, -0.08785098791122437, -0.024616174399852753, -0.08911864459514618, 0.05599839985370636, 0.07604087144136429, 0.07522077113389969, -0.09369881451129913, -0.06906233727931976, 0.06532558053731918, 0.09338922798633575, -0.0480629988014698, 0.016790270805358887, -0.05927431955933571, 0.04154730215668678, -0.010811764746904373, -0.02737397700548172, -0.19599847495555878, -0.020120104774832726, 0.008460487239062786, 0.03804144263267517, 0.03569728508591652, 0.016917835921049118, 0.07946288585662842, 0.061853330582380295, -0.05843716487288475, -0.02283610589802265, -0.007658396381884813, 0.002263224683701992, -0.12832307815551758, -0.210930734872818, -0.023497430607676506, -0.02480790950357914, 0.1299135982990265, -0.2182222306728363, 0.027584819123148918, -0.0375431627035141, 0.08018506318330765, 0.02030917815864086, -0.012558194808661938, -0.029949570074677467, 0.070867158472538, -0.03419371321797371, -0.052577074617147446, 0.07274068892002106, -0.004368371330201626, -0.07067135721445084, -0.04803341627120972, -0.09545346349477768, 0.15079350769519806, 0.11766278743743896, -0.09797258675098419, -0.10716857016086578, -0.008737980388104916, -0.05547838285565376, -0.03546224907040596, -0.04618135094642639, 0.04132992401719093, 0.16860714554786682, 0.0004593951744027436, 0.14656242728233337, -0.06011393666267395, -0.03280690684914589, 0.02474055252969265, -0.03337541222572327, 0.0276432316750288, 0.1320342719554901, 0.12481117993593216, -0.12017776817083359, 0.1356016993522644, 0.10617699474096298, -0.06403912603855133, 0.15217377245426178, -0.02696419693529606, -0.06737779080867767, -0.028583547100424767, -0.03428153693675995, -0.013213828206062317, 0.09303015470504761, -0.10662871599197388, 0.00011241027095820755, 0.01859530620276928, 0.025563308969140053, 0.0076435930095613, -0.20378915965557098, -0.04538022726774216, 0.027204886078834534, -0.0465453565120697, -0.03171400725841522, 0.009030049666762352, 0.017209632322192192, 0.11780901253223419, 0.007780800107866526, -0.10452674329280853, 0.018075354397296906, 0.00586672592908144, -0.07301679998636246, 0.21370241045951843, -0.07990214228630066, -0.15816882252693176, -0.11423222720623016, -0.09538353234529495, -0.047778431326150894, 0.011583776213228703, 0.056450311094522476, -0.08344060927629471, -0.01878543011844158, -0.062178295105695724, 0.022599197924137115, -0.003590845502912998, 0.028104865923523903, -0.022542405873537064, 0.002115346025675535, 0.043402478098869324, -0.10348860174417496, -0.013787791132926941, -0.06585806608200073, -0.05077582225203514, 0.0576351061463356, 0.03460973873734474, 0.11031129211187363, 0.13518548011779785, -0.020981397479772568, 0.016662701964378357, -0.03826583921909332, 0.24963054060935974, -0.07667648047208786, -0.0040848106145858765, 0.13763634860515594, -0.00015057034033816308, 0.05046076700091362, 0.14918245375156403, 0.06977161020040512, -0.09175323694944382, 0.012074054218828678, 0.036064159125089645, -0.026502396911382675, -0.2031455636024475, -0.05096297711133957, -0.04890923202037811, -0.006060848478227854, 0.10739810019731522, 0.0228958111256361, 0.035432666540145874, 0.06324858218431473, 0.04417300596833229, 0.055778250098228455, -0.03372587263584137, 0.06422901898622513, 0.11269155144691467, 0.04534927010536194, 0.1284191608428955, -0.03177056834101677, -0.08178840577602386, 0.04466671869158745, -0.033148232847452164, 0.2281903177499771, -0.00010355489212088287, 0.0919414684176445, 0.04731091484427452, 0.14443780481815338, -0.0004659570986405015, 0.08535198122262955, 0.005494314711540937, -0.04202789440751076, -0.021649464964866638, -0.0307372584939003, -0.04191974550485611, 0.02067592926323414, -0.04609498381614685, 0.038368988782167435, -0.1299927532672882, 0.010964097455143929, 0.047326140105724335, 0.24386626482009888, 0.039099227637052536, -0.3252081274986267, -0.09139738976955414, -0.007131498772650957, -0.044116389006376266, -0.03226722404360771, 0.01986294984817505, 0.0844026580452919, -0.11047246307134628, 0.028464151546359062, -0.0753796398639679, 0.08926736563444138, -0.06445670872926712, 0.048335492610931396, 0.0759955570101738, 0.10304449498653412, -0.015170660801231861, 0.077295683324337, -0.2909773588180542, 0.30309346318244934, 0.005538379307836294, 0.06760264188051224, -0.07652883976697922, -0.008403478190302849, 0.04854941368103027, 0.06676542013883591, 0.07391103357076645, -0.014122329652309418, -0.048545464873313904, -0.22228582203388214, -0.05481969565153122, 0.015305653214454651, 0.10515270382165909, -0.04345158860087395, 0.10636300593614578, -0.04151316359639168, 0.004376313183456659, 0.06875535100698471, -0.049107808619737625, -0.06363698095083237, -0.07434742152690887, -0.0029568932950496674, 0.017873503267765045, -0.01999148726463318, -0.044883254915475845, -0.11667972803115845, -0.11012543737888336, 0.11488540470600128, 0.020265797153115273, -0.028979266062378883, -0.12668932974338531, 0.10159528255462646, 0.10842856764793396, -0.09975960850715637, 0.037031520158052444, 0.008921509608626366, 0.051109734922647476, 0.039556361734867096, -0.059723976999521255, 0.11424432694911957, -0.06961281597614288, -0.17057111859321594, -0.05893092602491379, 0.09165730327367783, 0.04029723256826401, 0.06477910280227661, -0.010629137046635151, 0.031683776527643204, -0.04220302402973175, -0.08633603900671005, 0.024273954331874847, -0.03733091428875923, 0.07158131152391434, 0.041272375732660294, -0.07172953337430954, 0.019719138741493225, -0.05910457298159599, -0.0323726087808609, 0.1830950379371643, 0.24854378402233124, -0.0977858230471611, 0.03491148352622986, 0.046855390071868896, -0.06755702197551727, -0.2016225904226303, 0.03194575384259224, 0.05868292972445488, 0.0028050444088876247, 0.06426860392093658, -0.19821883738040924, 0.10016458481550217, 0.08854446560144424, -0.010512136854231358, 0.08797264099121094, -0.30143773555755615, -0.12612955272197723, 0.1236136332154274, 0.13600163161754608, 0.10471237450838089, -0.13661353290081024, -0.013420140370726585, 0.0008840117370709777, -0.09787020832300186, 0.10083185136318207, -0.03864568844437599, 0.14032447338104248, -0.03507707640528679, 0.10194995999336243, 0.016272999346256256, -0.050857868045568466, 0.09361555427312851, 0.025878751650452614, 0.11375945061445236, -0.06202458590269089, -0.057699304074048996, 0.016023622825741768, -0.04055904597043991, 0.031838566064834595, -0.059626560658216476, 0.036681853234767914, -0.10724565386772156, -0.01806122437119484, -0.08964803069829941, 0.032562077045440674, -0.035767581313848495, -0.07410713285207748, -0.034533996134996414, 0.06249479576945305, 0.048730749636888504, -0.020926235243678093, 0.12261680513620377, 0.0029747099615633488, 0.13745738565921783, 0.08772795647382736, 0.07224131375551224, -0.04583212733268738, -0.03661218285560608, -0.015210901387035847, -0.008008048869669437, 0.032419025897979736, -0.12778465449810028, 0.02370483987033367, 0.14393775165081024, 0.0258623119443655, 0.15187764167785645, 0.08127395808696747, -0.011276227422058582, -0.0028130412101745605, 0.07288582623004913, -0.16424141824245453, -0.0657692551612854, -0.010972809977829456, -0.09915971010923386, -0.10772832483053207, 0.038325805217027664, 0.10274241864681244, -0.0671202540397644, -0.005316461902111769, -0.010139522142708302, 0.019940432161092758, -0.03551902249455452, 0.20074908435344696, 0.060266051441431046, 0.046151116490364075, -0.10702165216207504, 0.06745348125696182, 0.05152878165245056, -0.045201271772384644, 0.0073084053583443165, 0.09589162468910217, -0.08918096125125885, -0.043029483407735825, 0.06203576549887657, 0.195978045463562, -0.07976333796977997, -0.022024771198630333, -0.14633651077747345, -0.1050054058432579, 0.07963944971561432, 0.15962673723697662, 0.10203630477190018, 0.004589000251144171, -0.06933873146772385, 0.011664057150483131, -0.1287599503993988, 0.1023716852068901, 0.059104613959789276, 0.0610961839556694, -0.14148569107055664, 0.16771399974822998, -0.008407129906117916, 0.03903021290898323, -0.027488980442285538, 0.017660658806562424, -0.11952457576990128, 0.00887452345341444, -0.08389722555875778, -0.04344423487782478, -0.041157517582178116, 0.007475164718925953, -0.007646312937140465, -0.06885627657175064, -0.05429952219128609, -0.008498839102685452, -0.1169717013835907, -0.017316395416855812, 0.028554627671837807, 0.06833663582801819, -0.12396826595067978, -0.04793228209018707, 0.021141622215509415, -0.05234362185001373, 0.07920458912849426, 0.05040178447961807, 0.014094941318035126, 0.0508282333612442, -0.13251373171806335, 0.010125498287379742, 0.04361478239297867, 0.002652629977092147, 0.07231294363737106, -0.09915360808372498, 0.0017998467665165663, -0.022551508620381355, 0.06485830247402191, 0.03080376237630844, 0.0596003383398056, -0.13096648454666138, 0.027079861611127853, -0.03212776780128479, -0.07415467500686646, -0.06737218797206879, 0.03891310468316078, 0.08186788111925125, 0.013372126035392284, 0.18999508023262024, -0.08060228079557419, 0.05354318395256996, -0.21958889067173004, -0.00043291738256812096, -0.008811899460852146, -0.1178395003080368, -0.11594908684492111, -0.09029975533485413, 0.07730673998594284, -0.05013662949204445, 0.1307634711265564, 0.06363900750875473, 0.051739953458309174, 0.02147023007273674, -0.02335982583463192, 0.0039151739329099655, 0.0060676573775708675, 0.1778697818517685, 0.043620385229587555, -0.04884621873497963, 0.07791262865066528, 0.06595205515623093, 0.09437304735183716, 0.11424774676561356, 0.21606853604316711, 0.14018751680850983, 0.003997644875198603, 0.07417034357786179, 0.04357767850160599, -0.06972314417362213, -0.14595866203308105, 0.0264106597751379, -0.031909141689538956, 0.08317781239748001, -0.03125525265932083, 0.20897404849529266, 0.05430549383163452, -0.17504335939884186, 0.054247431457042694, -0.057004231959581375, -0.08957791328430176, -0.1154865175485611, -0.03114163875579834, -0.08800053596496582, -0.1398671418428421, 0.009023836813867092, -0.12338051944971085, 0.03521968051791191, 0.09863989800214767, 0.01397067029029131, -0.020324276760220528, 0.13444863259792328, -0.0030749088618904352, 0.03006981685757637, 0.061404433101415634, 0.008157733827829361, -0.016732970252633095, -0.10649165511131287, -0.07116860151290894, -0.028206555172801018, -0.006412872113287449, 0.028515920042991638, -0.06661851704120636, -0.07134365290403366, 0.03211423382163048, -0.02915320359170437, -0.10388439148664474, 0.014785703271627426, 0.022894810885190964, 0.07027175277471542, 0.05180327221751213, 0.011971142143011093, 0.013507939875125885, -0.012738223187625408, 0.25265032052993774, -0.0775439590215683, -0.07994487136602402, -0.10259805619716644, 0.2705097198486328, 0.04934559017419815, -0.011221006512641907, 0.026685060933232307, -0.06893272697925568, -0.005103966221213341, 0.25811752676963806, 0.2139165699481964, -0.10422740876674652, -0.0056063104420900345, 0.010229809209704399, -0.007732993457466364, -0.0012735610362142324, 0.10883945226669312, 0.11112690716981888, 0.05104583129286766, -0.09519947320222855, -0.02890409715473652, -0.05984153598546982, -0.027635741978883743, -0.026874400675296783, 0.07652916014194489, 0.0530332550406456, 0.009357929229736328, -0.040664076805114746, 0.0757954940199852, -0.08953683823347092, -0.12281832098960876, 0.042964037507772446, -0.21878434717655182, -0.16360634565353394, -0.022278087213635445, 0.07991804927587509, 0.02632526122033596, 0.06581931561231613, -0.02538270689547062, -0.002194066997617483, 0.04379107058048248, -0.013825812377035618, -0.0750921219587326, -0.09453186392784119, 0.09902308881282806, -0.08233408629894257, 0.19528552889823914, -0.050631333142519, 0.05156385153532028, 0.12542948126792908, 0.05407843366265297, -0.08152087777853012, 0.06367997825145721, 0.04419912025332451, -0.07771217823028564, 0.03677009791135788, 0.10430700331926346, -0.03704607114195824, 0.056680407375097275, 0.04148773103952408, -0.14211662113666534, 0.030302487313747406, -0.06945883482694626, -0.06618468463420868, -0.042398933321237564, -0.03444533050060272, -0.03665757179260254, 0.1418762356042862, 0.23578542470932007, -0.02370019070804119, 0.015408427454531193, -0.07469628751277924, 0.01917145773768425, 0.05262527987360954, 0.015585597604513168, -0.08141644299030304, -0.2342967391014099, 0.004101764410734177, 0.0763920247554779, -0.01433812826871872, -0.23585323989391327, -0.09030577540397644, 0.004088624846190214, -0.06713950634002686, -0.09608475118875504, 0.100770965218544, 0.06612107902765274, 0.04816534370183945, -0.04082684591412544, -0.06543397903442383, -0.07025980204343796, 0.16668400168418884, -0.16563421487808228, -0.08172247558832169 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-ar This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.4819 - Wer: 0.4244 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 400 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 11.0435 | 0.67 | 400 | 4.3104 | 1.0 | | 3.4451 | 1.34 | 800 | 3.1566 | 1.0 | | 3.1399 | 2.01 | 1200 | 3.0532 | 0.9990 | | 2.8538 | 2.68 | 1600 | 1.6994 | 0.9238 | | 1.7195 | 3.35 | 2000 | 0.8867 | 0.6727 | | 1.326 | 4.02 | 2400 | 0.6603 | 0.5834 | | 1.1561 | 4.69 | 2800 | 0.5809 | 0.5479 | | 1.0764 | 5.36 | 3200 | 0.5943 | 0.5495 | | 1.0144 | 6.03 | 3600 | 0.5344 | 0.5251 | | 0.965 | 6.7 | 4000 | 0.4844 | 0.4936 | | 0.927 | 7.37 | 4400 | 0.5048 | 0.5019 | | 0.8985 | 8.04 | 4800 | 0.5809 | 0.5267 | | 0.8684 | 8.71 | 5200 | 0.4740 | 0.4753 | | 0.8581 | 9.38 | 5600 | 0.4813 | 0.4834 | | 0.8334 | 10.05 | 6000 | 0.4515 | 0.4545 | | 0.8134 | 10.72 | 6400 | 0.4370 | 0.4543 | | 0.8002 | 11.39 | 6800 | 0.4225 | 0.4384 | | 0.7884 | 12.06 | 7200 | 0.4593 | 0.4565 | | 0.7675 | 12.73 | 7600 | 0.4752 | 0.4680 | | 0.7607 | 13.4 | 8000 | 0.4950 | 0.4771 | | 0.7475 | 14.07 | 8400 | 0.4373 | 0.4391 | | 0.7397 | 14.74 | 8800 | 0.4506 | 0.4541 | | 0.7289 | 15.41 | 9200 | 0.4840 | 0.4691 | | 0.722 | 16.08 | 9600 | 0.4701 | 0.4571 | | 0.7067 | 16.75 | 10000 | 0.4561 | 0.4461 | | 0.7033 | 17.42 | 10400 | 0.4384 | 0.4347 | | 0.6915 | 18.09 | 10800 | 0.4424 | 0.4290 | | 0.6854 | 18.76 | 11200 | 0.4635 | 0.4360 | | 0.6813 | 19.43 | 11600 | 0.4280 | 0.4147 | | 0.6776 | 20.1 | 12000 | 0.4610 | 0.4344 | | 0.67 | 20.77 | 12400 | 0.4540 | 0.4367 | | 0.6653 | 21.44 | 12800 | 0.4509 | 0.4234 | | 0.6609 | 22.11 | 13200 | 0.4874 | 0.4444 | | 0.6541 | 22.78 | 13600 | 0.4542 | 0.4230 | | 0.6528 | 23.45 | 14000 | 0.4732 | 0.4373 | | 0.6463 | 24.12 | 14400 | 0.4483 | 0.4188 | | 0.6399 | 24.79 | 14800 | 0.4731 | 0.4341 | | 0.6353 | 25.46 | 15200 | 0.5031 | 0.4412 | | 0.6358 | 26.13 | 15600 | 0.4986 | 0.4397 | | 0.6317 | 26.8 | 16000 | 0.5000 | 0.4360 | | 0.6262 | 27.47 | 16400 | 0.4958 | 0.4318 | | 0.6317 | 28.14 | 16800 | 0.4738 | 0.4234 | | 0.6205 | 28.81 | 17200 | 0.4853 | 0.4262 | | 0.6205 | 29.48 | 17600 | 0.4819 | 0.4244 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-ar", "results": []}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xls-r-300m-ar
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-large-xls-r-300m-ar ============================ This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common\_voice dataset. It achieves the following results on the evaluation set: * Loss: 0.4819 * Wer: 0.4244 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 3e-05 * train\_batch\_size: 32 * eval\_batch\_size: 4 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 64 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 400 * num\_epochs: 30 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 61, 159, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.13469435274600983, 0.07378916442394257, -0.0023688238579779863, 0.041209857910871506, 0.13552002608776093, -0.0017890377203002572, 0.0910828486084938, 0.1462439000606537, -0.09717334061861038, 0.07566332072019577, 0.09145790338516235, 0.09167617559432983, 0.06848963350057602, 0.1123962327837944, -0.024251354858279228, -0.31841230392456055, 0.01948319375514984, 0.01958228461444378, -0.12592005729675293, 0.11913465708494186, 0.12040463835000992, -0.10578016191720963, 0.017216259613633156, 0.03838644549250603, -0.13524717092514038, 0.017375130206346512, -0.017448917031288147, -0.06870976090431213, 0.11853905767202377, 0.035819847136735916, 0.09627665579319, 0.010294681414961815, 0.08981528878211975, -0.26463717222213745, 0.01370159350335598, 0.06034580245614052, 0.052195124328136444, 0.06837140023708344, 0.10149039328098297, -0.011271324940025806, 0.13636167347431183, -0.06873644143342972, 0.06635504961013794, 0.058880068361759186, -0.11099220812320709, -0.34343811869621277, -0.09285309165716171, 0.030706647783517838, 0.11832870543003082, 0.10157707333564758, -0.032949477434158325, 0.07562542706727982, -0.07030462473630905, 0.08501500636339188, 0.2164088487625122, -0.2376488596200943, -0.08044738322496414, -0.03789670020341873, 0.04646528512239456, 0.024573879316449165, -0.12458081543445587, -0.036877766251564026, 0.04024125263094902, 0.04458457976579666, 0.08223646134138107, 0.01657208986580372, -0.044042956084012985, 0.01246966328471899, -0.13224387168884277, -0.06201687827706337, 0.1556132584810257, 0.08725865185260773, -0.04547642171382904, -0.07824966311454773, -0.0011520935222506523, -0.19631093740463257, -0.04423356056213379, 0.014386848546564579, 0.026812860742211342, -0.037074435502290726, -0.11444352567195892, 0.016622111201286316, -0.09399972856044769, -0.09530109167098999, 0.014148731715977192, 0.16769538819789886, 0.04770675674080849, -0.02477758377790451, 0.0020727303344756365, 0.10155197978019714, 0.03600809723138809, -0.1393929272890091, -0.012987441383302212, 0.04890244081616402, -0.09150057286024094, -0.018201397731900215, -0.0497392900288105, -0.028176188468933105, 0.0019439734751358628, 0.11887383460998535, -0.027179202064871788, 0.07779285311698914, 0.012576735578477383, 0.024416819214820862, -0.09426745027303696, 0.19234713912010193, -0.06619704514741898, -0.0038721945602446795, -0.04277757555246353, 0.08654456585645676, -0.02748543582856655, -0.008819665759801865, -0.050761628895998, 0.017378950491547585, 0.10933803766965866, 0.04663586989045143, -0.034524619579315186, 0.009024243801832199, -0.06283129751682281, -0.022583013400435448, -0.023493867367506027, -0.10554970800876617, 0.022906895726919174, 0.021752381697297096, -0.0926571860909462, 0.01047680526971817, 0.003481574123725295, 0.029697956517338753, -0.01872900500893593, 0.09723427146673203, -0.06284128874540329, 0.006607856601476669, -0.08928706496953964, -0.10637558996677399, 0.03341318294405937, -0.01745384745299816, 0.009376795962452888, -0.06760048866271973, -0.11839506775140762, -0.05869212746620178, 0.04973448067903519, -0.037846360355615616, -0.07659346610307693, -0.0700334683060646, -0.07133382558822632, 0.04740482568740845, -0.0325821153819561, 0.1710852086544037, -0.057399336248636246, 0.11830870062112808, 0.0419917069375515, 0.03259734436869621, 0.020337853580713272, 0.07403131574392319, -0.04987359419465065, 0.03231262043118477, -0.11001867800951004, 0.07251815497875214, -0.08420491963624954, 0.06615161895751953, -0.1418767124414444, -0.13259053230285645, -0.01834309846162796, 0.0008990130154415965, 0.1081107035279274, 0.08858523517847061, -0.18515267968177795, -0.0919961929321289, 0.17012977600097656, -0.07638023048639297, -0.1153235211968422, 0.12948453426361084, -0.021930640563368797, 0.02626102790236473, 0.047755029052495956, 0.17307493090629578, 0.08083852380514145, -0.09119272977113724, 0.012966198846697807, -0.06256613880395889, 0.1309027224779129, 0.01877063326537609, 0.10731171816587448, -0.05464492365717888, 0.015811067074537277, -0.00010480054334038869, -0.028523795306682587, 0.06753397732973099, -0.10364490002393723, -0.0842060074210167, -0.024514861404895782, -0.09003562480211258, -0.004310780204832554, 0.06437209248542786, 0.04594724252820015, -0.10988394916057587, -0.11612116545438766, 0.03710677847266197, 0.11892522126436234, -0.10691497474908829, 0.03248346224427223, -0.08078031986951828, 0.03522293642163277, -0.01767163909971714, -0.022171227261424065, -0.16367246210575104, -0.012724142521619797, 0.03156348690390587, -0.03857916593551636, 0.02666155993938446, -0.02642745152115822, 0.07969874143600464, 0.03404391556978226, -0.044819176197052, -0.0698469877243042, -0.08804623037576675, -0.010600359179079533, -0.06929994374513626, -0.20065996050834656, -0.07820963114500046, -0.02129231207072735, 0.15555010735988617, -0.22151623666286469, 0.006307412404567003, 0.017481690272688866, 0.10947705060243607, 0.028004370629787445, -0.05309692397713661, -0.011438634246587753, 0.07879053801298141, -0.016104038804769516, -0.061315521597862244, 0.031895458698272705, 0.011156768538057804, -0.11941085755825043, 0.029181618243455887, -0.1035642921924591, 0.09713032096624374, 0.10035121440887451, -0.03609829023480415, -0.07262532413005829, -0.06045282259583473, -0.06864514201879501, -0.06355114281177521, -0.025826996192336082, -0.010485455393791199, 0.19258765876293182, 0.02756073884665966, 0.1192171722650528, -0.08026696741580963, -0.04253687337040901, 0.02649834007024765, 0.00007727612683083862, 0.0006768279708921909, 0.12899911403656006, 0.04651418700814247, -0.04283959046006203, 0.09403084963560104, 0.06983944028615952, -0.0790487602353096, 0.14997541904449463, -0.08131785690784454, -0.12614743411540985, -0.011690705083310604, 0.024195801466703415, 0.03214698284864426, 0.11426633596420288, -0.14995768666267395, -0.0008844295516610146, 0.02563512697815895, 0.037635110318660736, 0.03323303908109665, -0.22141806781291962, -0.01751687563955784, 0.04023023694753647, -0.07236277312040329, -0.0674092248082161, 0.00037163536762818694, -0.0028340991120785475, 0.08373700827360153, 0.004486642777919769, -0.048228003084659576, -0.01193778682500124, -0.029862629249691963, -0.08897685259580612, 0.19187386333942413, -0.10366244614124298, -0.14455954730510712, -0.1449786275625229, -0.03038160316646099, -0.0014454867923632264, -0.01319570280611515, 0.051422689110040665, -0.12348728626966476, -0.03782506659626961, -0.051246508955955505, 0.05556856840848923, -0.07046311348676682, 0.022192325443029404, 0.004061609506607056, 0.018255501985549927, 0.09476533532142639, -0.11336542665958405, 0.021501027047634125, -0.018224431201815605, -0.05126029998064041, 0.02274484932422638, 0.0163286030292511, 0.09535902738571167, 0.15908895432949066, 0.03939414769411087, 0.025589922443032265, -0.03799610957503319, 0.16956378519535065, -0.10421106964349747, -0.051991723477840424, 0.1243579313158989, 0.014249583706259727, 0.03160804137587547, 0.10573749244213104, 0.052929434925317764, -0.09519221633672714, 0.035925544798374176, 0.050791531801223755, -0.014520802535116673, -0.25488710403442383, -0.02702466771006584, -0.06790836155414581, -0.0419473759829998, 0.11421752721071243, 0.028154192492365837, -0.006469592917710543, 0.028549419716000557, -0.013802324421703815, -0.005070336628705263, 0.004838823806494474, 0.0669734850525856, 0.09186583012342453, 0.03970685973763466, 0.11517056822776794, -0.020382042974233627, -0.037770308554172516, 0.02550559863448143, 0.009438573382794857, 0.26250189542770386, 0.010859166271984577, 0.1719130575656891, 0.061306584626436234, 0.1580212414264679, 0.003540792502462864, 0.08768545836210251, 0.017758440226316452, -0.026031138375401497, 0.02468509040772915, -0.05044442415237427, -0.035650935024023056, 0.038865990936756134, 0.08366893231868744, 0.04141750559210777, -0.14324188232421875, -0.039464790374040604, 0.01362605020403862, 0.371014803647995, 0.07152482867240906, -0.29861214756965637, -0.10419770330190659, -0.0033782096579670906, -0.09938400238752365, -0.05426563695073128, 0.030571598559617996, 0.09484288841485977, -0.10468149930238724, 0.05096694827079773, -0.06282185018062592, 0.11025726050138474, -0.042268019169569016, -0.003729188581928611, 0.0732765644788742, 0.0658981055021286, 0.0010959728388115764, 0.07610619068145752, -0.2746231257915497, 0.31372135877609253, -0.02401876263320446, 0.0881630927324295, -0.030091997236013412, 0.02657630667090416, 0.036548275500535965, -0.04886847361922264, 0.050623517483472824, -0.011122959665954113, -0.09857743978500366, -0.19699572026729584, -0.07037841528654099, 0.03861099109053612, 0.12719935178756714, -0.04656432196497917, 0.1208285540342331, -0.03131389617919922, 0.0046197050251066685, 0.06350113451480865, -0.07624185085296631, -0.1115795224905014, -0.09206945449113846, 0.010651940479874611, 0.029689837247133255, 0.10654562711715698, -0.11224449425935745, -0.11404616385698318, -0.058215655386447906, 0.1435101330280304, -0.057324618101119995, -0.009403517469763756, -0.12470034509897232, 0.09080224484205246, 0.17986848950386047, -0.06427302211523056, 0.056405287235975266, 0.024388326331973076, 0.12626758217811584, 0.0379616916179657, 0.005859441123902798, 0.09811802208423615, -0.07956679165363312, -0.19099068641662598, -0.04651496186852455, 0.16841080784797668, 0.04359111934900284, 0.06577351689338684, -0.015933586284518242, 0.025434833019971848, -0.0359194241464138, -0.074410080909729, 0.05058268457651138, -0.000714299560058862, 0.005388566758483648, 0.06927312165498734, -0.034072332084178925, 0.02732420526444912, -0.08630587160587311, -0.07640170305967331, 0.16443896293640137, 0.2853694558143616, -0.06467364728450775, 0.005684003699570894, 0.020869027823209763, -0.04700491949915886, -0.1303137242794037, 0.03370973467826843, 0.15271365642547607, 0.045475080609321594, -0.005109791643917561, -0.234331414103508, 0.05679506063461304, 0.09734165668487549, -0.025113781914114952, 0.08086982369422913, -0.2952686846256256, -0.13568350672721863, 0.11864110082387924, 0.10095549374818802, -0.02401653863489628, -0.1481538563966751, -0.05688831955194473, -0.02448776178061962, -0.12185592204332352, 0.07616695761680603, -0.007488556206226349, 0.12341434508562088, -0.0061565181240439415, 0.08702178299427032, 0.02617567591369152, -0.04741300642490387, 0.15868966281414032, -0.006662493571639061, 0.05432552471756935, 0.002329588867723942, 0.07172542810440063, 0.0020375449676066637, -0.045231256633996964, 0.01435034442692995, -0.06033909320831299, 0.020033150911331177, -0.14912745356559753, -0.03549144044518471, -0.09807074815034866, 0.028743360191583633, -0.03305714949965477, -0.034092847257852554, -0.014226270839571953, 0.041571930050849915, 0.05155027657747269, 0.011339918710291386, 0.129498690366745, -0.057029884308576584, 0.16471998393535614, 0.054491445422172546, 0.09579233080148697, -0.009659572504460812, -0.0826793760061264, -0.009927934966981411, -0.011129851453006268, 0.04777522012591362, -0.11990360915660858, 0.03291851282119751, 0.15130679309368134, 0.0515017956495285, 0.16175684332847595, 0.05935252830386162, -0.07900525629520416, 0.024910489097237587, 0.07473236322402954, -0.05517518147826195, -0.10525406152009964, -0.017633818089962006, 0.07248656451702118, -0.14523664116859436, 0.006918689236044884, 0.09429628401994705, -0.05280108004808426, -0.017385659739375114, 0.0014545498415827751, 0.014279027469456196, -0.07443840056657791, 0.2351742833852768, 0.03869939222931862, 0.08347594738006592, -0.09439300000667572, 0.08215101063251495, 0.04669364169239998, -0.1525934338569641, 0.017183661460876465, 0.06653590500354767, -0.030382363125681877, -0.01115984097123146, 0.012046189047396183, 0.06404974311590195, -0.019193531945347786, -0.0720650926232338, -0.12070922553539276, -0.1467827409505844, 0.08068373054265976, 0.09281002730131149, 0.040208153426647186, 0.03439589589834213, -0.03880731388926506, 0.04223204031586647, -0.11484982818365097, 0.07547655701637268, 0.10738866031169891, 0.07082825899124146, -0.12996722757816315, 0.15824726223945618, 0.012624511495232582, 0.011906358413398266, 0.013613250106573105, -0.01662943698465824, -0.08190129697322845, 0.042044710367918015, -0.12476081401109695, -0.03385234996676445, -0.04697548970580101, -0.0048907059244811535, 0.0186860840767622, -0.06310631334781647, -0.06577427685260773, 0.030404437333345413, -0.1330784559249878, -0.047490790486335754, -0.0002715167647693306, 0.06956885010004044, -0.0981195792555809, -0.02148638665676117, 0.06170440465211868, -0.11161839962005615, 0.0884210467338562, 0.07083312422037125, 0.02005591429769993, 0.06257379055023193, -0.11040586978197098, 0.009208143688738346, 0.0477973148226738, 0.0010505871614441276, 0.023405956104397774, -0.1711723655462265, -0.0035884424578398466, -0.004025849513709545, 0.04840497300028801, -0.0037269529420882463, 0.028500298038125038, -0.13301301002502441, -0.07130225002765656, -0.024840982630848885, -0.05932305380702019, -0.052967846393585205, 0.040630582720041275, 0.05545862764120102, 0.06687422841787338, 0.15998579561710358, -0.0829639881849289, 0.042585909366607666, -0.21929967403411865, 0.016270728781819344, -0.04847302660346031, -0.07302287220954895, -0.06342391669750214, -0.030033141374588013, 0.0856008380651474, -0.06486867368221283, 0.08810421824455261, -0.061191998422145844, 0.05921773239970207, 0.03249958157539368, -0.11679625511169434, 0.03492012247443199, 0.03797632455825806, 0.2647722363471985, 0.052787795662879944, -0.014916357584297657, 0.07748890668153763, 0.0020198284182697535, 0.047175146639347076, 0.16959668695926666, 0.14780829846858978, 0.17870484292507172, 0.02480897307395935, 0.09690064936876297, 0.06818673759698868, -0.11109158396720886, -0.10169978439807892, 0.09472813457250595, -0.01323322020471096, 0.12197843939065933, -0.003444995731115341, 0.24179279804229736, 0.10775446146726608, -0.2036779224872589, 0.058625489473342896, -0.03863034024834633, -0.08972300589084625, -0.09595318138599396, -0.027152735739946365, -0.0703408345580101, -0.18984375894069672, 0.02177145518362522, -0.12390671670436859, 0.06359978765249252, 0.06148873642086983, 0.03589027002453804, 0.013100014999508858, 0.13031497597694397, 0.037107400596141815, -0.013663274236023426, 0.12015435844659805, -0.0015578477177768946, -0.014140911400318146, -0.06004898622632027, -0.10254908353090286, 0.04876482114195824, -0.030830899253487587, 0.05610576644539833, -0.05483561009168625, -0.11260121315717697, 0.05526963621377945, 0.002577410778030753, -0.1041756272315979, 0.016759352758526802, -0.0054234834387898445, 0.07928597927093506, 0.09193595498800278, 0.03873264789581299, 0.0048627546057105064, -0.01563306339085102, 0.2580336034297943, -0.10385177284479141, -0.06515523791313171, -0.14137886464595795, 0.25403305888175964, 0.02865537628531456, -0.02619745209813118, 0.021571356803178787, -0.07168131321668625, -0.008714161813259125, 0.15781210362911224, 0.11404966562986374, -0.013429660350084305, -0.025903083384037018, -0.0022932037245482206, -0.016892530024051666, -0.0581127367913723, 0.08919288218021393, 0.13225902616977692, 0.043611206114292145, -0.06815124303102493, -0.04760069400072098, -0.061222851276397705, -0.04544898867607117, -0.01562664471566677, 0.06962182372808456, 0.019466735422611237, -0.025623762980103493, -0.025243427604436874, 0.12688902020454407, -0.06399281322956085, -0.10489575564861298, 0.009534171782433987, -0.17033791542053223, -0.18792134523391724, -0.04243918135762215, 0.034101080149412155, 0.0379495695233345, 0.04671601206064224, -0.020714662969112396, -0.014258530922234058, 0.11134493350982666, 0.0046556913293898106, -0.03854575380682945, -0.13683342933654785, 0.10288061946630478, -0.0974394828081131, 0.18676921725273132, -0.04112088307738304, 0.03960246592760086, 0.11430252343416214, 0.07869084179401398, -0.06764878332614899, 0.06385751813650131, 0.06524940580129623, -0.12149403244256973, 0.04727394878864288, 0.19973085820674896, -0.038885120302438736, 0.1367892175912857, 0.03317240998148918, -0.13021613657474518, 0.01958843693137169, -0.09289409220218658, -0.046350494027137756, -0.07179050892591476, -0.01483982801437378, -0.055276259779930115, 0.12390951812267303, 0.22370553016662598, -0.07603158056735992, -0.01654791459441185, -0.06694015115499496, 0.018426677212119102, 0.053473979234695435, 0.11544861644506454, -0.05098946392536163, -0.2898627817630768, 0.01444836612790823, 0.01132238656282425, -0.006995795760303736, -0.24904748797416687, -0.08302478492259979, 0.03644193336367607, -0.07617080211639404, -0.031586769968271255, 0.11826539039611816, 0.06327953189611435, 0.04492071270942688, -0.052991725504398346, -0.0787549540400505, -0.039856232702732086, 0.19568489491939545, -0.1688203364610672, -0.06793981790542603 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - EL dataset. It achieves the following results on the evaluation set: - Loss: 0.3218 - Wer: 0.3095 ## Training and evaluation data Evaluation is conducted in Notebook, you can see within the repo "notebook_evaluation_wav2vec2_el.ipynb" Test WER without LM wer = 31.1294 % cer = 7.9509 % Test WER using LM wer = 20.7340 % cer = 6.0466 % How to use eval.py ``` huggingface-cli login #login to huggingface for getting auth token to access the common voice v8 #running with LM !python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-el --dataset mozilla-foundation/common_voice_8_0 --config el --split test # running without LM !python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-el --dataset mozilla-foundation/common_voice_8_0 --config el --split test --greedy ``` ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 400 - num_epochs: 80.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 6.3683 | 8.77 | 500 | 3.1280 | 1.0 | | 1.9915 | 17.54 | 1000 | 0.6600 | 0.6444 | | 0.6565 | 26.32 | 1500 | 0.4208 | 0.4486 | | 0.4484 | 35.09 | 2000 | 0.3885 | 0.4006 | | 0.3573 | 43.86 | 2500 | 0.3548 | 0.3626 | | 0.3063 | 52.63 | 3000 | 0.3375 | 0.3430 | | 0.2751 | 61.4 | 3500 | 0.3359 | 0.3241 | | 0.2511 | 70.18 | 4000 | 0.3222 | 0.3108 | | 0.2361 | 78.95 | 4500 | 0.3205 | 0.3084 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
{"language": ["el"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard", "mozilla-foundation/common_voice_8_0", "robust-speech-event"], "datasets": ["mozilla-foundation/common_voice_8_0"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-el", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 8", "type": "mozilla-foundation/common_voice_8_0", "args": "el"}, "metrics": [{"type": "wer", "value": 20.9, "name": "Test WER using LM"}, {"type": "cer", "value": 6.0466, "name": "Test CER using LM"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xls-r-300m-el
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard", "mozilla-foundation/common_voice_8_0", "el", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #robust-speech-event #hf-asr-leaderboard #mozilla-foundation/common_voice_8_0 #el #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - EL dataset. It achieves the following results on the evaluation set: * Loss: 0.3218 * Wer: 0.3095 Training and evaluation data ---------------------------- Evaluation is conducted in Notebook, you can see within the repo "notebook\_evaluation\_wav2vec2\_el.ipynb" Test WER without LM wer = 31.1294 % cer = 7.9509 % Test WER using LM wer = 20.7340 % cer = 6.0466 % How to use URL Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-05 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 64 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 400 * num\_epochs: 80.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.1+cu102 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 80.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #robust-speech-event #hf-asr-leaderboard #mozilla-foundation/common_voice_8_0 #el #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 80.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 111, 160, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #robust-speech-event #hf-asr-leaderboard #mozilla-foundation/common_voice_8_0 #el #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 80.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.12867338955402374, 0.11383134126663208, -0.0053503867238759995, 0.05088794603943825, 0.12288543581962585, 0.015738720074295998, 0.09791316837072372, 0.14653952419757843, -0.08275110274553299, 0.08956418931484222, 0.07430341839790344, 0.08870711922645569, 0.0839126706123352, 0.08867949992418289, -0.005079356022179127, -0.29594382643699646, 0.01487681083381176, -0.021777601912617683, -0.11331191658973694, 0.10684893280267715, 0.09687558561563492, -0.09079272300004959, 0.033751241862773895, 0.032316554337739944, -0.08320451527833939, -0.00004341223757364787, -0.046322938054800034, -0.043740689754486084, 0.0831783339381218, 0.04275323823094368, 0.051752883940935135, 0.0328977108001709, 0.08338884264230728, -0.2457142174243927, 0.007203138433396816, 0.07081882655620575, 0.035581186413764954, 0.06004286929965019, 0.12765127420425415, -0.01564108207821846, 0.11811410635709763, -0.054553501307964325, 0.03782590478658676, 0.06648997962474823, -0.1006794348359108, -0.24773283302783966, -0.08315546065568924, 0.054155658930540085, 0.12838366627693176, 0.09209147840738297, -0.028533393517136574, 0.027561070397496223, -0.0960969626903534, 0.09390641748905182, 0.22761303186416626, -0.22013859450817108, -0.06795952469110489, -0.015621154569089413, 0.040738679468631744, 0.01831899583339691, -0.11089133471250534, -0.02851700223982334, 0.010351745411753654, 0.007809234783053398, 0.09393023699522018, 0.008635211735963821, 0.006277315318584442, -0.002750108251348138, -0.14331741631031036, -0.06337162107229233, 0.11696134507656097, 0.07301082462072372, -0.008194411173462868, -0.1008978858590126, -0.02759885974228382, -0.1970807909965515, -0.04570885747671127, 0.04637191817164421, 0.020494705066084862, -0.033668749034404755, -0.033790163695812225, 0.032184820622205734, -0.044817570596933365, -0.07422783225774765, 0.06743381917476654, 0.12111814320087433, 0.04722755774855614, -0.03200451657176018, -0.0062934355810284615, 0.12072055041790009, 0.0856204703450203, -0.172125443816185, -0.026886561885476112, 0.03479308634996414, -0.1148577630519867, -0.0033283173106610775, -0.007423547562211752, 0.030547073110938072, 0.039096999913454056, 0.13165590167045593, -0.010180533863604069, 0.08040273189544678, 0.027756864205002785, 0.023725656792521477, -0.07868248969316483, 0.17466585338115692, -0.0852934792637825, -0.0733453705906868, -0.04699142277240753, 0.13435979187488556, -0.013087984174489975, -0.01018697489053011, -0.06443947553634644, 0.028203466907143593, 0.09498538076877594, 0.06500371545553207, 0.0044393413700163364, 0.0354018434882164, -0.0585910864174366, -0.036087024956941605, 0.012224659323692322, -0.1250690370798111, 0.03825050964951515, 0.06720266491174698, -0.09989655762910843, 0.005318529438227415, -0.01660631038248539, 0.005455749575048685, -0.05068035051226616, 0.09168283641338348, -0.06058172881603241, -0.00027930992655456066, -0.09173861145973206, -0.08809595555067062, 0.036631759256124496, -0.0393943265080452, -0.002417383948341012, -0.055814314633607864, -0.10934368520975113, -0.06672759354114532, 0.05022575333714485, -0.06261178851127625, -0.07258202880620956, -0.06399544328451157, -0.09639877080917358, 0.053153760731220245, -0.02282237634062767, 0.18171316385269165, -0.0621771439909935, 0.08678647130727768, 0.02336714044213295, 0.04280976578593254, 0.12602683901786804, 0.0736391469836235, -0.02600107714533806, 0.06016532704234123, -0.11769741028547287, 0.10279674082994461, -0.11530756950378418, 0.06862927228212357, -0.11846552044153214, -0.10392218828201294, -0.007888101041316986, -0.0024756677448749542, 0.09156505018472672, 0.1296161264181137, -0.13716992735862732, -0.09234437346458435, 0.1602773815393448, -0.04656675457954407, -0.0775429904460907, 0.11593225598335266, -0.006359192077070475, -0.05259525403380394, 0.016970638185739517, 0.15923209488391876, 0.15356411039829254, -0.08519066870212555, -0.003303781384602189, -0.0433596596121788, 0.10209514200687408, 0.03558995574712753, 0.09082812070846558, -0.03007296659052372, 0.04196573793888092, 0.005986359901726246, -0.03800342604517937, 0.0686129629611969, -0.08428186923265457, -0.08453421294689178, -0.013705248944461346, -0.07721088081598282, 0.008215966634452343, 0.06204181909561157, 0.009539769031107426, -0.0604158416390419, -0.14005275070667267, -0.008627643808722496, 0.11846476793289185, -0.0917617455124855, 0.010182274505496025, -0.0832851380109787, 0.05779635161161423, -0.004545058123767376, 0.0024615998845547438, -0.15517030656337738, -0.03989392891526222, 0.03220267966389656, -0.07801079750061035, -0.0031352003570646048, -0.025282209739089012, 0.06681200116872787, 0.04456236585974693, -0.026806648820638657, -0.08088614046573639, -0.053896039724349976, -0.01846376620233059, -0.047344718128442764, -0.22531628608703613, -0.10251200944185257, -0.015040951780974865, 0.1813451200723648, -0.1956847757101059, 0.015445506200194359, 0.044329676777124405, 0.12840934097766876, 0.01736612431704998, -0.044874273240566254, 0.02350418083369732, 0.055504534393548965, -0.02165387198328972, -0.08108314871788025, 0.031716447323560715, 0.010640447027981281, -0.0850093811750412, 0.020643804222345352, -0.11466173082590103, 0.0914938822388649, 0.07651473581790924, 0.016352446749806404, -0.05691509693861008, -0.030018718913197517, -0.07017266750335693, -0.06344789266586304, -0.014263821765780449, -0.011748342774808407, 0.16085916757583618, 0.02015732228755951, 0.1154220849275589, -0.07088322192430496, -0.05120635777711868, 0.03207645192742348, 0.022762728855013847, 0.0008532080100849271, 0.15775789320468903, 0.07432406395673752, -0.0323978029191494, 0.09650208055973053, -0.005667462479323149, -0.06508121639490128, 0.1712339073419571, -0.07611432671546936, -0.09644526243209839, -0.0272330604493618, 0.02535473369061947, 0.029234856367111206, 0.10153566300868988, -0.1934785395860672, -0.017821861431002617, 0.03233770653605461, 0.016191333532333374, 0.03177325800061226, -0.18500198423862457, -0.002635337645187974, 0.0326622873544693, -0.08478885143995285, -0.00834660790860653, 0.007407051976770163, -0.004355499055236578, 0.08224640041589737, 0.0012606828240677714, -0.09716254472732544, -0.030551377683877945, -0.043154872953891754, -0.08735757321119308, 0.17190612852573395, -0.06987795233726501, -0.14292800426483154, -0.13080604374408722, -0.027929581701755524, -0.03639427572488785, -0.024085383862257004, 0.03613904491066933, -0.08843067288398743, -0.03417004644870758, -0.06337625533342361, 0.020193587988615036, -0.033886708319187164, 0.011644510552287102, 0.0278990026563406, 0.006353459320962429, 0.062477584928274155, -0.10548042505979538, 0.015309600159525871, -0.01510898768901825, -0.04036087170243263, 0.0036906120367348194, 0.02732527256011963, 0.09804581105709076, 0.18744449317455292, 0.0667809247970581, 0.04671507328748703, -0.01531843189150095, 0.18233807384967804, -0.13355936110019684, 0.006218167021870613, 0.11540953814983368, 0.018693605437874794, 0.053089771419763565, 0.14197491109371185, 0.0488801933825016, -0.06273863464593887, 0.004625820089131594, 0.0501362606883049, -0.022838104516267776, -0.2345174103975296, -0.02914191037416458, -0.07565850019454956, -0.021240662783384323, 0.10226312279701233, 0.034484367817640305, 0.022195201367139816, 0.019778015092015266, -0.03162969648838043, -0.019664226099848747, 0.049172669649124146, 0.043986327946186066, 0.03658726438879967, 0.04686226695775986, 0.11342037469148636, -0.014959320425987244, -0.03349384665489197, 0.014380230568349361, -0.005166694056242704, 0.24162626266479492, -0.015391891822218895, 0.18373116850852966, 0.05587244778871536, 0.1401219367980957, 0.0036911533679813147, 0.050663549453020096, 0.00008556653483537957, -0.0048118592239916325, 0.03907302767038345, -0.05617921054363251, -0.014001521281898022, 0.02467544563114643, 0.10256791114807129, 0.03125644475221634, -0.11082731932401657, 0.005996949039399624, 0.019640760496258736, 0.37537720799446106, 0.07679443061351776, -0.2950237989425659, -0.07008690387010574, 0.0022315436508506536, -0.06442568451166153, -0.04115791991353035, 0.028148239478468895, 0.09144624322652817, -0.08661390095949173, 0.06830860674381256, -0.059413302689790726, 0.09592920541763306, -0.08335861563682556, 0.0005433778860606253, 0.10564043372869492, 0.09497656673192978, 0.012374451383948326, 0.065216064453125, -0.24247664213180542, 0.2550165057182312, -0.015729032456874847, 0.07445383071899414, -0.04079807922244072, 0.05301237478852272, 0.04633563384413719, 0.0020572093781083822, 0.06844625622034073, -0.0018014096422120929, -0.07031286507844925, -0.17130813002586365, -0.10449875146150589, -0.0027217993047088385, 0.10689704865217209, -0.06344272196292877, 0.12283313274383545, -0.039174243807792664, -0.0493384450674057, 0.029675301164388657, -0.06599000841379166, -0.11721809953451157, -0.0979287400841713, 0.06136012077331543, 0.01762733981013298, 0.0850982740521431, -0.09295249730348587, -0.09767864644527435, -0.04874391853809357, 0.15522542595863342, -0.1267286092042923, -0.037466105073690414, -0.13573652505874634, 0.0423317514359951, 0.14956796169281006, -0.06492240726947784, 0.02921409346163273, 0.015016858465969563, 0.144276961684227, 0.038764551281929016, -0.024978265166282654, 0.09601476788520813, -0.08579450100660324, -0.19815607368946075, -0.04357534274458885, 0.17568376660346985, 0.029037432745099068, 0.0589718334376812, -0.025736885145306587, 0.010039606131613255, -0.009516718797385693, -0.0766097754240036, 0.06828390806913376, 0.04613013193011284, -0.010657751932740211, 0.05280708521604538, -0.03594909608364105, 0.012943089008331299, -0.08038843423128128, -0.037614766508340836, 0.09055592864751816, 0.22065208852291107, -0.07562323659658432, 0.029809167608618736, 0.03107224404811859, -0.06373520195484161, -0.15263842046260834, -0.007513928692787886, 0.1355741173028946, 0.034259017556905746, -0.03040437586605549, -0.21251925826072693, 0.008856925182044506, 0.07022617012262344, -0.024592390283942223, 0.11176528036594391, -0.37373656034469604, -0.12121696770191193, 0.08648114651441574, 0.052984002977609634, -0.03663695976138115, -0.16867941617965698, -0.06879093497991562, -0.011501702480018139, -0.10617558658123016, 0.03738750144839287, -0.011121012270450592, 0.12956100702285767, -0.01084769144654274, 0.02425825409591198, 0.014135845005512238, -0.05493535101413727, 0.14097708463668823, 0.032251834869384766, 0.044676318764686584, -0.01227102056145668, 0.019055062904953957, 0.029208512976765633, -0.07808535546064377, 0.02124127186834812, -0.06095903739333153, 0.021776922047138214, -0.17262865602970123, -0.025609727948904037, -0.10435807704925537, 0.022572990506887436, -0.05030081793665886, -0.012296760454773903, -0.009555994533002377, 0.04605729877948761, 0.08855937421321869, 0.028318408876657486, 0.10538507252931595, -0.055805496871471405, 0.1118822768330574, 0.13993534445762634, 0.08516470342874527, -0.008919987827539444, -0.14297527074813843, -0.03370528295636177, 0.023792492225766182, 0.04780242592096329, -0.10034599155187607, 0.03983136638998985, 0.14198271930217743, 0.04931741952896118, 0.13837362825870514, 0.0570547878742218, -0.08737041801214218, -0.00028815760742872953, 0.0578092560172081, -0.07266409695148468, -0.11915505677461624, -0.0372493676841259, 0.007349592633545399, -0.11626897007226944, -0.006662517320364714, 0.1081196516752243, -0.036434680223464966, 0.006281714420765638, 0.019837846979498863, 0.048513323068618774, -0.02863616868853569, 0.2227102369070053, 0.030450306832790375, 0.10720261931419373, -0.09684654325246811, 0.06875546276569366, 0.05337727442383766, -0.08728805929422379, 0.016238627955317497, 0.11113984882831573, -0.04815167933702469, -0.02929379977285862, -0.014678611420094967, 0.0821460634469986, 0.02562738209962845, -0.05930363014340401, -0.12396464496850967, -0.15293249487876892, 0.10384291410446167, 0.05439385771751404, 0.01972084864974022, 0.03329094126820564, -0.022529812529683113, 0.027795447036623955, -0.09234458953142166, 0.10479112714529037, 0.09468794614076614, 0.06460777670145035, -0.13183726370334625, 0.09994398802518845, 0.021243786439299583, 0.015329656191170216, 0.006952894385904074, -0.020878730341792107, -0.10083749890327454, 0.032852523028850555, -0.12031379342079163, -0.017658887431025505, -0.04229024425148964, 0.0008803210803307593, 0.0002058610989479348, -0.05250706151127815, -0.06301086395978928, 0.03467384725809097, -0.11183840036392212, -0.04856075346469879, -0.030498605221509933, 0.07061083614826202, -0.0997392013669014, -0.015087990090250969, 0.03111950494349003, -0.13734881579875946, 0.08820928633213043, 0.047201454639434814, 0.015734530985355377, 0.006333462428301573, -0.07154922187328339, -0.009316783398389816, 0.024567270651459694, 0.005963390693068504, 0.039919983595609665, -0.17458786070346832, 0.0032533607445657253, -0.03783360868692398, -0.0018462548032402992, -0.022713446989655495, -0.01622348092496395, -0.12798747420310974, -0.00020590348867699504, -0.019279539585113525, -0.045225340873003006, -0.04681774228811264, 0.07340894639492035, 0.08078210055828094, 0.015245991759002209, 0.15462367236614227, -0.05999713018536568, 0.06574926525354385, -0.23107975721359253, 0.005986938718706369, -0.001493159681558609, -0.06344661116600037, -0.021119531244039536, -0.020852675661444664, 0.09941810369491577, -0.06737863272428513, 0.0797543004155159, -0.005121510475873947, 0.045326974242925644, 0.021396618336439133, -0.0930146723985672, 0.036503542214632034, 0.06303849071264267, 0.13219882547855377, 0.049671802669763565, -0.009575829841196537, 0.0976337119936943, -0.019408678635954857, 0.05784986540675163, 0.12068939208984375, 0.1620965451002121, 0.11959738284349442, 0.07827509939670563, 0.08366914838552475, 0.0940370038151741, -0.14214767515659332, -0.14955557882785797, 0.12618795037269592, -0.06352007389068604, 0.15788957476615906, -0.031806088984012604, 0.18163473904132843, 0.12259985506534576, -0.18447723984718323, 0.07753097265958786, -0.02812992036342621, -0.08495109528303146, -0.09207294136285782, -0.09522733092308044, -0.06404650956392288, -0.16518935561180115, 0.017751121893525124, -0.09699945896863937, 0.06394266337156296, 0.06435201317071915, 0.039571937173604965, 0.02994244359433651, 0.10356416553258896, 0.08382459729909897, 0.0015874753007665277, 0.11684715747833252, 0.004057483281940222, -0.02166907675564289, -0.05618501082062721, -0.0921868309378624, 0.042658500373363495, -0.03137482702732086, 0.05661603435873985, -0.03664792701601982, -0.10633369535207748, 0.06641682982444763, 0.027475150302052498, -0.09537937492132187, 0.03106255829334259, -0.03111848421394825, 0.055824823677539825, 0.07320015132427216, 0.035824891179800034, -0.00288027198985219, -0.013919651508331299, 0.21372811496257782, -0.080815389752388, -0.04498577117919922, -0.11640387773513794, 0.18137173354625702, 0.009590656496584415, -0.00621765898540616, 0.028406042605638504, -0.06582364439964294, -0.029957713559269905, 0.18151023983955383, 0.13693667948246002, -0.02146591991186142, -0.02484273724257946, 0.007957665249705315, -0.008289196528494358, -0.02465130016207695, 0.08367939293384552, 0.12148881703615189, 0.0780956894159317, -0.03945261612534523, -0.012230895459651947, -0.024940911680459976, -0.07213981449604034, -0.030926955863833427, 0.09904871135950089, 0.02241968736052513, 0.0022161416709423065, -0.03145385533571243, 0.10119365155696869, -0.0895361378788948, -0.132883682847023, 0.040946051478385925, -0.15659627318382263, -0.1923368126153946, -0.04985702037811279, 0.06162592023611069, 0.03942263126373291, 0.07437950372695923, -0.0036550743971019983, -0.05507887527346611, 0.1119796633720398, 0.004950962960720062, -0.04077547416090965, -0.12240029126405716, 0.06578224152326584, -0.12035588920116425, 0.19710145890712738, -0.05015210807323456, 0.011316078715026379, 0.12161235511302948, 0.04513930529356003, -0.10614203661680222, 0.01578974723815918, 0.08104290068149567, -0.13840477168560028, 0.03893238678574562, 0.19203199446201324, -0.033158764243125916, 0.11653369665145874, 0.02047180011868477, -0.0901266410946846, 0.011162803508341312, -0.066097691655159, -0.019506342709064484, -0.06585800647735596, -0.013249307870864868, -0.027456607669591904, 0.1264367252588272, 0.21954399347305298, -0.06654802709817886, -0.016939379274845123, -0.05269023776054382, 0.005743459798395634, 0.015847615897655487, 0.12223056703805923, -0.04220162704586983, -0.2606096863746643, 0.024030618369579315, -0.010194594040513039, 0.028808744624257088, -0.19132794439792633, -0.08325077593326569, 0.034352101385593414, -0.07143034040927887, -0.05697128176689148, 0.11277665942907333, 0.05704004690051079, 0.05426700413227081, -0.051219262182712555, -0.092111736536026, -0.03163395822048187, 0.18016649782657623, -0.18907615542411804, -0.06504163891077042 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-ia This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.1452 - Wer: 0.1253 ## Training Procedure Training is conducted in Google Colab, the training notebook provided in the repo ## Training and evaluation data Language Model Created from texts from processed sentence in train + validation split of dataset (common voice 8.0 for Interlingua) Evaluation is conducted in Notebook, you can see within the repo "notebook_evaluation_wav2vec2_ia.ipynb" Test WER without LM wer = 20.1776 % cer = 4.7205 % Test WER using wer = 8.6074 % cer = 2.4147 % evaluation using eval.py ``` huggingface-cli login #login to huggingface for getting auth token to access the common voice v8 #running with LM python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-ia --dataset mozilla-foundation/common_voice_8_0 --config ia --split test # running without LM python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-ia --dataset mozilla-foundation/common_voice_8_0 --config ia --split test --greedy ``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 16 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 400 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 7.432 | 1.87 | 400 | 2.9636 | 1.0 | | 2.6922 | 3.74 | 800 | 2.2111 | 0.9977 | | 1.2581 | 5.61 | 1200 | 0.4864 | 0.4028 | | 0.6232 | 7.48 | 1600 | 0.2807 | 0.2413 | | 0.4479 | 9.35 | 2000 | 0.2219 | 0.1885 | | 0.3654 | 11.21 | 2400 | 0.1886 | 0.1606 | | 0.323 | 13.08 | 2800 | 0.1716 | 0.1444 | | 0.2935 | 14.95 | 3200 | 0.1687 | 0.1443 | | 0.2707 | 16.82 | 3600 | 0.1632 | 0.1382 | | 0.2559 | 18.69 | 4000 | 0.1507 | 0.1337 | | 0.2433 | 20.56 | 4400 | 0.1572 | 0.1358 | | 0.2338 | 22.43 | 4800 | 0.1489 | 0.1305 | | 0.2258 | 24.3 | 5200 | 0.1485 | 0.1278 | | 0.2218 | 26.17 | 5600 | 0.1470 | 0.1272 | | 0.2169 | 28.04 | 6000 | 0.1470 | 0.1270 | | 0.2117 | 29.91 | 6400 | 0.1452 | 0.1253 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"language": ["ia"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "robust-speech-event"], "datasets": ["mozilla-foundation/common_voice_8_0"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-ia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 8", "type": "mozilla-foundation/common_voice_8_0", "args": "ia"}, "metrics": [{"type": "wer", "value": 8.6074, "name": "Test WER using LM"}, {"type": "cer", "value": 2.4147, "name": "Test CER using LM"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xls-r-300m-ia
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "ia", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ia" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #ia #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us
wav2vec2-large-xls-r-300m-ia ============================ This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common\_voice dataset. It achieves the following results on the evaluation set: * Loss: 0.1452 * Wer: 0.1253 Training Procedure ------------------ Training is conducted in Google Colab, the training notebook provided in the repo Training and evaluation data ---------------------------- Language Model Created from texts from processed sentence in train + validation split of dataset (common voice 8.0 for Interlingua) Evaluation is conducted in Notebook, you can see within the repo "notebook\_evaluation\_wav2vec2\_ia.ipynb" Test WER without LM wer = 20.1776 % cer = 4.7205 % Test WER using wer = 8.6074 % cer = 2.4147 % evaluation using URL ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 3e-05 * train\_batch\_size: 16 * eval\_batch\_size: 4 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 400 * num\_epochs: 30 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #ia #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 111, 159, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #ia #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 4\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.13417483866214752, 0.10810285061597824, -0.00513931829482317, 0.046122509986162186, 0.11306290328502655, 0.009446452371776104, 0.10716062784194946, 0.1428627222776413, -0.08867528289556503, 0.10069473087787628, 0.08064551651477814, 0.08290591835975647, 0.07936400920152664, 0.0947943851351738, -0.007655202876776457, -0.2972838878631592, 0.023036466911435127, -0.027581479400396347, -0.1274603307247162, 0.10589376091957092, 0.09989114105701447, -0.08441708236932755, 0.03644651547074318, 0.0355822928249836, -0.10054205358028412, -0.0017242792528122663, -0.041734565049409866, -0.04332669824361801, 0.08356547355651855, 0.056743208318948746, 0.05190768092870712, 0.03533610329031944, 0.08543580770492554, -0.24601584672927856, 0.00898931547999382, 0.068290576338768, 0.03733684867620468, 0.060135360807180405, 0.11753213405609131, -0.018123043701052666, 0.10705737769603729, -0.04273213446140289, 0.03624819591641426, 0.07165109366178513, -0.10530055314302444, -0.23831292986869812, -0.08504655957221985, 0.0508384108543396, 0.124685138463974, 0.09303440153598785, -0.03249470144510269, 0.03065350651741028, -0.0833512544631958, 0.09177146852016449, 0.21868425607681274, -0.227402463555336, -0.07727105170488358, -0.01327072735875845, 0.04456324875354767, 0.016091004014015198, -0.1147652193903923, -0.01699560321867466, 0.02072657085955143, 0.006343476474285126, 0.07258763909339905, 0.00929793156683445, 0.013718805275857449, -0.007344232406467199, -0.1400424987077713, -0.0502568855881691, 0.13926301896572113, 0.07960743457078934, -0.01784924603998661, -0.09962253272533417, -0.02074209228157997, -0.1889057457447052, -0.04307915270328522, 0.03754279389977455, 0.019306978210806847, -0.026577992364764214, -0.05260762572288513, 0.02905392274260521, -0.04714926704764366, -0.08078708499670029, 0.05751322954893112, 0.13800625503063202, 0.039931416511535645, -0.029122892767190933, -0.005867087747901678, 0.1112857237458229, 0.082317054271698, -0.17212866246700287, -0.031287893652915955, 0.0373677983880043, -0.11652420461177826, -0.007769802585244179, -0.01313536986708641, 0.041028399020433426, 0.03592349961400032, 0.14311352372169495, 0.0012348354794085026, 0.08484896272420883, 0.022478418424725533, 0.019138222560286522, -0.07478597015142441, 0.17965714633464813, -0.09152239561080933, -0.07669180631637573, -0.04590218514204025, 0.1286284625530243, -0.011909149587154388, -0.012000949122011662, -0.062434110790491104, 0.033230677247047424, 0.08969210088253021, 0.06354548782110214, 0.003365045180544257, 0.038402874022722244, -0.05473362281918526, -0.029041819274425507, 0.014348863624036312, -0.12358114868402481, 0.035741910338401794, 0.06542357057332993, -0.09597423672676086, 0.020922843366861343, -0.018303142860531807, 0.008920738473534584, -0.03909527510404587, 0.07448296248912811, -0.06502238661050797, -0.001514712581411004, -0.08395811915397644, -0.08664308488368988, 0.03692513704299927, -0.02105230838060379, -0.00880276970565319, -0.055778127163648605, -0.10032283514738083, -0.06951975077390671, 0.052867379039525986, -0.06853361427783966, -0.07187512516975403, -0.07148317247629166, -0.10064370185136795, 0.06047941371798515, -0.016120409592986107, 0.16528737545013428, -0.05477225407958031, 0.08105386793613434, 0.034225236624479294, 0.03950494900345802, 0.11212945729494095, 0.06816200911998749, -0.02192896045744419, 0.06390396505594254, -0.1188931092619896, 0.10539431124925613, -0.11315729469060898, 0.06984962522983551, -0.12351425737142563, -0.09945137053728104, -0.0167945958673954, 0.005250450689345598, 0.09477659314870834, 0.13383232057094574, -0.14870195090770721, -0.09308706969022751, 0.16024723649024963, -0.05173836275935173, -0.08984233438968658, 0.12069588899612427, -0.00525411544367671, -0.03547309339046478, 0.021552864462137222, 0.1588047593832016, 0.1515524536371231, -0.09283802658319473, -0.010415446944534779, -0.046722423285245895, 0.108243927359581, 0.04356595128774643, 0.09335745871067047, -0.03990143910050392, 0.04074985533952713, 0.004472160246223211, -0.03315628692507744, 0.06036578118801117, -0.08580714464187622, -0.08473670482635498, -0.009332140907645226, -0.07585366070270538, 0.0014877045759931207, 0.05780931934714317, 0.013930150307714939, -0.06493522971868515, -0.14144499599933624, -0.0016171898460015655, 0.1208113431930542, -0.09987540543079376, 0.01718963496387005, -0.09121524542570114, 0.05638306587934494, -0.011967247352004051, 0.0017638426506891847, -0.15191295742988586, -0.041335321962833405, 0.03224560245871544, -0.06749112159013748, 0.008416533470153809, -0.030340421944856644, 0.07057158648967743, 0.04325081408023834, -0.024895960465073586, -0.07233782857656479, -0.044837698340415955, -0.01945827715098858, -0.043228067457675934, -0.22907978296279907, -0.09638805687427521, -0.018135156482458115, 0.1979474574327469, -0.19963866472244263, 0.012163824401795864, 0.04798116534948349, 0.12545812129974365, 0.01958632469177246, -0.05195820704102516, 0.017874468117952347, 0.05013683810830116, -0.027834169566631317, -0.07984139770269394, 0.023999180644750595, 0.014463838189840317, -0.09083051979541779, 0.023417016491293907, -0.12544944882392883, 0.09176953136920929, 0.08160151541233063, 0.01343547459691763, -0.06484632939100266, -0.03173152729868889, -0.06592556089162827, -0.06049826741218567, -0.01015059370547533, -0.020849157124757767, 0.15818770229816437, 0.02011134661734104, 0.1069214716553688, -0.06992324441671371, -0.04961497336626053, 0.03223537653684616, 0.018551155924797058, -0.010387497022747993, 0.14588342607021332, 0.07167568802833557, -0.036148592829704285, 0.09302634000778198, -0.0052072396501898766, -0.06364747136831284, 0.1659899204969406, -0.07610421627759933, -0.09153788536787033, -0.038509611040353775, 0.026172498241066933, 0.03323334828019142, 0.1081613302230835, -0.1837247908115387, -0.023233504965901375, 0.023239746689796448, 0.023652326315641403, 0.03251168131828308, -0.17984363436698914, 0.0012386215385049582, 0.032741524279117584, -0.09111639857292175, -0.012126905843615532, 0.007033160421997309, -0.017443805932998657, 0.07877891510725021, 0.007077267859131098, -0.0916064903140068, -0.03311701491475105, -0.04164749011397362, -0.0893130823969841, 0.17046557366847992, -0.0740690752863884, -0.13017095625400543, -0.14204664528369904, -0.026869596913456917, -0.03904244676232338, -0.020368456840515137, 0.028614025563001633, -0.0912618339061737, -0.03421730548143387, -0.06432089954614639, 0.026024101302027702, -0.039982568472623825, 0.014141482301056385, 0.038205165416002274, 0.0026273697149008512, 0.06639456003904343, -0.10665807127952576, 0.019317368045449257, -0.013145270757377148, -0.04044731333851814, -0.010259336791932583, 0.013192743062973022, 0.10278891026973724, 0.18518728017807007, 0.07858415693044662, 0.052764344960451126, -0.018532130867242813, 0.18733884394168854, -0.1370050311088562, 0.007516895886510611, 0.11239215731620789, 0.019433053210377693, 0.04362322390079498, 0.14827607572078705, 0.049693964421749115, -0.07683471590280533, 0.009290860965847969, 0.04998236894607544, -0.02242572046816349, -0.23070357739925385, -0.02735031023621559, -0.07678290456533432, -0.020025208592414856, 0.09098942577838898, 0.028889872133731842, 0.0250944122672081, 0.018844883888959885, -0.028532207012176514, -0.01705886423587799, 0.0522325374186039, 0.046301111578941345, 0.04659285396337509, 0.05092731490731239, 0.11710771918296814, -0.018630001693964005, -0.03403818979859352, 0.016187725588679314, -0.006133730057626963, 0.2424471378326416, -0.003096548141911626, 0.20084930956363678, 0.05476627126336098, 0.13229981064796448, 0.0020934282802045345, 0.050299473106861115, 0.006439555902034044, -0.009425508789718151, 0.04129103943705559, -0.056811198592185974, -0.011121622286736965, 0.022838382050395012, 0.11230283975601196, 0.028963536024093628, -0.1101992204785347, 0.012455991469323635, 0.023493491113185883, 0.3667847812175751, 0.08879832923412323, -0.2939850687980652, -0.07848111540079117, 0.004110739566385746, -0.06328509002923965, -0.03665771707892418, 0.03256005793809891, 0.09105033427476883, -0.08155019581317902, 0.07021145522594452, -0.05749916285276413, 0.09491259604692459, -0.07688692212104797, -0.0012708879075944424, 0.10217566043138504, 0.09596932679414749, 0.0070716445334255695, 0.06011529266834259, -0.22659429907798767, 0.26271918416023254, -0.013053089380264282, 0.08115595579147339, -0.03350769355893135, 0.05155550688505173, 0.045233387500047684, 0.001531789661385119, 0.06528231501579285, -0.006932368036359549, -0.08215981721878052, -0.16410858929157257, -0.10743951797485352, 0.00322407647036016, 0.11907205730676651, -0.06361430138349533, 0.12850436568260193, -0.04437049850821495, -0.046870552003383636, 0.03858134150505066, -0.059428174048662186, -0.12330614030361176, -0.10354004055261612, 0.05877319723367691, 0.01785055734217167, 0.08479912579059601, -0.08979737758636475, -0.09702388942241669, -0.04296857491135597, 0.16321100294589996, -0.11913634836673737, -0.024430690333247185, -0.13584785163402557, 0.04271179065108299, 0.15711760520935059, -0.06700785458087921, 0.02126440964639187, 0.013126029632985592, 0.14808940887451172, 0.043526019901037216, -0.02411622926592827, 0.09851910173892975, -0.085975781083107, -0.1903812140226364, -0.04323836788535118, 0.1795918196439743, 0.028249215334653854, 0.05844303220510483, -0.02091263234615326, 0.011762233451008797, -0.003619360039010644, -0.08062129467725754, 0.06786590069532394, 0.035938140004873276, -0.02476988174021244, 0.0449368841946125, -0.026940220966935158, 0.01983010210096836, -0.08090613037347794, -0.04015769064426422, 0.08892691135406494, 0.23559001088142395, -0.07391174882650375, 0.01791151985526085, 0.03426908701658249, -0.0609029158949852, -0.1546216607093811, 0.0010635724756866693, 0.13352210819721222, 0.037608202546834946, -0.05200980603694916, -0.21310293674468994, 0.013938580639660358, 0.06911535561084747, -0.025579823181033134, 0.10457872599363327, -0.35710805654525757, -0.130643829703331, 0.08853701502084732, 0.045128822326660156, -0.029932161793112755, -0.172881618142128, -0.07265150547027588, -0.016460338607430458, -0.10711932182312012, 0.03634217381477356, -0.012065582908689976, 0.1269378364086151, 0.0031223930418491364, 0.024433139711618423, 0.01042097993195057, -0.04884444177150726, 0.1485910415649414, 0.03073200210928917, 0.04021092504262924, -0.006716461852192879, 0.014273344539105892, 0.02682540751993656, -0.07507868111133575, 0.021952243521809578, -0.05753084644675255, 0.020210685208439827, -0.15311890840530396, -0.027270615100860596, -0.09943839907646179, 0.018739482387900352, -0.052279163151979446, -0.010824757628142834, -0.014767607674002647, 0.046629078686237335, 0.07728913426399231, 0.028164487332105637, 0.10509002953767776, -0.061790063977241516, 0.11386389285326004, 0.13229674100875854, 0.09544152766466141, -0.00002916793164331466, -0.12751203775405884, -0.020207993686199188, 0.01295793429017067, 0.0440848283469677, -0.09979023039340973, 0.0362606979906559, 0.13999857008457184, 0.0481584332883358, 0.14236418902873993, 0.05791627615690231, -0.08815445005893707, 0.003289520274847746, 0.06005469337105751, -0.06343306601047516, -0.1276523768901825, -0.03518831729888916, 0.005854051560163498, -0.1275710165500641, 0.0031376229599118233, 0.11762520670890808, -0.03288570046424866, 0.010704625397920609, 0.019512033089995384, 0.04232316091656685, -0.03798234462738037, 0.22538451850414276, 0.03225158154964447, 0.10068263113498688, -0.0952877327799797, 0.06698359549045563, 0.045751623809337616, -0.08243729919195175, 0.020303191617131233, 0.10085831582546234, -0.04255512356758118, -0.023172568529844284, -0.018364770337939262, 0.08226697146892548, 0.020466841757297516, -0.056902673095464706, -0.12857937812805176, -0.1558588445186615, 0.09913564473390579, 0.06733071058988571, 0.021170660853385925, 0.03798247128725052, -0.026905715465545654, 0.037890106439590454, -0.09531433880329132, 0.09885148704051971, 0.09842473268508911, 0.06328028440475464, -0.13580673933029175, 0.10823559761047363, 0.013781270943582058, 0.009076523594558239, 0.00615636957809329, -0.02536064013838768, -0.10472099483013153, 0.030503209680318832, -0.12688855826854706, -0.020028267055749893, -0.04826778545975685, -0.0008322561625391245, 0.011344077996909618, -0.05598171055316925, -0.07396900653839111, 0.028555167838931084, -0.11135118454694748, -0.04948217049241066, -0.029687820002436638, 0.06550536304712296, -0.10624346137046814, -0.007998822256922722, 0.038216788321733475, -0.14285612106323242, 0.08858893811702728, 0.055905237793922424, 0.02034277841448784, 0.009790820069611073, -0.06454075127840042, -0.00875572208315134, 0.020109649747610092, 0.0031056187581270933, 0.03504132106900215, -0.18058538436889648, -0.00006388208566932008, -0.03546474874019623, 0.015001807361841202, -0.018522411584854126, -0.008297414518892765, -0.11657088249921799, -0.0030277559999376535, -0.025154996663331985, -0.04253191128373146, -0.04417244344949722, 0.07429108023643494, 0.08200462907552719, 0.021359940990805626, 0.14999203383922577, -0.06431334465742111, 0.05862659588456154, -0.23224860429763794, 0.010145694948732853, -0.0015896218828856945, -0.060250602662563324, -0.014888763427734375, -0.02234717272222042, 0.09908264130353928, -0.06541753560304642, 0.07299715280532837, -0.025127574801445007, 0.04320529103279114, 0.02524738945066929, -0.09621664136648178, 0.033791910856962204, 0.06742477416992188, 0.1372470259666443, 0.05861585587263107, -0.008410618640482426, 0.0941186472773552, -0.01953805424273014, 0.05864555388689041, 0.11791227012872696, 0.16206596791744232, 0.11653696000576019, 0.06132159382104874, 0.09698661416769028, 0.09976516664028168, -0.14375700056552887, -0.12903498113155365, 0.12346134334802628, -0.07554204016923904, 0.14870774745941162, -0.026353124529123306, 0.1747981309890747, 0.12340984493494034, -0.18363647162914276, 0.06731993705034256, -0.03515056148171425, -0.090987429022789, -0.09991622716188431, -0.08626271784305573, -0.06899454444646835, -0.1670362800359726, 0.020590417087078094, -0.10421580076217651, 0.06338006258010864, 0.05264867842197418, 0.047520630061626434, 0.03402961418032646, 0.11155304312705994, 0.06537341326475143, -0.0007829529349692166, 0.11533311754465103, -0.003580802120268345, -0.02435566671192646, -0.05231037735939026, -0.09192754328250885, 0.04604865238070488, -0.032542940229177475, 0.06355220824480057, -0.031785234808921814, -0.10604247450828552, 0.07048151642084122, 0.022043997421860695, -0.10053251683712006, 0.02625262178480625, -0.026719368994235992, 0.056434985250234604, 0.07223173975944519, 0.03715841472148895, -0.004781622905284166, -0.009803581051528454, 0.20770622789859772, -0.0874243900179863, -0.04723961651325226, -0.13825735449790955, 0.17673173546791077, -0.001052828156389296, -0.0003975645813625306, 0.01800137385725975, -0.06588234007358551, -0.033786386251449585, 0.1896292120218277, 0.1470598429441452, -0.029347021132707596, -0.029267681762576103, 0.015183619223535061, -0.008211205713450909, -0.025942577049136162, 0.08452481031417847, 0.11047359555959702, 0.07927070558071136, -0.037780072540044785, -0.023258138447999954, -0.023997368291020393, -0.07685183733701706, -0.021242577582597733, 0.10198090225458145, 0.013384111225605011, 0.0009883320890367031, -0.028451962396502495, 0.10550926625728607, -0.09055441617965698, -0.1321565955877304, 0.033825043588876724, -0.16021397709846497, -0.1915574073791504, -0.047652535140514374, 0.050855521112680435, 0.04006544128060341, 0.07628796994686127, -0.004332731477916241, -0.05485544353723526, 0.12206314504146576, 0.0064260587096214294, -0.04075736179947853, -0.12374373525381088, 0.06877391040325165, -0.123978391289711, 0.19270706176757812, -0.051296137273311615, 0.012019800022244453, 0.12737849354743958, 0.045784514397382736, -0.10222910344600677, 0.019513018429279327, 0.081261545419693, -0.13838376104831696, 0.03882518410682678, 0.19202883541584015, -0.03956452012062073, 0.13043814897537231, 0.025304682552814484, -0.08979781717061996, 0.016359694302082062, -0.058059874922037125, -0.022642599418759346, -0.07494816184043884, -0.01538396067917347, -0.03684590011835098, 0.13099703192710876, 0.21879184246063232, -0.07215532660484314, -0.014973103068768978, -0.04086978733539581, 0.010397683829069138, 0.022709855809807777, 0.12139752507209778, -0.046633463352918625, -0.275168240070343, 0.01732088439166546, -0.02759096771478653, 0.023713309317827225, -0.20370663702487946, -0.08255016058683395, 0.037233930081129074, -0.06694181263446808, -0.05041858181357384, 0.1190161481499672, 0.06429930031299591, 0.05549224838614464, -0.052210304886102676, -0.07897093892097473, -0.030225267633795738, 0.17986081540584564, -0.18382956087589264, -0.06917708367109299 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - ID dataset. It achieves the following results on the evaluation set: - Loss: 0.3975 - Wer: 0.2633 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 0.78 | 100 | 4.5645 | 1.0 | | No log | 1.55 | 200 | 2.9016 | 1.0 | | No log | 2.33 | 300 | 2.2666 | 1.0982 | | No log | 3.1 | 400 | 0.6079 | 0.6376 | | 3.2188 | 3.88 | 500 | 0.4985 | 0.5008 | | 3.2188 | 4.65 | 600 | 0.4477 | 0.4469 | | 3.2188 | 5.43 | 700 | 0.3953 | 0.3915 | | 3.2188 | 6.2 | 800 | 0.4319 | 0.3921 | | 3.2188 | 6.98 | 900 | 0.4171 | 0.3698 | | 0.2193 | 7.75 | 1000 | 0.3957 | 0.3600 | | 0.2193 | 8.53 | 1100 | 0.3730 | 0.3493 | | 0.2193 | 9.3 | 1200 | 0.3780 | 0.3348 | | 0.2193 | 10.08 | 1300 | 0.4133 | 0.3568 | | 0.2193 | 10.85 | 1400 | 0.3984 | 0.3193 | | 0.1129 | 11.63 | 1500 | 0.3845 | 0.3174 | | 0.1129 | 12.4 | 1600 | 0.3882 | 0.3162 | | 0.1129 | 13.18 | 1700 | 0.3982 | 0.3008 | | 0.1129 | 13.95 | 1800 | 0.3902 | 0.3198 | | 0.1129 | 14.73 | 1900 | 0.4082 | 0.3237 | | 0.0765 | 15.5 | 2000 | 0.3732 | 0.3126 | | 0.0765 | 16.28 | 2100 | 0.3893 | 0.3001 | | 0.0765 | 17.05 | 2200 | 0.4168 | 0.3083 | | 0.0765 | 17.83 | 2300 | 0.4193 | 0.3044 | | 0.0765 | 18.6 | 2400 | 0.4006 | 0.3013 | | 0.0588 | 19.38 | 2500 | 0.3836 | 0.2892 | | 0.0588 | 20.16 | 2600 | 0.3761 | 0.2903 | | 0.0588 | 20.93 | 2700 | 0.3895 | 0.2930 | | 0.0588 | 21.71 | 2800 | 0.3885 | 0.2791 | | 0.0588 | 22.48 | 2900 | 0.3902 | 0.2891 | | 0.0448 | 23.26 | 3000 | 0.4200 | 0.2849 | | 0.0448 | 24.03 | 3100 | 0.4013 | 0.2799 | | 0.0448 | 24.81 | 3200 | 0.4039 | 0.2731 | | 0.0448 | 25.58 | 3300 | 0.3970 | 0.2647 | | 0.0448 | 26.36 | 3400 | 0.4081 | 0.2690 | | 0.0351 | 27.13 | 3500 | 0.4090 | 0.2674 | | 0.0351 | 27.91 | 3600 | 0.3953 | 0.2663 | | 0.0351 | 28.68 | 3700 | 0.4044 | 0.2650 | | 0.0351 | 29.46 | 3800 | 0.3969 | 0.2646 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.11.0
{"language": ["id"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "robust-speech-event"], "datasets": ["common_voice"], "model-index": [{"name": "XLS-R-300M - Indonesia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 8", "type": "mozilla-foundation/common_voice_8_0", "args": "sv-SE"}, "metrics": [{"type": "wer", "value": 38.098, "name": "Test WER"}, {"type": "cer", "value": 14.261, "name": "Test CER"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xls-r-300m-id
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "robust-speech-event", "id", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #robust-speech-event #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - ID dataset. It achieves the following results on the evaluation set: * Loss: 0.3975 * Wer: 0.2633 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0003 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 64 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 30.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.1+cu102 * Datasets 1.17.1.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.17.1.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #robust-speech-event #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.17.1.dev0\n* Tokenizers 0.11.0" ]
[ 91, 159, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #robust-speech-event #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.17.1.dev0\n* Tokenizers 0.11.0" ]
[ -0.13737544417381287, 0.16891443729400635, -0.005101142916828394, 0.045729830861091614, 0.09399552643299103, 0.011626206338405609, 0.08394631743431091, 0.15368600189685822, -0.05164054036140442, 0.1362622231245041, 0.1067264974117279, 0.08647013455629349, 0.0886663943529129, 0.14896075427532196, -0.008808729238808155, -0.2902718186378479, 0.02684761770069599, -0.033335331827402115, -0.08800826966762543, 0.10708009451627731, 0.08333712071180344, -0.09986496716737747, 0.020965948700904846, 0.00825074315071106, -0.08601880818605423, -0.0330517441034317, -0.058746837079524994, -0.04807676002383232, 0.09836218506097794, 0.01737651601433754, 0.026231342926621437, 0.03617710620164871, 0.09412116557359695, -0.26069238781929016, 0.006697565782815218, 0.06152006983757019, 0.04835700988769531, 0.07259341329336166, 0.10986008495092392, -0.0038312915712594986, 0.10295731574296951, -0.07664304226636887, 0.038596007972955704, 0.0526193268597126, -0.08128520846366882, -0.24905310571193695, -0.08214693516492844, 0.03972182050347328, 0.13790030777454376, 0.08769804239273071, -0.03697764128446579, 0.030700692906975746, -0.06795196235179901, 0.08508183062076569, 0.23577311635017395, -0.23152096569538116, -0.054145924746990204, -0.02997611090540886, 0.03328782692551613, 0.02297358028590679, -0.10173329710960388, -0.017033277079463005, 0.012879328802227974, 0.011380900628864765, 0.08795004338026047, 0.02427462302148342, 0.04964975267648697, -0.013421033509075642, -0.14062464237213135, -0.05652816966176033, 0.11918836832046509, 0.08661044389009476, -0.02083640918135643, -0.12943068146705627, -0.038135167211294174, -0.20424625277519226, -0.045505836606025696, 0.016280747950077057, 0.020864544436335564, -0.03554636240005493, -0.051516447216272354, 0.04309225082397461, -0.04372616112232208, -0.0687296986579895, 0.05261470377445221, 0.12089083343744278, 0.04965364187955856, -0.03884262219071388, 0.024524670094251633, 0.09260917454957962, 0.07362673431634903, -0.16226711869239807, -0.01220752578228712, 0.052258674055337906, -0.09882345050573349, 0.012422256171703339, -0.010326730087399483, 0.05238967388868332, 0.05745769292116165, 0.1379867047071457, -0.0006500474410131574, 0.08509965986013412, 0.03912808746099472, 0.0006009425269439816, -0.058831412345170975, 0.1486222892999649, -0.06513570249080658, -0.11248859763145447, -0.03815934434533119, 0.11427662521600723, 0.013243948109447956, -0.01691472716629505, -0.06482996791601181, 0.024114418774843216, 0.11410949379205704, 0.06037530303001404, 0.010997328907251358, 0.00805912259966135, -0.06567703187465668, -0.025559309870004654, 0.022021768614649773, -0.11721748858690262, 0.03346409648656845, 0.05986151471734047, -0.06684425473213196, 0.01243955921381712, -0.02466273494064808, 0.00746671948581934, -0.054388951510190964, 0.09623053669929504, -0.04915237799286842, -0.026103412732481956, -0.053254369646310806, -0.07970500737428665, 0.03927375748753548, -0.09647436439990997, -0.005263797007501125, -0.04245304688811302, -0.07997860014438629, -0.07663033902645111, 0.04611138999462128, -0.06826594471931458, -0.10568549484014511, -0.09501069784164429, -0.08555318415164948, 0.0619332380592823, -0.02507181279361248, 0.17082351446151733, -0.05957481265068054, 0.09562259912490845, 0.0016028683166950941, 0.06089525297284126, 0.10549306869506836, 0.07685883343219757, -0.015433062799274921, 0.04973935708403587, -0.09909211844205856, 0.10457096993923187, -0.09193041175603867, 0.025972958654165268, -0.1480170488357544, -0.09559411555528641, -0.0030901392456144094, -0.006655809469521046, 0.10694212466478348, 0.15128038823604584, -0.18107765913009644, -0.08876794576644897, 0.16211487352848053, -0.038305774331092834, -0.07355391979217529, 0.13393568992614746, -0.0021014083176851273, -0.07113153487443924, 0.014257710427045822, 0.20318518579006195, 0.09176667779684067, -0.09463514387607574, -0.0062685380689799786, -0.04293473809957504, 0.12304669618606567, 0.01898762956261635, 0.09559955447912216, -0.06987054646015167, 0.04632732644677162, -0.001993706915527582, -0.025970635935664177, 0.05373053625226021, -0.0832362100481987, -0.06765084713697433, -0.01041396614164114, -0.07708100229501724, 0.003476468613371253, 0.04358743503689766, 0.0005267203086987138, -0.07925137132406235, -0.13035021722316742, -0.050922345370054245, 0.11209387332201004, -0.11060589551925659, 0.02149963565170765, -0.06616014242172241, 0.06036410108208656, -0.002848540199920535, 0.0016331402584910393, -0.14920973777770996, 0.0056260693818330765, 0.05006426200270653, -0.05413522571325302, -0.0002758616174105555, -0.04670362547039986, 0.055732611566782, 0.035808682441711426, -0.020133433863520622, -0.07235443592071533, -0.024961713701486588, -0.010513998568058014, -0.06521936506032944, -0.23211820423603058, -0.07514143735170364, -0.02028445154428482, 0.1720026433467865, -0.16855524480342865, -0.005861970130354166, 0.029606999829411507, 0.12859074771404266, 0.015346975065767765, -0.070271335542202, 0.028628233820199966, 0.053206268697977066, -0.01683780550956726, -0.07815611362457275, 0.02685951441526413, 0.005067333113402128, -0.10427571833133698, 0.02590983361005783, -0.13302502036094666, 0.045811235904693604, 0.0828046053647995, 0.06659086048603058, -0.05782747268676758, -0.04801269993185997, -0.06017572432756424, -0.05362720787525177, -0.019025182351469994, 0.007152427453547716, 0.15572044253349304, 0.032741203904151917, 0.08856663107872009, -0.07290218025445938, -0.056825559586286545, 0.04353111609816551, 0.026115722954273224, -0.01247267983853817, 0.15958106517791748, 0.07862865924835205, -0.010105425491929054, 0.08943868428468704, 0.034218575805425644, -0.03439453989267349, 0.12486892938613892, -0.07442695647478104, -0.09662070870399475, -0.03226188197731972, 0.018979625776410103, 0.017423851415514946, 0.10855318605899811, -0.1787167340517044, -0.011162027716636658, 0.04079499468207359, 0.03216060996055603, 0.011530936695635319, -0.17010752856731415, 0.0030925245955586433, 0.03525911271572113, -0.09057901799678802, -0.022489918395876884, 0.010180213488638401, -0.017143679782748222, 0.08253710716962814, 0.014125659130513668, -0.08202635496854782, -0.0375521145761013, -0.04090972989797592, -0.09579295665025711, 0.15932303667068481, -0.09546615183353424, -0.13883279263973236, -0.0969761312007904, -0.027521727606654167, -0.014997821301221848, -0.01490053441375494, 0.050056494772434235, -0.09831757098436356, -0.03825652599334717, -0.06820272654294968, 0.012223353609442711, -0.044764287769794464, 0.04055306315422058, 0.041097432374954224, -0.017256535589694977, 0.044193118810653687, -0.08717429637908936, 0.01038463693112135, -0.021173903718590736, 0.006673122756183147, 0.00041070551378652453, 0.03159535303711891, 0.09145337343215942, 0.17037999629974365, 0.07025524973869324, 0.044475164264440536, -0.027862073853611946, 0.19223642349243164, -0.1320367306470871, 0.012784217484295368, 0.1175629273056984, 0.012509905733168125, 0.04857337474822998, 0.15598976612091064, 0.031865738332271576, -0.08051208406686783, 0.012893542647361755, 0.030806703492999077, -0.016790146008133888, -0.2281758189201355, -0.0469222292304039, -0.07185287773609161, -0.015858635306358337, 0.10620975494384766, 0.03337381035089493, -0.01920226402580738, 0.015674535185098648, -0.026010235771536827, -0.03129331395030022, 0.025628093630075455, 0.04223751276731491, 0.06192441284656525, 0.038566846400499344, 0.10662849247455597, -0.001965353498235345, -0.018407093361020088, 0.02565312758088112, -0.004724479280412197, 0.23124860227108002, -0.0043121157214045525, 0.19969217479228973, 0.03215405344963074, 0.16257677972316742, -0.004110604524612427, 0.053656067699193954, 0.01607387699186802, 0.009524056687951088, 0.018689747899770737, -0.05140423774719238, -0.03874092176556587, 0.01964269019663334, 0.13441383838653564, 0.009979012422263622, -0.11216071248054504, 0.03743646666407585, 0.009515411220490932, 0.36512455344200134, 0.09943018108606339, -0.2735837697982788, -0.07725933194160461, 0.0031849045772105455, -0.07256823778152466, -0.04213743656873703, 0.03795633465051651, 0.10910134017467499, -0.08129564672708511, 0.0704423263669014, -0.04494726285338402, 0.08891259133815765, -0.08825472742319107, 0.0034207236021757126, 0.06102617457509041, 0.09384647011756897, 0.010346311144530773, 0.04413103684782982, -0.2349245250225067, 0.2665626108646393, -0.02161593921482563, 0.07538215816020966, -0.06084253266453743, 0.04466371610760689, 0.013479011133313179, -0.05838361755013466, 0.1002129316329956, 0.0005410389858298004, -0.0969579666852951, -0.1204187422990799, -0.12955941259860992, 0.0105630187317729, 0.12294448912143707, -0.07566516846418381, 0.10880277305841446, -0.030846094712615013, -0.04979923740029335, 0.029048902913928032, -0.07748234272003174, -0.08998126536607742, -0.10562928020954132, 0.03963891416788101, 0.03217096999287605, 0.07097363471984863, -0.06871749460697174, -0.09338663518428802, -0.07373407483100891, 0.1392347812652588, -0.13523703813552856, -0.034975629299879074, -0.1336391419172287, 0.0443539023399353, 0.17536471784114838, -0.07163802534341812, 0.04566651210188866, 0.013141236267983913, 0.12138739228248596, 0.022924114018678665, -0.01333415973931551, 0.10334692150354385, -0.08447405695915222, -0.20326460897922516, -0.06037897616624832, 0.19429099559783936, 0.028187328949570656, 0.06881009787321091, -0.0309558417648077, 0.029632775112986565, -0.019039155915379524, -0.07350374758243561, 0.09728013724088669, 0.05414358526468277, 0.002735528163611889, 0.033689867705106735, -0.009593428112566471, 0.009428572840988636, -0.08368725329637527, -0.06879652291536331, 0.12258221954107285, 0.2735109031200409, -0.08714665472507477, 0.07360641658306122, 0.0517151914536953, -0.045734286308288574, -0.15336962044239044, -0.04006761312484741, 0.1409369856119156, 0.04979090765118599, -0.03489569574594498, -0.21189750730991364, 0.018376555293798447, 0.07198533415794373, -0.022300267592072487, 0.08223484456539154, -0.339899480342865, -0.13357096910476685, 0.09099358320236206, 0.046936120837926865, -0.006189498119056225, -0.142528235912323, -0.06711037456989288, -0.008775904774665833, -0.08877775818109512, 0.03874816372990608, 0.039054710417985916, 0.12233863770961761, 0.0003084396885242313, 0.029518524184823036, 0.018889665603637695, -0.04768696427345276, 0.13493980467319489, 0.028096511960029602, 0.014732273295521736, -0.007101526949554682, 0.022174976766109467, -0.0683874636888504, -0.06541386246681213, 0.022916613146662712, -0.0807470753788948, 0.03586071357131004, -0.1353427618741989, -0.024866383522748947, -0.08439648151397705, 0.012510783039033413, -0.040464840829372406, -0.0245169959962368, -0.012758903205394745, 0.030159199610352516, 0.10331251472234726, 0.01570994406938553, 0.09674955159425735, -0.0465218685567379, 0.09572800248861313, 0.10212225466966629, 0.0897657722234726, -0.013405240140855312, -0.11022873967885971, -0.019959306344389915, -0.011196237057447433, 0.0286584310233593, -0.09764795750379562, 0.020926816388964653, 0.14387477934360504, 0.05772282928228378, 0.12770280241966248, 0.04350702837109566, -0.08193767070770264, -0.004358314443379641, 0.04718257486820221, -0.08267813920974731, -0.16128383576869965, -0.01192488707602024, 0.0028154340106993914, -0.13134139776229858, -0.000246052019065246, 0.10504399985074997, -0.025355268269777298, -0.007678507827222347, 0.008719316683709621, 0.05373207852244377, -0.011125205084681511, 0.22374014556407928, 0.017973415553569794, 0.09681116044521332, -0.1116865873336792, 0.084532231092453, 0.0470217764377594, -0.11920035630464554, 0.05541156232357025, 0.09806402027606964, -0.05976850911974907, -0.006425690371543169, 0.03524525836110115, 0.0475517176091671, 0.05109499767422676, -0.042175035923719406, -0.09669457376003265, -0.14563991129398346, 0.11211855709552765, 0.05062500759959221, 0.016774218529462814, 0.03432454541325569, -0.030283022671937943, 0.028154566884040833, -0.10597480833530426, 0.10427483171224594, 0.11212154477834702, 0.056275248527526855, -0.1194918230175972, 0.09432252496480942, 0.0008537672110833228, 0.008830437436699867, 0.002946559339761734, -0.01010072696954012, -0.09129682928323746, 0.01811341755092144, -0.07769332081079483, -0.013277034275233746, -0.06954057514667511, -0.006606855429708958, 0.006445347797125578, -0.05894922837615013, -0.04394430294632912, 0.027189645916223526, -0.10164254903793335, -0.05963082239031792, -0.04339434206485748, 0.05858873948454857, -0.0982285663485527, -0.006951343268156052, 0.02972526289522648, -0.13448549807071686, 0.1147378459572792, 0.056089382618665695, 0.012303777039051056, 0.0034637274220585823, -0.075590580701828, -0.01322503574192524, 0.02962065301835537, 0.0038923893589526415, 0.03416278585791588, -0.18782323598861694, -0.01340432558208704, -0.02881833165884018, -0.0024864203296601772, -0.019245965406298637, 0.021874308586120605, -0.12198297679424286, -0.01020035706460476, -0.044645216315984726, -0.027675651013851166, -0.057570137083530426, 0.05821102485060692, 0.07774047553539276, 0.018542850390076637, 0.14237326383590698, -0.06590045243501663, 0.07244578748941422, -0.2122444063425064, 0.009696646593511105, -0.022375797852873802, -0.053935855627059937, -0.06257881224155426, -0.008613706566393375, 0.11193083226680756, -0.05160791426897049, 0.0519767627120018, -0.04890524968504906, 0.03420023247599602, 0.017148153856396675, -0.08721669018268585, 0.013865998014807701, 0.04451601207256317, 0.17064470052719116, 0.05303552746772766, -0.02880426123738289, 0.08033255487680435, -0.015497220680117607, 0.057678431272506714, 0.14110271632671356, 0.14795930683612823, 0.13164833188056946, 0.09512711316347122, 0.0883554071187973, 0.06706807762384415, -0.1478564292192459, -0.1582905352115631, 0.15171325206756592, -0.06863589584827423, 0.1449769288301468, -0.014829905703663826, 0.22196310758590698, 0.0768498107790947, -0.1953064203262329, 0.056435342878103256, -0.02330024167895317, -0.08075221627950668, -0.1100969985127449, -0.09540440142154694, -0.0818852037191391, -0.15415307879447937, 0.013577322475612164, -0.10284391790628433, 0.07005216926336288, 0.040400151163339615, 0.04738522693514824, 0.0397556833922863, 0.07363538444042206, 0.058510784059762955, -0.005430820398032665, 0.11939238011837006, 0.02425568550825119, -0.02939823642373085, -0.03484932705760002, -0.1009756401181221, 0.05216573551297188, -0.02468612603843212, 0.08249258249998093, -0.03600577637553215, -0.11227606981992722, 0.07563841342926025, 0.018017029389739037, -0.1020897924900055, 0.025575479492545128, -0.028295982629060745, 0.06092504784464836, 0.12860579788684845, 0.048365429043769836, -0.00838072132319212, -0.01616082526743412, 0.2021622508764267, -0.08956697583198547, -0.04788431152701378, -0.12110964208841324, 0.18757234513759613, 0.005240230821073055, -0.004037126898765564, 0.03797994926571846, -0.06812892109155655, -0.03375469893217087, 0.1570356786251068, 0.14463360607624054, -0.022347940132021904, -0.03094397485256195, 0.03479587286710739, -0.008313191123306751, -0.03141367807984352, 0.07639860361814499, 0.1310606300830841, 0.08440268784761429, -0.04236340522766113, -0.004153830464929342, -0.037696972489356995, -0.07187438011169434, -0.044045668095350266, 0.0740581825375557, 0.008047012612223625, -0.02705278806388378, -0.013572286814451218, 0.10470382869243622, -0.08894385397434235, -0.1439470648765564, 0.011410662904381752, -0.18090872466564178, -0.197593092918396, -0.027525197714567184, 0.04346166551113129, 0.06024886295199394, 0.04743588715791702, 0.005623906385153532, -0.02315644733607769, 0.1422627717256546, 0.0068298508413136005, -0.03696853667497635, -0.08470787853002548, 0.05480975657701492, -0.09378060698509216, 0.1754504144191742, -0.032636385411024094, 0.034468088299036026, 0.10912569612264633, 0.0761873722076416, -0.10839606076478958, 0.030719485133886337, 0.08867795020341873, -0.12220311909914017, 0.05779207497835159, 0.19446420669555664, -0.039326515048742294, 0.1142992302775383, 0.043751802295446396, -0.07407685369253159, 0.0024455939419567585, -0.06767649948596954, -0.02766658365726471, -0.05976622551679611, 0.0040782601572573185, -0.035637643188238144, 0.13587437570095062, 0.18187488615512848, -0.07730729132890701, -0.025035016238689423, -0.034093692898750305, 0.0023866358678787947, 0.014475664123892784, 0.12758439779281616, -0.02863902971148491, -0.26233071088790894, 0.027329735457897186, -0.02673327922821045, 0.026156239211559296, -0.17470751702785492, -0.07751355320215225, 0.010645764879882336, -0.05566728860139847, -0.07781040668487549, 0.10924405604600906, 0.0365859717130661, 0.0353752076625824, -0.05667721852660179, -0.06433102488517761, -0.02299966849386692, 0.17971043288707733, -0.18327713012695312, -0.06171240657567978 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - MN dataset. It achieves the following results on the evaluation set: - Loss: 0.5502 - Wer: 0.4042 ## Training and evaluation data Evaluation is conducted in Notebook, you can see within the repo "notebook_evaluation_wav2vec2_mn.ipynb" Test WER without LM wer = 58.2171 % cer = 16.0670 % Test WER using wer = 31.3919 % cer = 10.2565 % How to use eval.py ``` huggingface-cli login #login to huggingface for getting auth token to access the common voice v8 #running with LM python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-mn --dataset mozilla-foundation/common_voice_8_0 --config mn --split test # running without LM python eval.py --model_id ayameRushia/wav2vec2-large-xls-r-300m-mn --dataset mozilla-foundation/common_voice_8_0 --config mn --split test --greedy ``` ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 40.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 6.35 | 400 | 0.9380 | 0.7902 | | 3.2674 | 12.7 | 800 | 0.5794 | 0.5309 | | 0.7531 | 19.05 | 1200 | 0.5749 | 0.4815 | | 0.5382 | 25.4 | 1600 | 0.5530 | 0.4447 | | 0.4293 | 31.75 | 2000 | 0.5709 | 0.4237 | | 0.4293 | 38.1 | 2400 | 0.5476 | 0.4059 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.18.3 - Tokenizers 0.11.0
{"language": ["mn"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "robust-speech-event"], "datasets": ["mozilla-foundation/common_voice_8_0"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-mn", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 8", "type": "mozilla-foundation/common_voice_8_0", "args": "mn"}, "metrics": [{"type": "wer", "value": 31.3919, "name": "Test WER using LM"}, {"type": "cer", "value": 10.2565, "name": "Test CER using LM"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Dev Data", "type": "speech-recognition-community-v2/dev_data", "args": "mn"}, "metrics": [{"type": "wer", "value": 65.26, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Test Data", "type": "speech-recognition-community-v2/eval_data", "args": "mn"}, "metrics": [{"type": "wer", "value": 63.09, "name": "Test WER"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xls-r-300m-mn
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "mozilla-foundation/common_voice_8_0", "mn", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "mn" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #mn #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - MN dataset. It achieves the following results on the evaluation set: * Loss: 0.5502 * Wer: 0.4042 Training and evaluation data ---------------------------- Evaluation is conducted in Notebook, you can see within the repo "notebook\_evaluation\_wav2vec2\_mn.ipynb" Test WER without LM wer = 58.2171 % cer = 16.0670 % Test WER using wer = 31.3919 % cer = 10.2565 % How to use URL Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0003 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 64 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 40.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.1+cu102 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 40.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #mn #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 40.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 111, 159, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #mozilla-foundation/common_voice_8_0 #mn #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 40.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.1+cu102\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.13510292768478394, 0.12448471039533615, -0.005370603874325752, 0.04649150371551514, 0.1182214617729187, 0.005783065687865019, 0.10255559533834457, 0.1448727399110794, -0.09206388890743256, 0.10490849614143372, 0.07884195446968079, 0.08272586762905121, 0.08366817981004715, 0.09909352660179138, -0.01032936479896307, -0.3011786937713623, 0.015171813778579235, -0.02831410802900791, -0.12334714084863663, 0.10375779867172241, 0.09769953042268753, -0.08943444490432739, 0.03392615169286728, 0.0308302640914917, -0.0931830033659935, -0.0016571036539971828, -0.04625971242785454, -0.04690468683838844, 0.07733843475580215, 0.04881257936358452, 0.04430224746465683, 0.029223764315247536, 0.08838308602571487, -0.24671955406665802, 0.008854218758642673, 0.07135868817567825, 0.04031864181160927, 0.05949140712618828, 0.12494239211082458, -0.01796465367078781, 0.11558917909860611, -0.046840228140354156, 0.03533884510397911, 0.07035069912672043, -0.0975971445441246, -0.2422982007265091, -0.07877104729413986, 0.04857983440160751, 0.12464989721775055, 0.09908168017864227, -0.031914904713630676, 0.025178337469697, -0.08108464628458023, 0.09142693132162094, 0.22155819833278656, -0.21285012364387512, -0.07613492757081985, -0.021892916411161423, 0.039684515446424484, 0.018446624279022217, -0.11831621080636978, -0.022005418315529823, 0.01176515780389309, 0.0085220318287611, 0.08102405816316605, 0.007278361823409796, 0.018402136862277985, -0.0021932078525424004, -0.1411372721195221, -0.04927460104227066, 0.1312028020620346, 0.07785064727067947, -0.011315803974866867, -0.10197105258703232, -0.020354891195893288, -0.18890626728534698, -0.04693036898970604, 0.04263010621070862, 0.01862151548266411, -0.02860061451792717, -0.05322267487645149, 0.027957838028669357, -0.044179584830999374, -0.07610970735549927, 0.05593959614634514, 0.1327914297580719, 0.03699522092938423, -0.0316392220556736, -0.0013234446523711085, 0.10815618187189102, 0.08011207729578018, -0.17143765091896057, -0.02682657539844513, 0.03765667602419853, -0.11201804131269455, -0.0023152155335992575, -0.0072321039624512196, 0.03533301502466202, 0.03859566152095795, 0.13237424194812775, -0.00485709123313427, 0.08686034381389618, 0.02090463973581791, 0.022186316549777985, -0.07772542536258698, 0.17503872513771057, -0.0864148885011673, -0.08111061155796051, -0.045168187469244, 0.13604719936847687, -0.013154910877346992, -0.010485256090760231, -0.06528806686401367, 0.033807676285505295, 0.08566734939813614, 0.0634884312748909, 0.0067957001738250256, 0.032863810658454895, -0.05578623712062836, -0.030194206163287163, 0.0036193816922605038, -0.12679576873779297, 0.04313821718096733, 0.07021363824605942, -0.09722594916820526, 0.012473450042307377, -0.01992516778409481, 0.009084596298635006, -0.050172388553619385, 0.08166501671075821, -0.05789995566010475, 0.0005727445241063833, -0.0875236764550209, -0.08660827577114105, 0.03636831045150757, -0.026336994022130966, -0.008845280855894089, -0.056231122463941574, -0.09635043889284134, -0.07405088096857071, 0.049601152539253235, -0.06866626441478729, -0.07344406843185425, -0.07447640597820282, -0.10007050633430481, 0.0582854226231575, -0.01728067360818386, 0.17356152832508087, -0.05606057122349739, 0.08531597256660461, 0.02861942909657955, 0.039350658655166626, 0.12041887640953064, 0.07349690049886703, -0.020012449473142624, 0.05831879377365112, -0.11019918322563171, 0.10435432195663452, -0.11448425054550171, 0.07036293298006058, -0.12120945751667023, -0.10134126991033554, -0.024154076352715492, 0.001606086501851678, 0.09711319208145142, 0.13107824325561523, -0.1483086496591568, -0.09574408829212189, 0.1522880345582962, -0.04460706561803818, -0.08355081081390381, 0.11755010485649109, -0.0019480414921417832, -0.047483451664447784, 0.017446722835302353, 0.1571461707353592, 0.15454860031604767, -0.08926502615213394, -0.005684494040906429, -0.04666360467672348, 0.10916060209274292, 0.03817083314061165, 0.09263189882040024, -0.042618513107299805, 0.04266977682709694, 0.0049439226277172565, -0.02803809754550457, 0.06478425115346909, -0.08393174409866333, -0.08377784490585327, -0.009566173888742924, -0.07614847272634506, -0.0028947845567017794, 0.06343050301074982, 0.009011426940560341, -0.0633859783411026, -0.13840574026107788, -0.012363116256892681, 0.11567697674036026, -0.09714851528406143, 0.01357978954911232, -0.08505525439977646, 0.05792233720421791, -0.006633734330534935, 0.0022533079609274864, -0.15388719737529755, -0.026962589472532272, 0.033292219042778015, -0.06597540527582169, 0.007101446855813265, -0.039779748767614365, 0.07013380527496338, 0.04127499833703041, -0.025737443938851357, -0.07321301847696304, -0.04358641430735588, -0.018579084426164627, -0.04528426751494408, -0.22654466331005096, -0.09092071652412415, -0.015454010106623173, 0.18529415130615234, -0.19908355176448822, 0.010336220264434814, 0.05183982849121094, 0.12558507919311523, 0.02033604122698307, -0.052269358187913895, 0.022699005901813507, 0.05601730942726135, -0.020663458853960037, -0.08349907398223877, 0.02745436131954193, 0.015172137878835201, -0.08477892726659775, 0.021347172558307648, -0.1296403706073761, 0.08971232175827026, 0.08173254877328873, 0.014803611673414707, -0.05982537567615509, -0.04056532308459282, -0.0688333660364151, -0.06009063497185707, -0.0168681088835001, -0.014242064207792282, 0.16157600283622742, 0.01930369809269905, 0.10878840833902359, -0.07191210985183716, -0.05086437612771988, 0.03219988942146301, 0.022300561890006065, -0.010107536800205708, 0.15001219511032104, 0.08209332823753357, -0.027440696954727173, 0.09434562176465988, -0.008212417364120483, -0.06634388118982315, 0.16635340452194214, -0.07505295425653458, -0.09471002966165543, -0.03494436293840408, 0.02474208176136017, 0.03299465402960777, 0.10391253232955933, -0.18916025757789612, -0.022836903110146523, 0.026101864874362946, 0.019241487607359886, 0.034117668867111206, -0.18021906912326813, 0.0005729469121433794, 0.03193974867463112, -0.08796387910842896, -0.005557055119425058, 0.003555359784513712, -0.010859833098948002, 0.08090256154537201, 0.0030076538678258657, -0.09115613996982574, -0.03491343930363655, -0.04437003284692764, -0.090042345225811, 0.1660325527191162, -0.07954021543264389, -0.14031991362571716, -0.13585278391838074, -0.021192004904150963, -0.038820572197437286, -0.020879628136754036, 0.028315206989645958, -0.088034488260746, -0.03602753207087517, -0.06878641992807388, 0.02134104073047638, -0.041409533470869064, 0.015177938155829906, 0.03349389135837555, 0.010463577695190907, 0.061251815408468246, -0.10183034092187881, 0.01798912137746811, -0.010845507495105267, -0.03547036275267601, -0.007609230466187, 0.02060086652636528, 0.09971695393323898, 0.18308643996715546, 0.07470861077308655, 0.05325128510594368, -0.019538989290595055, 0.17975540459156036, -0.13950207829475403, 0.009516805410385132, 0.1109941229224205, 0.02570197917521, 0.049283020198345184, 0.14447931945323944, 0.04838903993368149, -0.06945772469043732, 0.012368372641503811, 0.049309808760881424, -0.020676974207162857, -0.22275744378566742, -0.028977110981941223, -0.08054150640964508, -0.01980854943394661, 0.10043925791978836, 0.032711826264858246, 0.017487112432718277, 0.018027132377028465, -0.028966793790459633, -0.02100534737110138, 0.05344006419181824, 0.04548009857535362, 0.041733648627996445, 0.050208479166030884, 0.11477510631084442, -0.014532074332237244, -0.03594563901424408, 0.019768143072724342, -0.008955394849181175, 0.24321117997169495, -0.010197172872722149, 0.18972671031951904, 0.04911300912499428, 0.13563506305217743, 0.0014492588816210628, 0.05071214213967323, 0.003902936354279518, -0.005916443653404713, 0.04341898858547211, -0.05919724330306053, -0.012888622470200062, 0.024577762931585312, 0.12088858336210251, 0.029520219191908836, -0.10895238071680069, 0.011626609601080418, 0.01976311206817627, 0.368625283241272, 0.08013838529586792, -0.28926947712898254, -0.0670858770608902, 0.004154934082180262, -0.06423770636320114, -0.041279587894678116, 0.03530420735478401, 0.10370148718357086, -0.08314523100852966, 0.0702296793460846, -0.05730119347572327, 0.09741146862506866, -0.07214327156543732, -0.005964362528175116, 0.10235657542943954, 0.10038571804761887, 0.006768538150936365, 0.06366846710443497, -0.23669390380382538, 0.2544383108615875, -0.014814896509051323, 0.08154884725809097, -0.0413348451256752, 0.052836447954177856, 0.04776144400238991, -0.007168116047978401, 0.06476669013500214, -0.005830490496009588, -0.08926411718130112, -0.16802090406417847, -0.10577330738306046, 0.00028500607004389167, 0.11963162571191788, -0.05921056494116783, 0.1236017569899559, -0.042161211371421814, -0.05176905542612076, 0.034936971962451935, -0.06907206773757935, -0.11652582138776779, -0.10630851984024048, 0.0577315054833889, 0.026389816775918007, 0.08407015353441238, -0.09079878777265549, -0.09471973776817322, -0.03802110627293587, 0.15460363030433655, -0.1261463165283203, -0.029508644714951515, -0.13695353269577026, 0.0455881766974926, 0.15816490352153778, -0.06376269459724426, 0.0251662265509367, 0.015260269865393639, 0.15012703835964203, 0.043751828372478485, -0.02130695804953575, 0.09582226723432541, -0.0840597078204155, -0.19224299490451813, -0.044472724199295044, 0.18268416821956635, 0.024101993069052696, 0.05923997610807419, -0.027430158108472824, 0.013945831917226315, -0.010985806584358215, -0.08013496547937393, 0.06839444488286972, 0.041838061064481735, -0.011982106603682041, 0.05226661264896393, -0.03153674304485321, 0.014744565822184086, -0.0789559930562973, -0.043399326503276825, 0.09011906385421753, 0.21755912899971008, -0.07286660373210907, 0.023840857669711113, 0.033776622265577316, -0.06236717104911804, -0.1564478874206543, -0.0019410334061831236, 0.13849270343780518, 0.03930530324578285, -0.04866291582584381, -0.21213924884796143, 0.009148456156253815, 0.06719410419464111, -0.021702008321881294, 0.11202052980661392, -0.3647509515285492, -0.12795504927635193, 0.08727572113275528, 0.046374838799238205, -0.047999341040849686, -0.17361575365066528, -0.0718284323811531, -0.018799657002091408, -0.08932927250862122, 0.03659658133983612, -0.008877210319042206, 0.12748636305332184, 0.001564315753057599, 0.023129291832447052, 0.014728240668773651, -0.051401782780885696, 0.14448828995227814, 0.025098303332924843, 0.038953423500061035, -0.010903329588472843, 0.013964159414172173, 0.021328002214431763, -0.07506376504898071, 0.024018842726945877, -0.06295884400606155, 0.022433819249272346, -0.16133958101272583, -0.02873731032013893, -0.09643085300922394, 0.017950808629393578, -0.04713292419910431, -0.010001269169151783, -0.012313351966440678, 0.04559805989265442, 0.08230240643024445, 0.02769205905497074, 0.09720584750175476, -0.06426694989204407, 0.11161280423402786, 0.14247912168502808, 0.09960334748029709, -0.006795878522098064, -0.12898436188697815, -0.02200457826256752, 0.016072440892457962, 0.04080311954021454, -0.09163299947977066, 0.03695613145828247, 0.14281465113162994, 0.04861273244023323, 0.1415233463048935, 0.056449487805366516, -0.08510762453079224, 0.002073985291644931, 0.05906349793076515, -0.06275391578674316, -0.12499850243330002, -0.041571084409952164, 0.018415581434965134, -0.12704403698444366, -0.008400766178965569, 0.10973531007766724, -0.032222021371126175, 0.011471627280116081, 0.01944475807249546, 0.046475913375616074, -0.031200354918837547, 0.2274957150220871, 0.0346103310585022, 0.10029103606939316, -0.09430216997861862, 0.07111048698425293, 0.049180127680301666, -0.08313418179750443, 0.021239468827843666, 0.11256065219640732, -0.042968858033418655, -0.022778160870075226, -0.008024629205465317, 0.08058429509401321, 0.03646869212388992, -0.056599099189043045, -0.12276780605316162, -0.15631742775440216, 0.10410262644290924, 0.061446413397789, 0.01426741387695074, 0.03221374377608299, -0.02860189601778984, 0.03451206162571907, -0.09755714237689972, 0.10190855711698532, 0.09920107573270798, 0.05932622030377388, -0.12828968465328217, 0.11051987111568451, 0.012724842876195908, 0.008407793007791042, 0.005870963912457228, -0.02316822111606598, -0.10472437739372253, 0.035071808844804764, -0.11645971983671188, -0.018402058631181717, -0.047765664756298065, -0.0037002938333898783, 0.009368019178509712, -0.0558919832110405, -0.07054092735052109, 0.030579017475247383, -0.11052830517292023, -0.04978353530168533, -0.03425808995962143, 0.06306644529104233, -0.10249950736761093, -0.01069455873221159, 0.034627608954906464, -0.13898888230323792, 0.0848136618733406, 0.05478156730532646, 0.014193180948495865, 0.0018783302512019873, -0.06972245872020721, -0.011182566173374653, 0.021776720881462097, 0.005513748154044151, 0.0346694216132164, -0.17983952164649963, 0.004427871201187372, -0.030908366665244102, 0.010633617639541626, -0.020906226709485054, -0.007628302089869976, -0.12064461410045624, -0.002665602834895253, -0.025371940806508064, -0.04579322785139084, -0.045053765177726746, 0.07450364530086517, 0.07608326524496078, 0.01744941622018814, 0.14970459043979645, -0.062388304620981216, 0.06258895248174667, -0.23265965282917023, 0.01064209546893835, -0.001659768051467836, -0.06129467859864235, -0.010792681016027927, -0.016000675037503242, 0.10639839619398117, -0.06187548488378525, 0.07268090546131134, -0.010310659185051918, 0.03982215002179146, 0.02319982834160328, -0.09943164885044098, 0.043386247009038925, 0.06660941243171692, 0.13689403235912323, 0.05894668772816658, -0.013992172665894032, 0.09765777736902237, -0.02143867500126362, 0.059865374118089676, 0.12187905609607697, 0.15387557446956635, 0.11843618005514145, 0.07342714816331863, 0.0883684903383255, 0.10176104307174683, -0.14482998847961426, -0.12968653440475464, 0.13139210641384125, -0.07371474802494049, 0.15039710700511932, -0.034805312752723694, 0.18916544318199158, 0.12196030467748642, -0.18409988284111023, 0.0722668394446373, -0.030971424654126167, -0.08968433737754822, -0.09808678925037384, -0.0984141081571579, -0.07130807638168335, -0.16694019734859467, 0.019080063328146935, -0.10158927738666534, 0.06673697382211685, 0.048761218786239624, 0.04603325203061104, 0.03328791260719299, 0.10295131802558899, 0.0675627589225769, -0.005103710573166609, 0.12452992796897888, 0.00003365783049957827, -0.020354947075247765, -0.05232609808444977, -0.09306797385215759, 0.04738110303878784, -0.03892175853252411, 0.06786935031414032, -0.03119041956961155, -0.10995252430438995, 0.06819138675928116, 0.02326057106256485, -0.09544812142848969, 0.029067520052194595, -0.03370150923728943, 0.052530430257320404, 0.07893090695142746, 0.03754004091024399, -0.0053343456238508224, -0.011514718644320965, 0.20908625423908234, -0.08933839946985245, -0.04993231222033501, -0.13018137216567993, 0.166080504655838, 0.0037527349777519703, -0.0026344137731939554, 0.019587641581892967, -0.06734421849250793, -0.031271908432245255, 0.18633022904396057, 0.14283423125743866, -0.03040706180036068, -0.02813531458377838, 0.01731802523136139, -0.008414732292294502, -0.026952074840664864, 0.08451183885335922, 0.11999797075986862, 0.08416924625635147, -0.03400770574808121, -0.01985894702374935, -0.024344375357031822, -0.07340490818023682, -0.02929292991757393, 0.10101243853569031, 0.018727576360106468, 0.0002727096725720912, -0.02794148586690426, 0.10502444952726364, -0.08894168585538864, -0.1260712444782257, 0.030960679054260254, -0.1590745449066162, -0.19174502789974213, -0.04386809095740318, 0.049391552805900574, 0.04072150960564613, 0.06982462853193283, -0.0018827030435204506, -0.04937616363167763, 0.1175832599401474, 0.005763930268585682, -0.03984384983778, -0.12178248167037964, 0.06272365152835846, -0.131851926445961, 0.19004115462303162, -0.05294662341475487, 0.00827016681432724, 0.12490665912628174, 0.046847082674503326, -0.10099872946739197, 0.024222707375884056, 0.08183789998292923, -0.14186768233776093, 0.04503100365400314, 0.19842368364334106, -0.039544470608234406, 0.13173972070217133, 0.023281820118427277, -0.07341562956571579, 0.015062165446579456, -0.06658107042312622, -0.019662991166114807, -0.07050587981939316, -0.012401510030031204, -0.0361657477915287, 0.1262744814157486, 0.2136147916316986, -0.07116582989692688, -0.017006345093250275, -0.043531544506549835, 0.009674946777522564, 0.01567632146179676, 0.12438587099313736, -0.043161541223526, -0.2706766426563263, 0.01651136949658394, -0.018763167783617973, 0.024384358897805214, -0.2038712054491043, -0.0800413116812706, 0.03360268473625183, -0.06943804025650024, -0.05460450053215027, 0.11647497862577438, 0.053519390523433685, 0.05012842267751694, -0.055084217339754105, -0.09434580057859421, -0.025648372247815132, 0.18298090994358063, -0.18766458332538605, -0.063428595662117 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Indonesia Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) in Indonesia using the [Common Voice](https://huggingface.co/datasets/common_voice) When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "id", split="test[:2%]"). processor = Wav2Vec2Processor.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model = Wav2Vec2ForCTC.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("common_voice", "id", split="test") wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model = Wav2Vec2ForCTC.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model.to("cuda") chars_to_ignore_regex = '[\\,\\?\\.\\!\\-\\;\\:\\"\\โ€œ]' resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() speech_array, sampling_rate = torchaudio.load(batch["path"]) resampler = torchaudio.transforms.Resample(sampling_rate, 16_000) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: WER = 20.072720 % ## Training Training using common voice dataset
{"language": "id", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "model-index": [{"name": "XLSR Wav2Vec2 Indonesia by Ayame Rushia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice id", "type": "common_voice", "args": "id"}, "metrics": [{"type": "wer", "value": "???", "name": "Test WER"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xlsr-indo-base
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "id", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Indonesia Fine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. Test Result: WER = 20.072720 % ## Training Training using common voice dataset
[ "# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\nThe model can be used directly (without a language model) as follows:", "## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 20.072720 %", "## Training\nTraining using common voice dataset" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\nThe model can be used directly (without a language model) as follows:", "## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 20.072720 %", "## Training\nTraining using common voice dataset" ]
[ 80, 60, 20, 34, 8 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\nThe model can be used directly (without a language model) as follows:## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 20.072720 %## Training\nTraining using common voice dataset" ]
[ -0.12025351077318192, 0.009708316996693611, -0.0012697172351181507, -0.08034420758485794, 0.06670685857534409, -0.05689452588558197, 0.15604336559772491, 0.08423783630132675, -0.01679180935025215, -0.046105023473501205, 0.003982253838330507, 0.021104294806718826, 0.08461610972881317, 0.07008467614650726, -0.019566627219319344, -0.27264946699142456, 0.061515651643276215, 0.004543761257082224, 0.17271625995635986, 0.15138886868953705, 0.0914001539349556, -0.05226314067840576, 0.013924689963459969, 0.14009854197502136, -0.09117182344198227, 0.002138967625796795, -0.006765191443264484, -0.15117350220680237, 0.12386472523212433, 0.045364152640104294, 0.06974399089813232, 0.07878392934799194, 0.06373395770788193, -0.15263120830059052, 0.020064275711774826, -0.009571810252964497, 0.046267881989479065, 0.0017155568348243833, 0.033622920513153076, 0.02792222611606121, 0.11253847181797028, 0.125790074467659, -0.05877002328634262, 0.03420362249016762, -0.06491449475288391, -0.2224620133638382, -0.03367052972316742, 0.038629770278930664, 0.13463476300239563, 0.12077535688877106, -0.07527565211057663, 0.09388919919729233, -0.14878708124160767, 0.04909946769475937, 0.07281786948442459, -0.2016242891550064, 0.008899586275219917, 0.03471701964735985, 0.06307817995548248, 0.07242844998836517, -0.04555140808224678, -0.0002123868907801807, 0.02280612289905548, 0.03946385160088539, 0.0007889344706200063, -0.09398341923952103, -0.18589378893375397, -0.06502243131399155, -0.10469015687704086, -0.0043280888348817825, 0.24408403038978577, 0.007782162632793188, -0.0953918993473053, -0.10174132883548737, 0.020976709201931953, 0.005599224008619785, -0.03188437968492508, -0.099010169506073, -0.00748042855411768, 0.03992059826850891, 0.04082480072975159, -0.013594507239758968, -0.10260628163814545, -0.11620329320430756, -0.026691749691963196, 0.09769138693809509, 0.01876894198358059, 0.018630068749189377, -0.12049301713705063, 0.044311683624982834, -0.12216288596391678, -0.08919715881347656, -0.03563322126865387, 0.03043391741812229, -0.04658254608511925, 0.06280648708343506, -0.059656430035829544, -0.16013140976428986, 0.055116068571805954, -0.07644829899072647, 0.021616201847791672, 0.02225637249648571, 0.0247914120554924, 0.0702960267663002, 0.023445283994078636, 0.14777028560638428, -0.0884464681148529, -0.06770828366279602, 0.017948070541024208, 0.01442689262330532, -0.007379597518593073, -0.03316629305481911, -0.05064552649855614, -0.07390903681516647, 0.046687375754117966, 0.05227116122841835, -0.08874819427728653, 0.022135183215141296, -0.03664128854870796, -0.057982828468084335, 0.0749756395816803, -0.10574426501989365, -0.06969568133354187, 0.032984934747219086, 0.011439281515777111, 0.1979353129863739, 0.06911651045084, 0.04728693887591362, -0.058852918446063995, -0.0025058751925826073, 0.03508436307311058, 0.040862392634153366, -0.01335783302783966, -0.05667334794998169, -0.004650891292840242, -0.0022507074754685163, -0.017697200179100037, -0.1045609712600708, -0.10780639201402664, -0.07931454479694366, -0.019625147804617882, 0.018610820174217224, -0.027056563645601273, -0.09300658851861954, -0.0005271944683045149, -0.03439886495471001, -0.09733787924051285, 0.02864527516067028, -0.04918881878256798, 0.06699846684932709, 0.0857158973813057, 0.08642164617776871, -0.001847237115725875, 0.09828824549913406, -0.08390749990940094, -0.00031330925412476063, -0.032000914216041565, 0.11326184868812561, -0.032019417732954025, -0.05373423174023628, -0.10507114231586456, -0.10216818004846573, -0.07798110693693161, 0.05519458279013634, 0.03184725344181061, 0.06742262840270996, -0.2222083956003189, -0.10159296542406082, 0.1961425542831421, -0.1519676297903061, -0.054265689104795456, 0.18538351356983185, 0.020496290177106857, 0.15922382473945618, 0.08991359919309616, 0.21435561776161194, 0.11394073814153671, -0.13101018965244293, 0.0716828852891922, 0.04172680154442787, -0.020065998658537865, -0.07397128641605377, 0.06790988147258759, -0.03806821256875992, -0.012370328418910503, 0.06070161610841751, -0.05212729796767235, 0.09114918112754822, -0.023868296295404434, -0.06910569220781326, 0.010171579197049141, -0.10469244420528412, 0.060872435569763184, -0.0011155563406646252, 0.05405239388346672, -0.0181905385106802, -0.02176055684685707, 0.08070613443851471, 0.1482086479663849, -0.12778271734714508, 0.017294039949774742, -0.13486166298389435, 0.05961839482188225, -0.14081615209579468, 0.011896547861397266, -0.12078548967838287, 0.196232408285141, -0.016708994284272194, 0.009251839481294155, 0.08299969881772995, 0.14478586614131927, 0.008077332749962807, -0.03489241003990173, -0.0176320169121027, -0.04376298561692238, 0.052819084376096725, -0.0035554668866097927, 0.022427866235375404, -0.014853387139737606, -0.03071535751223564, -0.02632381208240986, 0.08004233241081238, -0.10166481137275696, -0.042848020792007446, -0.04608898237347603, -0.0330524817109108, -0.04425919055938721, 0.019566504284739494, 0.11293748021125793, 0.10905582457780838, 0.04593956470489502, 0.039907149970531464, 0.03402357175946236, 0.026108568534255028, -0.09495260566473007, 0.2052375227212906, -0.12679260969161987, 0.06525304168462753, 0.10045222193002701, -0.08265905827283859, -0.00025782460579648614, 0.11413740366697311, -0.007952986285090446, -0.055981528013944626, -0.08116789907217026, 0.021878303959965706, 0.28927910327911377, 0.015543431043624878, 0.10897347331047058, -0.08447521179914474, 0.038681428879499435, 0.03627035766839981, -0.12865884602069855, 0.058006517589092255, 0.03475360572338104, -0.014221473596990108, 0.0075475540943443775, 0.012444362975656986, -0.007058158982545137, -0.1274377852678299, 0.18709076941013336, -0.037306129932403564, -0.08694227039813995, 0.0023296603467315435, -0.013277915306389332, -0.041637834161520004, 0.012314511463046074, -0.26993289589881897, -0.05665811896324158, 0.06644756346940994, 0.06937560439109802, 0.0666748657822609, -0.12500834465026855, 0.00189721561037004, 0.012144173495471478, -0.11417713016271591, -0.17260701954364777, 0.07983312755823135, -0.03628704696893692, 0.01788974553346634, -0.11859502643346786, -0.09316409379243851, 0.010714155621826649, -0.042824093252420425, -0.14880822598934174, 0.10475838929414749, -0.01859801635146141, -0.32438644766807556, -0.14566099643707275, -0.009946607053279877, -0.011540163308382034, -0.03097572550177574, 0.09639892727136612, -0.10064876079559326, -0.023822324350476265, -0.007018607575446367, 0.0257895328104496, 0.04030579328536987, -0.018837174400687218, -0.0506151057779789, -0.02376127615571022, 0.07164541631937027, -0.14983101189136505, -0.011462918482720852, -0.02226782962679863, -0.050628628581762314, 0.055112216621637344, -0.03075452335178852, 0.014434395357966423, 0.21240536868572235, 0.01486619096249342, 0.018416840583086014, 0.038195401430130005, 0.181717649102211, -0.09637895971536636, -0.04247462749481201, 0.2431999146938324, -0.020475510507822037, -0.023062337189912796, 0.07326291501522064, 0.02229364775121212, -0.07771319150924683, -0.004659549333155155, -0.021072790026664734, -0.057942554354667664, -0.20929068326950073, -0.08972866833209991, -0.05996287986636162, -0.09231097251176834, -0.11238414794206619, -0.01903025433421135, 0.10190892219543457, 0.023557599633932114, 0.024117199704051018, -0.05417686328291893, 0.002020855201408267, -0.03252338990569115, 0.15357442200183868, -0.013307818211615086, 0.09766174107789993, -0.08216667175292969, -0.0533120296895504, -0.000267225899733603, 0.03411657735705376, 0.12713705003261566, 0.11260637640953064, 0.08496925234794617, 0.11839321255683899, 0.15336088836193085, 0.14831554889678955, 0.0686192661523819, -0.11324707418680191, -0.014909306541085243, -0.009338794276118279, -0.0638299360871315, -0.04864504560828209, 0.014472332783043385, 0.10010931640863419, -0.028518669307231903, -0.010837090201675892, -0.033499523997306824, 0.020452994853258133, 0.16268394887447357, 0.05407920107245445, -0.1697462946176529, -0.0748145654797554, -0.04246550798416138, -0.10427804291248322, 0.039399661123752594, 0.03623979538679123, 0.12750039994716644, -0.15201634168624878, 0.02716829814016819, 0.03919176012277603, 0.0947888046503067, -0.0770508423447609, 0.036370206624269485, -0.10886548459529877, 0.04935603216290474, 0.021254660561680794, 0.11166336387395859, -0.22358126938343048, 0.22839324176311493, -0.008775057271122932, 0.10723502188920975, -0.0563478097319603, -0.02408471144735813, 0.024111909791827202, 0.05847049877047539, 0.08948797732591629, 0.03569208085536957, 0.05445468798279762, -0.06924175471067429, -0.08513413369655609, 0.07086988538503647, -0.0021301221568137407, 0.1188666895031929, 0.01201615296304226, 0.01642482541501522, -0.015155087225139141, 0.013776581734418869, 0.019856801256537437, -0.10976991802453995, -0.016876094043254852, 0.030354280024766922, 0.20440274477005005, 0.13791002333164215, -0.018262607976794243, -0.08103898912668228, -0.08624932169914246, 0.06761296093463898, 0.031022043898701668, -0.06597594916820526, -0.03814704716205597, -0.13079425692558289, 0.12786352634429932, -0.07703866809606552, -0.04111892729997635, 0.07411038130521774, 0.06695036590099335, -0.010267937555909157, 0.000545032205991447, 0.04102021083235741, -0.07824898511171341, -0.09633617103099823, 0.010593902319669724, 0.15627846121788025, 0.07226511090993881, 0.07594984769821167, 0.07268397510051727, -0.059123676270246506, 0.022359171882271767, -0.09526057541370392, -0.03595314547419548, 0.11261553317308426, -0.17964287102222443, -0.033998604863882065, 0.07110842317342758, -0.1353462189435959, -0.11796131730079651, -0.0758824497461319, 0.13360127806663513, 0.02367221750319004, -0.0007993820472620428, 0.16219013929367065, 0.28640076518058777, -0.05097949132323265, -0.1470259726047516, -0.11072906851768494, 0.07806406170129776, 0.10158369690179825, -0.0657704696059227, -0.12100601941347122, 0.11585491895675659, 0.06769290566444397, -0.03930824622511864, -0.10380668938159943, -0.2789151966571808, -0.13094764947891235, 0.1855594366788864, -0.022614238783717155, 0.19205927848815918, -0.02792932465672493, -0.06630365550518036, -0.028903454542160034, -0.09655249118804932, -0.015244380570948124, -0.059750013053417206, 0.12253762781620026, 0.01750432513654232, 0.13196036219596863, 0.02284345217049122, 0.001053866813890636, 0.13539710640907288, 0.10375431925058365, -0.007821801118552685, 0.010529998689889908, 0.06415826082229614, -0.0038176702801138163, 0.06726814061403275, 0.19273649156093597, -0.09609406441450119, 0.011013569310307503, -0.06749316304922104, -0.104007788002491, -0.12221527844667435, 0.05626202002167702, 0.053980909287929535, -0.04427645355463028, 0.06661578267812729, -0.025932054966688156, 0.0019296688260510564, 0.02686142735183239, -0.032736677676439285, -0.13327693939208984, -0.027222391217947006, 0.12087716907262802, 0.17997021973133087, -0.10989529639482498, -0.09250456839799881, -0.09569050371646881, -0.025111600756645203, 0.1107509657740593, -0.06060230731964111, 0.01406877115368843, 0.022966280579566956, 0.03348619118332863, 0.15653812885284424, -0.010474788956344128, -0.1436873972415924, 0.12033439427614212, 0.053719352930784225, -0.0013928498374298215, -0.11720402538776398, -0.004816486034542322, -0.01498700212687254, 0.015788782387971878, 0.03495389595627785, 0.07427803426980972, -0.1164160892367363, -0.029913028702139854, -0.05074382945895195, 0.010497333481907845, -0.17958588898181915, 0.23736639320850372, 0.01835962012410164, 0.08107882738113403, -0.11579007655382156, 0.06644313037395477, 0.02132759615778923, -0.01637144386768341, 0.02358224242925644, -0.0014287931844592094, -0.0685885101556778, -0.051691751927137375, -0.05880632996559143, 0.08243051916360855, 0.040839094668626785, -0.18297480046749115, -0.039031095802783966, -0.08461243659257889, -0.04300693795084953, 0.06384193152189255, 0.048945628106594086, 0.033082835376262665, -0.10257784277200699, -0.07191312313079834, -0.07898814231157303, 0.006490207277238369, 0.12315043061971664, -0.01993435062468052, -0.10270409286022186, 0.11536730825901031, 0.12014693766832352, 0.07666381448507309, -0.05299313738942146, -0.11992467194795609, -0.006434171460568905, 0.07925499230623245, -0.1339689940214157, 0.007492627948522568, -0.11025019735097885, 0.0011806953698396683, 0.008482467383146286, -0.07679586112499237, -0.013629600405693054, 0.05720105394721031, -0.09097623080015182, 0.06701606512069702, 0.00941256619989872, 0.07045138627290726, -0.06817833334207535, 0.041489485651254654, 0.053212717175483704, -0.03486950322985649, 0.05847763270139694, 0.1524113118648529, -0.13181330263614655, 0.13296908140182495, -0.19989760220050812, -0.05089781433343887, 0.10707132518291473, 0.06678234785795212, 0.026173019781708717, -0.111958347260952, 0.04486750811338425, 0.13571876287460327, 0.058650191873311996, -0.014402671717107296, 0.06723013520240784, -0.08662668615579605, -0.019913824275135994, -0.10141713917255402, -0.009443231858313084, -0.0502975694835186, -0.008224991150200367, 0.10299943387508392, 0.12220844626426697, 0.14861294627189636, -0.10608818382024765, 0.09358052164316177, -0.10681159794330597, 0.011850125156342983, -0.07775436341762543, -0.061838071793317795, -0.06837695091962814, -0.10595016926527023, 0.04582328721880913, -0.04224653169512749, 0.14400741457939148, 0.0014159317361190915, 0.07289092242717743, -0.0732669085264206, -0.07923880964517593, -0.02480108104646206, -0.03144349530339241, 0.26695385575294495, 0.08558497577905655, 0.021681521087884903, -0.021041959524154663, -0.012090887874364853, -0.00559714762493968, 0.07413649559020996, -0.039172571152448654, 0.1798555999994278, 0.028678013011813164, 0.11422352492809296, 0.13026094436645508, -0.02540331892669201, -0.08001385629177094, -0.08427178859710693, -0.10086165368556976, 0.03190886974334717, -0.032490335404872894, 0.16292643547058105, 0.13505835831165314, -0.07013534754514694, 0.10220170766115189, 0.02060036174952984, -0.08450914919376373, -0.13265454769134521, -0.205713152885437, -0.06907738000154495, -0.16770383715629578, 0.009730021469295025, -0.08868315815925598, 0.013485785573720932, 0.05582602694630623, 0.032441288232803345, -0.06871045380830765, 0.17781797051429749, -0.019696572795510292, -0.0833815261721611, 0.10105692595243454, -0.10579478740692139, -0.0156637541949749, -0.08394330739974976, 0.05075753107666969, 0.1487368643283844, 0.05624083802103996, 0.06842677295207977, 0.010035171173512936, -0.08232836425304413, 0.0005917262751609087, -0.06330636143684387, -0.09879758954048157, -0.02883068658411503, -0.06570909917354584, 0.07746714353561401, 0.09842444956302643, 0.12113454192876816, -0.05867627263069153, 0.004724664147943258, 0.10920018702745438, -0.04594656080007553, -0.17934651672840118, -0.13615694642066956, 0.1741054654121399, 0.011325475759804249, -0.006345327012240887, 0.003443734021857381, -0.033494994044303894, -0.02024558000266552, 0.2839007079601288, 0.17562486231327057, 0.045415833592414856, 0.044313397258520126, -0.03666254132986069, -0.00411800155416131, -0.07468608021736145, 0.046513237059116364, 0.10540946573019028, 0.22257296741008759, 0.014217579737305641, 0.07286214828491211, -0.10141003876924515, -0.11234456300735474, 0.02788720093667507, 0.031164085492491722, -0.02477812021970749, -0.09994710981845856, 0.00902047473937273, 0.14533357322216034, -0.1530851125717163, -0.0872546136379242, -0.12319456040859222, -0.018002143129706383, -0.10960555076599121, -0.03413702920079231, 0.00236299610696733, 0.1581382155418396, 0.012437044642865658, -0.0641694888472557, 0.032004356384277344, 0.16819533705711365, 0.02013520896434784, -0.045525964349508286, -0.06813060492277145, 0.051178399473428726, -0.041846465319395065, -0.02618664689362049, 0.04437031224370003, 0.13085432350635529, -0.010263584554195404, 0.13858656585216522, -0.006492546759545803, 0.27051350474357605, -0.02266199141740799, -0.04433056339621544, 0.039204757660627365, 0.14731213450431824, -0.005007206927984953, 0.09155203402042389, 0.0049376231618225574, -0.1527457982301712, 0.03479352593421936, -0.1013936921954155, 0.0075834705494344234, -0.08869242668151855, 0.0031680441461503506, -0.015779461711645126, 0.055733177810907364, 0.09956978261470795, -0.07141651958227158, -0.05058762803673744, -0.05753835663199425, 0.10960307717323303, 0.01882547326385975, -0.09994561225175858, -0.08686438202857971, -0.1910272091627121, -0.03717918321490288, -0.09646967053413391, -0.0024724705144762993, -0.15243209898471832, -0.0005590239306911826, -0.014877364039421082, -0.08282090723514557, 0.0473492331802845, -0.007514612749218941, 0.06516895443201065, 0.061881255358457565, 0.013903923332691193, -0.002311636693775654, 0.050433382391929626, 0.11651892215013504, -0.18751303851604462, -0.10682309418916702 ]
null
null
transformers
# Wav2Vec2-Large-XLSR-53-Indonesia Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) in Indonesia using the [Common Voice](https://huggingface.co/datasets/common_voice) When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "id", split="test[:2%]"). processor = Wav2Vec2Processor.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model = Wav2Vec2ForCTC.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("common_voice", "id", split="test") wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model = Wav2Vec2ForCTC.from_pretrained("ayameRushia/wav2vec2-large-xlsr-indonesia-demo") model.to("cuda") chars_to_ignore_regex = '[\,\?\.\!\-\;\:\"\โ€œ]' resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() speech_array, sampling_rate = torchaudio.load(batch["path"]) resampler = torchaudio.transforms.Resample(sampling_rate, 16_000) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**: WER = 19.830319 % ## Training Training using common voice dataset
{"language": "id", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "model-index": [{"name": "XLSR Wav2Vec2 Indonesia by Ayame Rushia", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice id", "type": "common_voice", "args": "id"}, "metrics": [{"type": "wer", "value": 19.830319, "name": "Test WER"}]}]}]}
automatic-speech-recognition
ayameRushia/wav2vec2-large-xlsr-indonesia
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "id", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "id" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-Indonesia Fine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the {language} test data of Common Voice. Test Result: WER = 19.830319 % ## Training Training using common voice dataset
[ "# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\nThe model can be used directly (without a language model) as follows:", "## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 19.830319 %", "## Training\nTraining using common voice dataset" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\nThe model can be used directly (without a language model) as follows:", "## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 19.830319 %", "## Training\nTraining using common voice dataset" ]
[ 80, 60, 20, 34, 8 ]
[ "passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #id #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Indonesia\nFine-tuned facebook/wav2vec2-large-xlsr-53 in Indonesia using the Common Voice\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\nThe model can be used directly (without a language model) as follows:## Evaluation\nThe model can be evaluated as follows on the {language} test data of Common Voice.\n\nTest Result: \nWER = 19.830319 %## Training\nTraining using common voice dataset" ]
[ -0.11943415552377701, 0.007512805983424187, -0.0013085979735478759, -0.07939015328884125, 0.06721214205026627, -0.05767597630620003, 0.1542372852563858, 0.0848492830991745, -0.01803058758378029, -0.04575289785861969, 0.003714438062161207, 0.02349478006362915, 0.0845046192407608, 0.06816030293703079, -0.019583936780691147, -0.27028581500053406, 0.06175879389047623, 0.004670352675020695, 0.17390385270118713, 0.15219484269618988, 0.09209112823009491, -0.050700899213552475, 0.014274746179580688, 0.14108508825302124, -0.08859241753816605, 0.0016650577308610082, -0.006840987596660852, -0.15245650708675385, 0.1244242936372757, 0.04574022814631462, 0.07093901187181473, 0.07853053510189056, 0.06330377608537674, -0.15244480967521667, 0.02021864242851734, -0.009119255468249321, 0.04710850864648819, 0.0016617759829387069, 0.034190285950899124, 0.024687139317393303, 0.10808288305997849, 0.12739235162734985, -0.056869711726903915, 0.0356164276599884, -0.0665818601846695, -0.2212865948677063, -0.03440794348716736, 0.039174944162368774, 0.13663658499717712, 0.12039822340011597, -0.07534611225128174, 0.0933985635638237, -0.14888280630111694, 0.049237966537475586, 0.07211240381002426, -0.1976521760225296, 0.008656902238726616, 0.03456037491559982, 0.0635162964463234, 0.07208706438541412, -0.044265665113925934, 0.000199501242605038, 0.022994333878159523, 0.0394570417702198, 0.002124568447470665, -0.09354052692651749, -0.18728125095367432, -0.06570419669151306, -0.10447976738214493, -0.0043412307277321815, 0.2422858029603958, 0.007077709771692753, -0.09604751318693161, -0.10113774240016937, 0.02209407463669777, 0.005221610888838768, -0.031098151579499245, -0.09871409088373184, -0.008136828429996967, 0.03907236456871033, 0.04253525659441948, -0.01423849631100893, -0.10215545445680618, -0.11685866117477417, -0.026387542486190796, 0.09944631159305573, 0.018135959282517433, 0.018274959176778793, -0.12020310759544373, 0.04493771493434906, -0.12191175669431686, -0.08931136131286621, -0.0351216085255146, 0.029736846685409546, -0.04724632203578949, 0.06352954357862473, -0.05987175926566124, -0.16049808263778687, 0.05395042523741722, -0.07373861968517303, 0.022956714034080505, 0.022186432033777237, 0.02368037961423397, 0.07058961689472198, 0.024107974022626877, 0.1492924690246582, -0.08873416483402252, -0.0650472491979599, 0.018422875553369522, 0.016717685386538506, -0.007709043566137552, -0.03278500959277153, -0.04981345310807228, -0.0728689506649971, 0.047127481549978256, 0.05370194464921951, -0.08722513914108276, 0.022549346089363098, -0.03798718750476837, -0.059007059782743454, 0.07681230455636978, -0.10554257035255432, -0.07066617906093597, 0.03287172317504883, 0.011219891719520092, 0.19581463932991028, 0.07044968008995056, 0.04836646839976311, -0.05877671018242836, 0.000047806348447920755, 0.03535451740026474, 0.040514025837183, -0.01332076359540224, -0.05585947632789612, -0.004714180249720812, -0.0032437723129987717, -0.017406024038791656, -0.1054011732339859, -0.10837510973215103, -0.07959365099668503, -0.019335607066750526, 0.0183684304356575, -0.027892567217350006, -0.0912570133805275, 0.0003005577891599387, -0.034532032907009125, -0.09834543615579605, 0.02910519577562809, -0.05002486705780029, 0.06747744232416153, 0.08510436862707138, 0.08522219955921173, 0.0007042840588837862, 0.09935197979211807, -0.08464061468839645, -0.00011300348705844954, -0.037452783435583115, 0.11447197943925858, -0.033363454043865204, -0.05403929203748703, -0.10561856627464294, -0.10287243872880936, -0.07742783427238464, 0.05523289740085602, 0.0324750579893589, 0.06725506484508514, -0.221244215965271, -0.10106869786977768, 0.1920681893825531, -0.15246188640594482, -0.05535377934575081, 0.18650133907794952, 0.019229521974921227, 0.15745559334754944, 0.08922994136810303, 0.21667799353599548, 0.11332013458013535, -0.13084730505943298, 0.07148215919733047, 0.04038556292653084, -0.02106105163693428, -0.07208091020584106, 0.06832820177078247, -0.03870846703648567, -0.014305750839412212, 0.06050548329949379, -0.054502565413713455, 0.09227408468723297, -0.023817023262381554, -0.06940554827451706, 0.011463156901299953, -0.10452452301979065, 0.06140422075986862, -0.0003329860046505928, 0.0520612858235836, -0.0193401500582695, -0.02170095220208168, 0.07726654410362244, 0.14812812209129333, -0.1270737498998642, 0.017535535618662834, -0.1341642588376999, 0.06122848764061928, -0.1411971002817154, 0.010866673663258553, -0.12008745223283768, 0.1941894292831421, -0.017482347786426544, 0.009367682039737701, 0.08298410475254059, 0.14069534838199615, 0.007328834384679794, -0.036242883652448654, -0.018289456143975258, -0.04494035989046097, 0.05265440791845322, -0.0035654606763273478, 0.023263493552803993, -0.01527394074946642, -0.030416641384363174, -0.025689544156193733, 0.07916386425495148, -0.09819838404655457, -0.04294428601861, -0.04974990338087082, -0.03372124955058098, -0.045584097504615784, 0.01912808045744896, 0.11421174556016922, 0.10921011865139008, 0.04602769389748573, 0.040450457483530045, 0.03452768921852112, 0.025577116757631302, -0.0951559916138649, 0.2077718824148178, -0.12701934576034546, 0.0636424571275711, 0.09971098601818085, -0.08498866856098175, -0.0015448047779500484, 0.11562506854534149, -0.007838748395442963, -0.05606419965624809, -0.08103851974010468, 0.022603793069720268, 0.2940884232521057, 0.015276716090738773, 0.10886288434267044, -0.0855289101600647, 0.0385405533015728, 0.03668026626110077, -0.12896372377872467, 0.057722657918930054, 0.03543740510940552, -0.013635345734655857, 0.01007404737174511, 0.012666130438446999, -0.007044664584100246, -0.12680970132350922, 0.18506671488285065, -0.03650674223899841, -0.08668600767850876, 0.0025965275708585978, -0.01422822754830122, -0.04179786145687103, 0.012858526781201363, -0.26875925064086914, -0.05629269406199455, 0.0663943663239479, 0.06982041895389557, 0.06679636985063553, -0.12382768094539642, 0.0011863631661981344, 0.013275811448693275, -0.11388307809829712, -0.17156928777694702, 0.07903873175382614, -0.035129278898239136, 0.0184407290071249, -0.11854957789182663, -0.09421960264444351, 0.010274909436702728, -0.04286675155162811, -0.149420827627182, 0.1050698459148407, -0.01847541704773903, -0.3254888653755188, -0.14637291431427002, -0.010313237085938454, -0.013167151249945164, -0.03099053166806698, 0.09659802168607712, -0.10187181085348129, -0.023926015943288803, -0.006922903936356306, 0.027314119040966034, 0.040189918130636215, -0.020180094987154007, -0.0507887527346611, -0.025058632716536522, 0.07208166271448135, -0.1496700495481491, -0.011476978659629822, -0.02244686521589756, -0.050065431743860245, 0.05557186156511307, -0.0302625373005867, 0.01568686030805111, 0.21308840811252594, 0.014400126412510872, 0.017045432701706886, 0.037714872509241104, 0.18248020112514496, -0.09724227339029312, -0.04303986579179764, 0.24247074127197266, -0.022472906857728958, -0.023914465680718422, 0.07227343320846558, 0.022387467324733734, -0.07750104367733002, -0.00433498527854681, -0.019139079377055168, -0.05887189880013466, -0.21143372356891632, -0.0901474729180336, -0.059027839452028275, -0.09247774630784988, -0.11310594528913498, -0.01936258189380169, 0.10382339358329773, 0.024634949862957, 0.023447440937161446, -0.05665246769785881, 0.002493936801329255, -0.03235615789890289, 0.15506376326084137, -0.013140820898115635, 0.09898676723241806, -0.0806768611073494, -0.052946995943784714, -0.00005146750481799245, 0.0345952995121479, 0.12443262338638306, 0.11051123589277267, 0.08327309042215347, 0.1192437931895256, 0.15324565768241882, 0.14789006114006042, 0.06850937008857727, -0.11457877606153488, -0.014297806657850742, -0.0096527598798275, -0.0636100172996521, -0.05050438642501831, 0.012800531461834908, 0.09868477284908295, -0.0282429251819849, -0.011667334474623203, -0.03461439162492752, 0.020439056679606438, 0.1629808098077774, 0.05218956992030144, -0.17000728845596313, -0.07475344091653824, -0.04289168491959572, -0.10262829065322876, 0.03991243988275528, 0.037893958389759064, 0.12283205986022949, -0.15250298380851746, 0.027724556624889374, 0.039654750376939774, 0.09598211199045181, -0.07659677416086197, 0.03601629287004471, -0.10806795209646225, 0.04828828573226929, 0.01970571093261242, 0.1114496961236, -0.2224113941192627, 0.22876636683940887, -0.008243465796113014, 0.10708961635828018, -0.055200591683387756, -0.023833023384213448, 0.025070583447813988, 0.05800129473209381, 0.09022699296474457, 0.035688772797584534, 0.05543547123670578, -0.07168741524219513, -0.08668476343154907, 0.07162453979253769, -0.004596555605530739, 0.11756150424480438, 0.011799578554928303, 0.01698402687907219, -0.01499099563807249, 0.01380858477205038, 0.019034048542380333, -0.10885085165500641, -0.01557922549545765, 0.030057178810238838, 0.20212742686271667, 0.13794399797916412, -0.01783192716538906, -0.08144410699605942, -0.08792637288570404, 0.067003034055233, 0.032896291464567184, -0.06746481359004974, -0.03802098333835602, -0.1351298838853836, 0.12945304811000824, -0.07716590166091919, -0.0417315848171711, 0.072988361120224, 0.06448029726743698, -0.010246798396110535, -0.0007004960207268596, 0.040983833372592926, -0.07937907427549362, -0.09616020321846008, 0.010558413341641426, 0.15723903477191925, 0.07218149304389954, 0.07639782875776291, 0.0722147598862648, -0.058933768421411514, 0.02298806421458721, -0.09520035982131958, -0.03508184105157852, 0.11095833033323288, -0.17786024510860443, -0.031338922679424286, 0.07110319286584854, -0.13579091429710388, -0.12054243683815002, -0.074910007417202, 0.13240155577659607, 0.024249760434031487, -0.000700380012858659, 0.16225726902484894, 0.2883891463279724, -0.05211937800049782, -0.14580287039279938, -0.11095023900270462, 0.07759586721658707, 0.10092959553003311, -0.0641365721821785, -0.12096967548131943, 0.11457637697458267, 0.06934578716754913, -0.0393294095993042, -0.1063733920454979, -0.2780759632587433, -0.13147173821926117, 0.187249094247818, -0.022727996110916138, 0.19056978821754456, -0.028102824464440346, -0.06517377495765686, -0.02610093727707863, -0.09757070988416672, -0.01658180169761181, -0.05497455969452858, 0.12193269282579422, 0.017559340223670006, 0.13155902922153473, 0.02347090095281601, 0.000926822132896632, 0.13576063513755798, 0.10357080399990082, -0.007195457816123962, 0.009808480739593506, 0.06397128850221634, -0.006773042492568493, 0.06677626818418503, 0.19267858564853668, -0.0949321910738945, 0.01216354314237833, -0.0700075700879097, -0.10472167283296585, -0.12318011373281479, 0.056141916662454605, 0.05369410291314125, -0.045129843056201935, 0.06570079177618027, -0.02507481724023819, 0.00231498503126204, 0.026371072977781296, -0.03306702896952629, -0.13354898989200592, -0.0241171233355999, 0.12059665471315384, 0.18106287717819214, -0.11019349098205566, -0.09517352283000946, -0.09505551308393478, -0.02548389323055744, 0.11079175025224686, -0.05892931669950485, 0.014703977853059769, 0.023955892771482468, 0.032695818692445755, 0.15719541907310486, -0.01035338081419468, -0.14377188682556152, 0.12067493051290512, 0.05275887995958328, -0.0013076727045699954, -0.11457697302103043, -0.004992605186998844, -0.018229853361845016, 0.01621616818010807, 0.03348942846059799, 0.07482738792896271, -0.11722084879875183, -0.030759409070014954, -0.050483718514442444, 0.011747735552489758, -0.1790306717157364, 0.23664377629756927, 0.019115369766950607, 0.08126787096261978, -0.11576948314905167, 0.06712121516466141, 0.021464381366968155, -0.018146736547350883, 0.02418757602572441, -0.0015824552392587066, -0.0674641951918602, -0.051788344979286194, -0.05982125550508499, 0.08431366831064224, 0.038028325885534286, -0.1826031357049942, -0.03958728164434433, -0.08510114252567291, -0.04268461465835571, 0.06516096740961075, 0.048978496342897415, 0.03252099081873894, -0.10113368183374405, -0.07200859487056732, -0.07943037152290344, 0.007732865400612354, 0.12297099828720093, -0.020006973296403885, -0.1017577052116394, 0.11299514025449753, 0.12112174183130264, 0.0779237449169159, -0.0531245619058609, -0.12012510746717453, -0.006334100384265184, 0.08042676001787186, -0.13579736649990082, 0.007402041926980019, -0.11066750437021255, 0.001959532266482711, 0.008062344044446945, -0.07751736789941788, -0.013861134648323059, 0.056924860924482346, -0.09072700142860413, 0.06712258607149124, 0.009650961495935917, 0.07165983319282532, -0.07011637836694717, 0.04066073149442673, 0.05384644865989685, -0.03400162234902382, 0.05894658342003822, 0.15276199579238892, -0.1316787153482437, 0.13175244629383087, -0.19989818334579468, -0.050124041736125946, 0.10679497569799423, 0.06645308434963226, 0.02415238693356514, -0.11407845467329025, 0.045344214886426926, 0.13476799428462982, 0.05663906782865524, -0.014067905023694038, 0.06771344691514969, -0.08798964321613312, -0.019243726506829262, -0.09868071228265762, -0.009823145344853401, -0.0505029670894146, -0.007956692948937416, 0.103035569190979, 0.1235981434583664, 0.1475878208875656, -0.10545921325683594, 0.09496022015810013, -0.10749926418066025, 0.011518007144331932, -0.07801807671785355, -0.060947220772504807, -0.06846585124731064, -0.10545748472213745, 0.0446685291826725, -0.04233565554022789, 0.14406678080558777, 0.003125022863969207, 0.07222659885883331, -0.0723903477191925, -0.07896462827920914, -0.023304972797632217, -0.031166724860668182, 0.2634623646736145, 0.0832299292087555, 0.021360313519835472, -0.019950605928897858, -0.011847367510199547, -0.0059204744175076485, 0.0734068751335144, -0.03903845325112343, 0.1813574731349945, 0.03260619193315506, 0.11349962651729584, 0.13058853149414062, -0.026738623157143593, -0.08051124960184097, -0.08773212134838104, -0.10037051886320114, 0.031833697110414505, -0.030461877584457397, 0.16165226697921753, 0.13451196253299713, -0.0698777288198471, 0.10425523668527603, 0.019657636061310768, -0.08444438129663467, -0.13241367042064667, -0.2033379077911377, -0.06793797761201859, -0.1667378693819046, 0.009225773625075817, -0.08967068046331406, 0.013250559568405151, 0.05830621346831322, 0.032728612422943115, -0.0694253072142601, 0.17898154258728027, -0.01984618790447712, -0.08339818567037582, 0.10093050450086594, -0.10574568808078766, -0.014520462602376938, -0.08453913778066635, 0.0508587546646595, 0.14886631071567535, 0.057115428149700165, 0.06794769316911697, 0.010177096351981163, -0.08196863532066345, 0.0009965004865080118, -0.06354527920484543, -0.10004789382219315, -0.028862114995718002, -0.06642439216375351, 0.07893246412277222, 0.10076672583818436, 0.1207738071680069, -0.0567210428416729, 0.0054570832289755344, 0.11124919354915619, -0.044270455837249756, -0.17611664533615112, -0.13630886375904083, 0.1750698685646057, 0.009803280234336853, -0.007756194565445185, 0.003658304689452052, -0.032646019011735916, -0.020193172618746758, 0.2846928536891937, 0.176587775349617, 0.04518307372927666, 0.04324354976415634, -0.03804565593600273, -0.003826987696811557, -0.07269903272390366, 0.04629511758685112, 0.10455545783042908, 0.22185947000980377, 0.012924393638968468, 0.07375653088092804, -0.10148777812719345, -0.11186649650335312, 0.029868129640817642, 0.02987806126475334, -0.024429291486740112, -0.0998140424489975, 0.008703512139618397, 0.14644929766654968, -0.15497267246246338, -0.0871327742934227, -0.12348563969135284, -0.020455917343497276, -0.11072442680597305, -0.03468433767557144, 0.0026660102885216475, 0.15719392895698547, 0.012983365915715694, -0.06355686485767365, 0.03090311586856842, 0.16803865134716034, 0.020288608968257904, -0.044380106031894684, -0.07029230147600174, 0.050711046904325485, -0.03663279488682747, -0.023846013471484184, 0.0451558493077755, 0.12981188297271729, -0.010464917868375778, 0.13920530676841736, -0.007265379652380943, 0.269185870885849, -0.022060053423047066, -0.043903250247240067, 0.03968833014369011, 0.14701175689697266, -0.004128157626837492, 0.09221285581588745, 0.00543260620906949, -0.1534397006034851, 0.03502007946372032, -0.10276943445205688, 0.007578008342534304, -0.0877387598156929, 0.0027757601346820593, -0.01591758243739605, 0.05554899200797081, 0.10005373507738113, -0.07039957493543625, -0.0519661121070385, -0.05777149647474289, 0.10824154317378998, 0.01807621493935585, -0.09873130917549133, -0.0869942381978035, -0.19139312207698822, -0.03619610518217087, -0.09526853263378143, -0.0033977916464209557, -0.1544276922941208, -0.000004935457127430709, -0.01344505324959755, -0.08334192633628845, 0.04783782735466957, -0.007282114587724209, 0.063158318400383, 0.062330711632966995, 0.014554598368704319, -0.00010783439211081713, 0.05049660801887512, 0.11672420799732208, -0.18760646879673004, -0.10630457103252411 ]
null
null
transformers
# `false-positives-scancode-bert-base-uncased-L8-1` ## Intended Use This model is intended to be used for Sentence Classification which is used for results analysis in [`scancode-results-analyzer`](https://github.com/nexB/scancode-results-analyzer). `scancode-results-analyzer` helps detect faulty scans in [`scancode-toolkit`](https://github.com/nexB/scancode-results-analyzer) by using statistics and nlp modeling, among other tools, to make Scancode better. #### How to use Refer [quickstart](https://github.com/nexB/scancode-results-analyzer#quickstart---local-machine) section in `scancode-results-analyzer` documentation, for installing and getting started. - [Link to Code](https://github.com/nexB/scancode-results-analyzer/blob/master/src/results_analyze/nlp_models.py) Then in `NLPModelsPredict` class, function `predict_basic_false_positive` uses this classifier to predict sentances as either valid license tags or false positives. #### Limitations and bias As this model is a fine-tuned version of the [`bert-base-uncased`](https://huggingface.co/bert-base-uncased) model, it has the same biases, but as the task it is fine-tuned to is a very specific field (license tags vs false positives) without those intended biases, it's safe to assume those don't apply at all here. ## Training and Fine-Tuning Data The BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers). Then this `bert-base-uncased` model was fine-tuned on Scancode Rule texts, specifically trained in the context of sentence classification, where the two classes are - License Tags - False Positives of License Tags ## Training procedure For fine-tuning procedure and training, refer `scancode-results-analyzer` code. - [Link to Code](https://github.com/nexB/scancode-results-analyzer/blob/master/src/results_analyze/nlp_models.py) In `NLPModelsTrain` class, function `prepare_input_data_false_positive` prepares the training data. In `NLPModelsTrain` class, function `train_basic_false_positive_classifier` fine-tunes this classifier. 1. Model - [BertBaseUncased](https://huggingface.co/bert-base-uncased) (Weights 0.5 GB) 2. Sentence Length - 8 3. Labels - 2 (False Positive/License Tag) 4. After 4-6 Epochs of Fine-Tuning with learning rate 2e-5 (6 secs each on an RTX 2060) Note: The classes aren't balanced. ## Eval results - Accuracy on the training data (90%) : 0.99 (+- 0.005) - Accuracy on the validation data (10%) : 0.96 (+- 0.015) The errors have lower confidence scores using thresholds on confidence scores almost makes it a perfect classifier as the classification task is comparatively easier. Results are stable, in the sence fine-tuning accuracy is very easily achieved every time, though more learning epochs makes the data overfit, i.e. the training loss decreases, but the validation loss increases, even though accuracies are very stable even on overfitting.
{"language": "en", "license": "apache-2.0", "tags": ["license", "sentence-classification", "scancode", "license-compliance"], "datasets": ["bookcorpus", "wikipedia", "scancode-rules"], "version": 1.0}
fill-mask
ayansinha/false-positives-scancode-bert-base-uncased-L8-1
[ "transformers", "tf", "bert", "fill-mask", "license", "sentence-classification", "scancode", "license-compliance", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:scancode-rules", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# 'false-positives-scancode-bert-base-uncased-L8-1' ## Intended Use This model is intended to be used for Sentence Classification which is used for results analysis in 'scancode-results-analyzer'. 'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools, to make Scancode better. #### How to use Refer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started. - Link to Code Then in 'NLPModelsPredict' class, function 'predict_basic_false_positive' uses this classifier to predict sentances as either valid license tags or false positives. #### Limitations and bias As this model is a fine-tuned version of the 'bert-base-uncased' model, it has the same biases, but as the task it is fine-tuned to is a very specific field (license tags vs false positives) without those intended biases, it's safe to assume those don't apply at all here. ## Training and Fine-Tuning Data The BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers). Then this 'bert-base-uncased' model was fine-tuned on Scancode Rule texts, specifically trained in the context of sentence classification, where the two classes are - License Tags - False Positives of License Tags ## Training procedure For fine-tuning procedure and training, refer 'scancode-results-analyzer' code. - Link to Code In 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the training data. In 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes this classifier. 1. Model - BertBaseUncased (Weights 0.5 GB) 2. Sentence Length - 8 3. Labels - 2 (False Positive/License Tag) 4. After 4-6 Epochs of Fine-Tuning with learning rate 2e-5 (6 secs each on an RTX 2060) Note: The classes aren't balanced. ## Eval results - Accuracy on the training data (90%) : 0.99 (+- 0.005) - Accuracy on the validation data (10%) : 0.96 (+- 0.015) The errors have lower confidence scores using thresholds on confidence scores almost makes it a perfect classifier as the classification task is comparatively easier. Results are stable, in the sence fine-tuning accuracy is very easily achieved every time, though more learning epochs makes the data overfit, i.e. the training loss decreases, but the validation loss increases, even though accuracies are very stable even on overfitting.
[ "# 'false-positives-scancode-bert-base-uncased-L8-1'", "## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.", "#### How to use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_false_positive' uses this classifier to\npredict sentances as either valid license tags or false positives.", "#### Limitations and bias\n\nAs this model is a fine-tuned version of the 'bert-base-uncased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific field\n(license tags vs false positives) without those intended biases, it's safe to assume\nthose don't apply at all here.", "## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-uncased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the two classes are\n\n\t- License Tags \n\t- False Positives of License Tags", "## Training procedure\n\nFor fine-tuning procedure and training, refer 'scancode-results-analyzer' code.\n\n- Link to Code\n\nIn 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the\ntraining data.\n\nIn 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes\nthis classifier.\n\n1. Model - BertBaseUncased (Weights 0.5 GB)\n2. Sentence Length - 8\n3. Labels - 2 (False Positive/License Tag)\n4. After 4-6 Epochs of Fine-Tuning with learning rate 2e-5 (6 secs each on an RTX 2060)\n\nNote: The classes aren't balanced.", "## Eval results\n\n- Accuracy on the training data (90%) : 0.99 (+- 0.005) \n- Accuracy on the validation data (10%) : 0.96 (+- 0.015)\n\nThe errors have lower confidence scores using thresholds on confidence scores almost\nmakes it a perfect classifier as the classification task is comparatively easier.\n\nResults are stable, in the sence fine-tuning accuracy is very easily achieved every\ntime, though more learning epochs makes the data overfit, i.e. the training loss \ndecreases, but the validation loss increases, even though accuracies are very stable\neven on overfitting." ]
[ "TAGS\n#transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# 'false-positives-scancode-bert-base-uncased-L8-1'", "## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.", "#### How to use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_false_positive' uses this classifier to\npredict sentances as either valid license tags or false positives.", "#### Limitations and bias\n\nAs this model is a fine-tuned version of the 'bert-base-uncased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific field\n(license tags vs false positives) without those intended biases, it's safe to assume\nthose don't apply at all here.", "## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-uncased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the two classes are\n\n\t- License Tags \n\t- False Positives of License Tags", "## Training procedure\n\nFor fine-tuning procedure and training, refer 'scancode-results-analyzer' code.\n\n- Link to Code\n\nIn 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the\ntraining data.\n\nIn 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes\nthis classifier.\n\n1. Model - BertBaseUncased (Weights 0.5 GB)\n2. Sentence Length - 8\n3. Labels - 2 (False Positive/License Tag)\n4. After 4-6 Epochs of Fine-Tuning with learning rate 2e-5 (6 secs each on an RTX 2060)\n\nNote: The classes aren't balanced.", "## Eval results\n\n- Accuracy on the training data (90%) : 0.99 (+- 0.005) \n- Accuracy on the validation data (10%) : 0.96 (+- 0.015)\n\nThe errors have lower confidence scores using thresholds on confidence scores almost\nmakes it a perfect classifier as the classification task is comparatively easier.\n\nResults are stable, in the sence fine-tuning accuracy is very easily achieved every\ntime, though more learning epochs makes the data overfit, i.e. the training loss \ndecreases, but the validation loss increases, even though accuracies are very stable\neven on overfitting." ]
[ 86, 24, 86, 81, 85, 102, 175, 144 ]
[ "passage: TAGS\n#transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# 'false-positives-scancode-bert-base-uncased-L8-1'## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.#### How to use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_false_positive' uses this classifier to\npredict sentances as either valid license tags or false positives.#### Limitations and bias\n\nAs this model is a fine-tuned version of the 'bert-base-uncased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific field\n(license tags vs false positives) without those intended biases, it's safe to assume\nthose don't apply at all here.## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-uncased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the two classes are\n\n\t- License Tags \n\t- False Positives of License Tags" ]
[ -0.04031326249241829, 0.13766920566558838, -0.006441679783165455, 0.053930316120386124, 0.0382624976336956, -0.014384493231773376, 0.06086663901805878, 0.09505443274974823, 0.011709480546414852, 0.07558996230363846, -0.01341312937438488, 0.053697358816862106, 0.08797045052051544, 0.052468325942754745, -0.004049098584800959, -0.15762414038181305, 0.045229848474264145, -0.04340432211756706, 0.1012033000588417, 0.077041856944561, 0.12379249930381775, -0.04752998426556587, 0.05568401888012886, 0.04806157201528549, -0.05417029187083244, -0.0010619147215038538, 0.030769506469368935, -0.01326095499098301, 0.045783881098032, 0.056746531277894974, 0.07679847627878189, -0.003091794904321432, 0.02404400147497654, -0.20990703999996185, 0.009489279240369797, 0.0809701457619667, 0.012540210969746113, 0.03150084614753723, 0.1100751981139183, -0.01768052764236927, 0.12191619724035263, -0.05515339598059654, 0.07441035658121109, 0.08625948429107666, -0.11239561438560486, -0.17178405821323395, -0.10093794763088226, 0.11083880811929703, 0.017990922555327415, 0.06598208844661713, -0.03908706083893776, 0.05881673842668533, -0.023333001881837845, 0.05382971093058586, 0.0678129494190216, -0.25165075063705444, -0.027385594323277473, 0.013129398226737976, 0.03551344946026802, -0.0016688862815499306, -0.09267570078372955, -0.011053155176341534, -0.010694324970245361, 0.016708780080080032, 0.033922936767339706, -0.028981070965528488, 0.0022440003231167793, -0.01919342763721943, -0.12266058474779129, -0.06748837232589722, 0.05603456869721413, -0.006745133548974991, -0.08960821479558945, -0.20705941319465637, -0.056331511586904526, 0.004917428828775883, 0.019279269501566887, -0.03372754156589508, -0.0009791775373741984, 0.0528857484459877, 0.07521214336156845, -0.11797253042459488, -0.0894542708992958, 0.009724574163556099, -0.09675078839063644, 0.0996168926358223, 0.016368882730603218, 0.022653408348560333, 0.05553867295384407, 0.12907204031944275, 0.019767701625823975, -0.09743793308734894, -0.04751238599419594, -0.03289032354950905, -0.11234330385923386, -0.05062182992696762, -0.032666511833667755, -0.025834035128355026, -0.000007282038950506831, 0.20077122747898102, -0.06370559334754944, 0.046822529286146164, -0.10885429382324219, 0.0005268189706839621, 0.016180580481886864, 0.11344239115715027, 0.012529920786619186, 0.017727412283420563, 0.013091286644339561, 0.0035353614948689938, 0.04126223176717758, -0.025478297844529152, -0.019449757412075996, -0.06379639357328415, 0.04905840754508972, 0.0727546438574791, 0.017385344952344894, 0.038660697638988495, -0.10459960252046585, -0.027128588408231735, 0.08828607201576233, -0.15962232649326324, 0.04564424976706505, 0.07578149437904358, -0.03159051388502121, -0.020579691976308823, 0.04067307338118553, 0.013627596199512482, -0.08273305743932724, 0.04646637290716171, -0.048576634377241135, -0.005823610816150904, -0.08045860379934311, -0.07048385590314865, 0.041354238986968994, -0.04562712088227272, -0.0881333276629448, -0.07214662432670593, -0.17662246525287628, -0.021846121177077293, 0.045716091990470886, -0.05171136185526848, 0.02237481251358986, 0.0073285275138914585, -0.018076319247484207, -0.008710168302059174, 0.0026089586317539215, -0.011632979847490788, -0.015301127918064594, 0.051167119294404984, -0.026347259059548378, 0.01113602053374052, 0.010745582170784473, 0.01853625848889351, -0.1279989778995514, 0.03599834442138672, -0.14657148718833923, 0.15653365850448608, -0.07200801372528076, -0.0507064089179039, -0.1154000461101532, 0.0070365844294428825, -0.010418025776743889, 0.0047734444960951805, 0.011845745146274567, 0.09352530539035797, -0.17961488664150238, 0.009472903795540333, 0.1235973909497261, -0.1396520435810089, -0.07887256890535355, 0.09214016795158386, -0.06042896956205368, 0.12276458740234375, 0.09599296003580093, 0.10521379858255386, 0.08994947373867035, -0.1577148586511612, -0.032477572560310364, 0.010503406636416912, -0.007776040118187666, 0.17715796828269958, 0.0562346875667572, -0.07404300570487976, -0.05827736482024193, -0.0014532431960105896, -0.0991319864988327, -0.027300773188471794, 0.018803870305418968, -0.05416589975357056, -0.008916021324694157, -0.009836405515670776, 0.035105906426906586, -0.014540544711053371, -0.07630562037229538, -0.024121005088090897, -0.12652871012687683, 0.027779823169112206, 0.04847485199570656, -0.05046697333455086, 0.03327203541994095, -0.14242137968540192, 0.04884318634867668, 0.01802227832376957, -0.03645087033510208, -0.17333419620990753, -0.148153617978096, 0.05581384152173996, -0.15636588633060455, 0.08350152522325516, -0.029043124988675117, 0.010882105678319931, 0.015434637665748596, -0.04015621915459633, 0.00034018210135400295, -0.011413042433559895, -0.004888626281172037, -0.023679275065660477, -0.09622422605752945, -0.01969802752137184, -0.004913245793431997, 0.1984674334526062, -0.16145087778568268, 0.022570019587874413, 0.1474987417459488, 0.11670449376106262, 0.03785083070397377, -0.05804511532187462, -0.013935297727584839, -0.020106321200728416, 0.009489894844591618, -0.04738893359899521, 0.03696649894118309, 0.06265483051538467, -0.08750896155834198, 0.10399179905653, -0.16611146926879883, -0.14852015674114227, 0.07306890189647675, 0.06167575716972351, -0.12353990226984024, -0.04895782470703125, -0.014785107225179672, -0.010766981169581413, -0.06986004114151001, -0.07665592432022095, 0.10433051735162735, 0.053165238350629807, 0.06604031473398209, -0.10099230706691742, -0.08102522790431976, -0.03126465156674385, -0.03444366902112961, -0.027596615254878998, 0.06914824992418289, -0.02876586839556694, -0.24572737514972687, 0.0863448902964592, 0.06843195110559464, -0.00905382726341486, 0.11144173890352249, -0.012629742734134197, -0.057801470160484314, -0.06883607059717178, 0.08984968811273575, 0.04556552693247795, 0.024266494438052177, 0.01979539357125759, 0.06617983430624008, 0.049273669719696045, -0.03804517164826393, 0.016382187604904175, -0.12365459650754929, 0.044369909912347794, 0.0432717464864254, -0.041836030781269073, 0.04480966925621033, 0.03053205832839012, 0.0056099798530340195, 0.08471585810184479, 0.007446909788995981, 0.05087800696492195, 0.050146639347076416, -0.029931891709566116, -0.13102634251117706, 0.09985223412513733, -0.10292042046785355, -0.21186205744743347, -0.17965003848075867, -0.02355600707232952, -0.02998405694961548, 0.018534043803811073, 0.0010579433292150497, 0.030412040650844574, -0.0687146931886673, -0.11824163794517517, -0.00012002625589957461, 0.01625838875770569, -0.09508319944143295, -0.06531254947185516, 0.014802882447838783, 0.06931984424591064, -0.10026837885379791, -0.01981288753449917, 0.02514326386153698, -0.10467120260000229, -0.026072753593325615, -0.004091352224349976, 0.054232168942689896, 0.11832728236913681, 0.0016384805785492063, -0.02492719143629074, -0.02508169412612915, 0.21614079177379608, -0.08190403878688812, 0.08506729453802109, 0.17167140543460846, -0.09522060304880142, 0.07932157814502716, 0.16187556087970734, 0.04672829806804657, -0.04174816235899925, 0.050376445055007935, 0.07698068767786026, -0.003912958316504955, -0.24481093883514404, -0.06175331771373749, -0.04081881791353226, -0.07415883243083954, 0.07924997806549072, 0.05267034471035004, 0.06550682336091995, 0.045618753880262375, -0.07024642825126648, -0.03282589465379715, 0.1049051359295845, 0.09987892210483551, 0.19330726563930511, 0.011471414007246494, 0.0779888853430748, -0.06127047911286354, -0.00410437909886241, 0.07656659185886383, -0.009142271243035793, 0.07795994728803635, -0.025796500965952873, 0.14346717298030853, 0.08040666580200195, 0.06531871110200882, 0.0491495244204998, -0.0016716120298951864, -0.045267559587955475, 0.0359906330704689, -0.008056556805968285, -0.09585265815258026, -0.054928116500377655, 0.04764522239565849, 0.000388167827622965, 0.014852609485387802, 0.015552746132016182, -0.0959327220916748, 0.0646551251411438, 0.09986212104558945, 0.05934688076376915, -0.16007716953754425, -0.05985327810049057, 0.03672676905989647, -0.04334068298339844, -0.08898534625768661, -0.026180077344179153, 0.08052806556224823, -0.14403600990772247, 0.0906248465180397, -0.024163709953427315, 0.07886790484189987, -0.08087765425443649, -0.025417551398277283, 0.08224818110466003, 0.05012620985507965, -0.015199017710983753, 0.07296199351549149, -0.2053450495004654, 0.027767695486545563, 0.044440362602472305, 0.0803733617067337, -0.050407733768224716, 0.06318767368793488, 0.016020841896533966, -0.01441188808530569, 0.14088653028011322, -0.011154999025166035, -0.015110809355974197, -0.08404968678951263, -0.024059904739260674, -0.030177293345332146, 0.0940941721200943, -0.09871397912502289, 0.06182342767715454, -0.032700713723897934, 0.025275563821196556, -0.01940341107547283, 0.03673022985458374, -0.14355789124965668, -0.17323757708072662, 0.039417341351509094, 0.006947565358132124, 0.04813654348254204, -0.06146412342786789, -0.0038766830693930387, -0.05516178160905838, 0.14988552033901215, -0.19118888676166534, -0.09068509191274643, -0.1191604807972908, -0.025187471881508827, 0.10162020474672318, -0.06051075831055641, 0.04455271363258362, -0.039484668523073196, 0.1522204875946045, -0.034735579043626785, -0.05605725198984146, 0.025596776977181435, -0.03274581953883171, -0.17550957202911377, -0.05377655103802681, 0.15251652896404266, 0.11337488144636154, 0.047877006232738495, 0.02978096343576908, 0.05493735894560814, 0.0109811807051301, -0.0970231220126152, -0.009505684487521648, 0.14315325021743774, 0.018797654658555984, 0.04593796283006668, -0.08485636115074158, -0.10559085756540298, -0.09797584265470505, -0.008737119846045971, 0.08938635885715485, 0.14958623051643372, -0.04930885136127472, 0.08848327398300171, 0.13110435009002686, -0.1023210734128952, -0.20121166110038757, 0.017544563859701157, 0.07639531791210175, -0.00558410445228219, 0.03079076297581196, -0.17897044122219086, 0.10784337669610977, 0.08856319636106491, -0.030300896614789963, -0.0014308395329862833, -0.2233985811471939, -0.10658878087997437, 0.06005203723907471, 0.020921669900417328, -0.1846262812614441, -0.08364476263523102, -0.09089479595422745, -0.021955903619527817, -0.13636015355587006, 0.19855691492557526, -0.07480327785015106, 0.03413568064570427, 0.08350472897291183, 0.04118260368704796, 0.055366143584251404, -0.004053472075611353, 0.11467383056879044, -0.013343960978090763, 0.04983510076999664, -0.07418835908174515, -0.0047881766222417355, 0.0852104052901268, -0.03470358997583389, 0.11011751741170883, 0.003017807612195611, 0.02208840474486351, -0.04212014377117157, -0.06102435290813446, -0.05332072079181671, 0.07498431950807571, -0.08927726000547409, -0.06393499672412872, -0.06837179511785507, 0.0819091945886612, 0.0885036513209343, -0.038616351783275604, 0.001967168878763914, -0.09844273328781128, 0.03543471544981003, 0.19245505332946777, 0.13058046996593475, 0.10944738984107971, -0.09204229712486267, -0.02178392931818962, -0.026592159643769264, 0.05225107818841934, 0.005801075603812933, 0.06213260442018509, 0.06663957983255386, 0.008059973828494549, 0.14884622395038605, 0.021217549219727516, -0.1584874838590622, -0.0207193735986948, 0.008938525803387165, -0.0890979990363121, -0.12222245335578918, -0.0010523368837311864, 0.111147440969944, -0.10412690043449402, -0.0868341326713562, 0.12031267583370209, 0.006455635651946068, -0.022690704092383385, 0.056279826909303665, 0.032766979187726974, -0.006829261314123869, 0.06482395529747009, 0.0381491482257843, 0.019116440787911415, -0.058618560433387756, 0.07852320373058319, 0.12876464426517487, -0.11578933894634247, 0.0037393784150481224, 0.05748570337891579, -0.07253054529428482, -0.04821782186627388, -0.07615306228399277, 0.08265101909637451, -0.0318506620824337, -0.05374244973063469, 0.06127838417887688, -0.09747844934463501, 0.029218774288892746, 0.1677258312702179, 0.04446486756205559, 0.07827146351337433, -0.0381641685962677, 0.009957161732017994, -0.02018718235194683, 0.11834166198968887, 0.004308727104216814, 0.04255004972219467, -0.03432846441864967, 0.11354847252368927, 0.00008993459778139368, -0.04729216545820236, -0.0024265837855637074, -0.01620120368897915, -0.07841991633176804, -0.006950173527002335, -0.04537815600633621, 0.06237760931253433, -0.07960601150989532, 0.007002591155469418, 0.01809680461883545, 0.0006961643812246621, 0.017580129206180573, 0.017744921147823334, -0.02779896929860115, -0.04969208315014839, 0.0038623071741312742, 0.07379823923110962, -0.18820995092391968, -0.023771775886416435, 0.08993346989154816, -0.09902941435575485, 0.07979228347539902, -0.011241920292377472, -0.02911301515996456, -0.01730927638709545, -0.14113770425319672, -0.021151501685380936, -0.03947524353861809, 0.045783113688230515, -0.010135308839380741, -0.16974195837974548, 0.017363861203193665, 0.008344360627233982, -0.08293388783931732, 0.01743851602077484, 0.11515377461910248, -0.10298853367567062, 0.014104992151260376, 0.08464482426643372, -0.03202248364686966, -0.06958889216184616, 0.07965172082185745, 0.08496090024709702, 0.048066191375255585, 0.12605619430541992, -0.04187199845910072, 0.029286745935678482, -0.11219661682844162, -0.0040800608694553375, 0.013729245401918888, 0.02198873460292816, -0.056550122797489166, -0.03782903403043747, 0.03381219878792763, -0.02464781329035759, 0.12845095992088318, 0.02255203388631344, -0.047279104590415955, 0.05050785094499588, 0.029081134125590324, -0.004938426427543163, 0.0080612413585186, 0.12475858628749847, 0.011952285654842854, -0.014554542489349842, 0.03354262560606003, -0.0767747238278389, -0.0838659256696701, -0.013902057893574238, 0.09350477159023285, 0.10888350754976273, 0.045009005814790726, -0.025485800579190254, 0.03060740977525711, -0.0072455741465091705, -0.02879668027162552, 0.0900692418217659, -0.023666957393288612, 0.013006912544369698, -0.0728461891412735, 0.15427522361278534, 0.0874127671122551, -0.1736833155155182, 0.0994676798582077, 0.008601664565503597, -0.05711397901177406, -0.12453080713748932, -0.2611639201641083, -0.04486382007598877, 0.0466262586414814, 0.0026153663638979197, -0.0798128992319107, 0.05166516453027725, 0.0635630264878273, 0.00415911665186286, -0.016317257657647133, 0.12196475267410278, -0.1495428830385208, -0.11560597270727158, 0.0304441899061203, 0.027197638526558876, 0.050718583166599274, -0.007973657920956612, 0.011540323495864868, 0.049378398805856705, 0.08982640504837036, 0.06278698891401291, 0.06379028409719467, 0.09399215877056122, -0.012639501132071018, -0.05876365676522255, -0.025194793939590454, 0.027534330263733864, -0.004879562184214592, -0.016290761530399323, 0.14709050953388214, 0.04420265182852745, 0.01511642336845398, -0.001231727423146367, 0.1700618714094162, -0.03667335584759712, -0.12295754253864288, -0.1348976194858551, 0.1981636881828308, -0.0318462997674942, 0.03349979221820831, 0.027657605707645416, -0.103660948574543, 0.05049360170960426, 0.1294245570898056, 0.09379355609416962, -0.08456499129533768, 0.029085934162139893, 0.02872605249285698, 0.010336240753531456, 0.008227730169892311, 0.03517845645546913, -0.009995012544095516, 0.23820079863071442, -0.045254744589328766, 0.10678070038557053, -0.034739263355731964, -0.030108964070677757, -0.039922695606946945, 0.14141982793807983, -0.08995185792446136, 0.0013359851436689496, -0.07889311760663986, 0.08253282308578491, -0.031725890934467316, -0.22140555083751678, 0.0563875250518322, 0.016845565289258957, -0.054471418261528015, 0.022833723574876785, 0.03467576205730438, -0.017615115270018578, 0.018568245694041252, -0.0019176133209839463, 0.004369060043245554, 0.07838593423366547, 0.00720860343426466, -0.031132029369473457, 0.011464100331068039, 0.07345081120729446, -0.02558894269168377, 0.22972500324249268, 0.03350937366485596, 0.11435920000076294, 0.09517411887645721, 0.017059046775102615, -0.12317025661468506, 0.08224159479141235, 0.07045634835958481, -0.05580379441380501, 0.01994330994784832, 0.07487954944372177, 0.029323037713766098, 0.11739887297153473, 0.11075848340988159, -0.07707499712705612, 0.1036578044295311, 0.012868452817201614, -0.06505827605724335, -0.15643420815467834, 0.05634869262576103, -0.12899568676948547, 0.12427369505167007, 0.1617853194475174, -0.03732623532414436, 0.01121518760919571, -0.051468703895807266, -0.025569716468453407, 0.042464010417461395, 0.0684979036450386, 0.042089641094207764, -0.08233676850795746, 0.06120958551764488, 0.05643376335501671, 0.07780028134584427, -0.16914337873458862, -0.06620418280363083, 0.032203301787376404, -0.006257216911762953, -0.033146750181913376, 0.0719686970114708, 0.030494309961795807, -0.001967074116691947, -0.021023528650403023, -0.11629537492990494, 0.008850245736539364, 0.07701808959245682, -0.042824432253837585, -0.014016173779964447 ]
null
null
transformers
# `lic-class-scancode-bert-base-cased-L32-1` ## Intended Use This model is intended to be used for Sentence Classification which is used for results analysis in [`scancode-results-analyzer`](https://github.com/nexB/scancode-results-analyzer). `scancode-results-analyzer` helps detect faulty scans in [`scancode-toolkit`](https://github.com/nexB/scancode-results-analyzer) by using statistics and nlp modeling, among other tools, to make Scancode better. ## How to Use Refer [quickstart](https://github.com/nexB/scancode-results-analyzer#quickstart---local-machine) section in `scancode-results-analyzer` documentation, for installing and getting started. - [Link to Code](https://github.com/nexB/scancode-results-analyzer/blob/master/src/results_analyze/nlp_models.py) Then in `NLPModelsPredict` class, function `predict_basic_lic_class` uses this classifier to predict sentances as either valid license tags or false positives. ## Limitations and Bias As this model is a fine-tuned version of the [`bert-base-cased`](https://huggingface.co/bert-base-cased) model, it has the same biases, but as the task it is fine-tuned to is a very specific task (license text/notice/tag/referance) without those intended biases, it's safe to assume those don't apply at all here. ## Training and Fine-Tuning Data The BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers). Then this `bert-base-cased` model was fine-tuned on Scancode Rule texts, specifically trained in the context of sentence classification, where the four classes are - License Text - License Notice - License Tag - License Referance ## Training Procedure For fine-tuning procedure and training, refer `scancode-results-analyzer` code. - [Link to Code](https://github.com/nexB/scancode-results-analyzer/blob/master/src/results_analyze/nlp_models.py) In `NLPModelsTrain` class, function `prepare_input_data_false_positive` prepares the training data. In `NLPModelsTrain` class, function `train_basic_false_positive_classifier` fine-tunes this classifier. 1. Model - [BertBaseCased](https://huggingface.co/bert-base-cased) (Weights 0.5 GB) 2. Sentence Length - 32 3. Labels - 4 (License Text/Notice/Tag/Referance) 4. After 4 Epochs of Fine-Tuning with learning rate 2e-5 (60 secs each on an RTX 2060) Note: The classes aren't balanced. ## Eval Results - Accuracy on the training data (90%) : 0.98 (+- 0.01) - Accuracy on the validation data (10%) : 0.84 (+- 0.01) ## Further Work 1. Apllying Splitting/Aggregation Strategies 2. Data Augmentation according to Vaalidation Errors 3. Bigger/Better Suited Models
{"language": "en", "license": "apache-2.0", "tags": ["license", "sentence-classification", "scancode", "license-compliance"], "datasets": ["bookcorpus", "wikipedia", "scancode-rules"], "version": 1.0}
fill-mask
ayansinha/lic-class-scancode-bert-base-cased-L32-1
[ "transformers", "tf", "bert", "fill-mask", "license", "sentence-classification", "scancode", "license-compliance", "en", "dataset:bookcorpus", "dataset:wikipedia", "dataset:scancode-rules", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# 'lic-class-scancode-bert-base-cased-L32-1' ## Intended Use This model is intended to be used for Sentence Classification which is used for results analysis in 'scancode-results-analyzer'. 'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools, to make Scancode better. ## How to Use Refer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started. - Link to Code Then in 'NLPModelsPredict' class, function 'predict_basic_lic_class' uses this classifier to predict sentances as either valid license tags or false positives. ## Limitations and Bias As this model is a fine-tuned version of the 'bert-base-cased' model, it has the same biases, but as the task it is fine-tuned to is a very specific task (license text/notice/tag/referance) without those intended biases, it's safe to assume those don't apply at all here. ## Training and Fine-Tuning Data The BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers). Then this 'bert-base-cased' model was fine-tuned on Scancode Rule texts, specifically trained in the context of sentence classification, where the four classes are - License Text - License Notice - License Tag - License Referance ## Training Procedure For fine-tuning procedure and training, refer 'scancode-results-analyzer' code. - Link to Code In 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the training data. In 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes this classifier. 1. Model - BertBaseCased (Weights 0.5 GB) 2. Sentence Length - 32 3. Labels - 4 (License Text/Notice/Tag/Referance) 4. After 4 Epochs of Fine-Tuning with learning rate 2e-5 (60 secs each on an RTX 2060) Note: The classes aren't balanced. ## Eval Results - Accuracy on the training data (90%) : 0.98 (+- 0.01) - Accuracy on the validation data (10%) : 0.84 (+- 0.01) ## Further Work 1. Apllying Splitting/Aggregation Strategies 2. Data Augmentation according to Vaalidation Errors 3. Bigger/Better Suited Models
[ "# 'lic-class-scancode-bert-base-cased-L32-1'", "## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.", "## How to Use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_lic_class' uses this classifier to\npredict sentances as either valid license tags or false positives.", "## Limitations and Bias\n\nAs this model is a fine-tuned version of the 'bert-base-cased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific task\n(license text/notice/tag/referance) without those intended biases, it's safe to assume\nthose don't apply at all here.", "## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-cased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the four classes are\n\n\t- License Text\n\t- License Notice\n\t- License Tag\n\t- License Referance", "## Training Procedure\n\nFor fine-tuning procedure and training, refer 'scancode-results-analyzer' code.\n\n- Link to Code\n\nIn 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the\ntraining data.\n\nIn 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes\nthis classifier.\n\n1. Model - BertBaseCased (Weights 0.5 GB)\n2. Sentence Length - 32\n3. Labels - 4 (License Text/Notice/Tag/Referance)\n4. After 4 Epochs of Fine-Tuning with learning rate 2e-5 (60 secs each on an RTX 2060)\n\nNote: The classes aren't balanced.", "## Eval Results\n\n- Accuracy on the training data (90%) : 0.98 (+- 0.01) \n- Accuracy on the validation data (10%) : 0.84 (+- 0.01)", "## Further Work\n\n1. Apllying Splitting/Aggregation Strategies\n2. Data Augmentation according to Vaalidation Errors\n3. Bigger/Better Suited Models" ]
[ "TAGS\n#transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# 'lic-class-scancode-bert-base-cased-L32-1'", "## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.", "## How to Use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_lic_class' uses this classifier to\npredict sentances as either valid license tags or false positives.", "## Limitations and Bias\n\nAs this model is a fine-tuned version of the 'bert-base-cased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific task\n(license text/notice/tag/referance) without those intended biases, it's safe to assume\nthose don't apply at all here.", "## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-cased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the four classes are\n\n\t- License Text\n\t- License Notice\n\t- License Tag\n\t- License Referance", "## Training Procedure\n\nFor fine-tuning procedure and training, refer 'scancode-results-analyzer' code.\n\n- Link to Code\n\nIn 'NLPModelsTrain' class, function 'prepare_input_data_false_positive' prepares the\ntraining data.\n\nIn 'NLPModelsTrain' class, function 'train_basic_false_positive_classifier' fine-tunes\nthis classifier.\n\n1. Model - BertBaseCased (Weights 0.5 GB)\n2. Sentence Length - 32\n3. Labels - 4 (License Text/Notice/Tag/Referance)\n4. After 4 Epochs of Fine-Tuning with learning rate 2e-5 (60 secs each on an RTX 2060)\n\nNote: The classes aren't balanced.", "## Eval Results\n\n- Accuracy on the training data (90%) : 0.98 (+- 0.01) \n- Accuracy on the validation data (10%) : 0.84 (+- 0.01)", "## Further Work\n\n1. Apllying Splitting/Aggregation Strategies\n2. Data Augmentation according to Vaalidation Errors\n3. Bigger/Better Suited Models" ]
[ 86, 21, 86, 78, 88, 104, 179, 44, 38 ]
[ "passage: TAGS\n#transformers #tf #bert #fill-mask #license #sentence-classification #scancode #license-compliance #en #dataset-bookcorpus #dataset-wikipedia #dataset-scancode-rules #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# 'lic-class-scancode-bert-base-cased-L32-1'## Intended Use\n\nThis model is intended to be used for Sentence Classification which is used for results\nanalysis in 'scancode-results-analyzer'.\n\n'scancode-results-analyzer' helps detect faulty scans in 'scancode-toolkit' by using statistics and nlp modeling, among other tools,\nto make Scancode better.## How to Use\n\nRefer quickstart section in 'scancode-results-analyzer' documentation, for installing and getting started.\n\n- Link to Code\n\nThen in 'NLPModelsPredict' class, function 'predict_basic_lic_class' uses this classifier to\npredict sentances as either valid license tags or false positives.## Limitations and Bias\n\nAs this model is a fine-tuned version of the 'bert-base-cased' model,\nit has the same biases, but as the task it is fine-tuned to is a very specific task\n(license text/notice/tag/referance) without those intended biases, it's safe to assume\nthose don't apply at all here.## Training and Fine-Tuning Data\n\nThe BERT model was pretrained on BookCorpus, a dataset consisting of 11,038 unpublished books and English Wikipedia (excluding lists, tables and headers).\n\nThen this 'bert-base-cased' model was fine-tuned on Scancode Rule texts, specifically\ntrained in the context of sentence classification, where the four classes are\n\n\t- License Text\n\t- License Notice\n\t- License Tag\n\t- License Referance" ]
[ -0.04472174495458603, 0.15657564997673035, -0.00612260028719902, 0.04665049538016319, 0.04596731439232826, -0.0027946175541728735, 0.07285820692777634, 0.09863945096731186, 0.03382302448153496, 0.07375029474496841, -0.011146560311317444, 0.07966703921556473, 0.0914342850446701, 0.04661605879664421, 0.0072822305373847485, -0.16334396600723267, 0.054621871560811996, -0.036592885851860046, 0.06959258764982224, 0.08060131967067719, 0.12409800291061401, -0.055739037692546844, 0.06803949177265167, 0.05431957170367241, -0.057469312101602554, 0.009887638501822948, 0.02406669408082962, -0.02888207696378231, 0.059058357030153275, 0.051577333360910416, 0.08149614185094833, -0.0027773280162364244, 0.03310650959610939, -0.20616964995861053, 0.0070103867910802364, 0.08590137958526611, 0.02289905585348606, 0.039734065532684326, 0.11803711950778961, -0.01605874113738537, 0.13689178228378296, -0.05345567688345909, 0.1023307517170906, 0.08042185008525848, -0.10741784423589706, -0.13481402397155762, -0.1116180345416069, 0.10423141717910767, 0.02861509658396244, 0.06006481871008873, -0.017498495057225227, 0.05246398597955704, -0.01634649746119976, 0.05733112618327141, 0.09500036388635635, -0.261745423078537, -0.01911407895386219, 0.05576256290078163, 0.050448015332221985, -0.016106657683849335, -0.09571370482444763, -0.013304800726473331, -0.016135090962052345, 0.02321433648467064, 0.04428018629550934, -0.036117926239967346, -0.00852519366890192, -0.03263036534190178, -0.11048174649477005, -0.08222993463277817, 0.05629557743668556, -0.013983707875013351, -0.08158303797245026, -0.20953872799873352, -0.0676940381526947, 0.003729377407580614, 0.00731477839872241, -0.03815435245633125, 0.017790040001273155, 0.05389140173792839, 0.06563041359186172, -0.12928009033203125, -0.08415432274341583, -0.0007845778600312769, -0.07093454152345657, 0.08178094029426575, 0.011768855154514313, 0.01623818837106228, 0.04993433505296707, 0.14478695392608643, 0.051860056817531586, -0.08693855255842209, -0.06755094230175018, -0.041801851242780685, -0.10362216830253601, -0.051548853516578674, -0.021848436444997787, -0.023294657468795776, -0.0005445951246656477, 0.22356516122817993, -0.032700080424547195, 0.07264325767755508, -0.10918116569519043, 0.007231699302792549, -0.005737503990530968, 0.14826521277427673, 0.015649955719709396, -0.006215584930032492, 0.025819750502705574, -0.007469853851944208, 0.029833195731043816, -0.028268983587622643, -0.018406324088573456, -0.0971151813864708, 0.039520446211099625, 0.07618677616119385, 0.016303442418575287, 0.07755881547927856, -0.08368328958749771, -0.027698254212737083, 0.08041143417358398, -0.17305679619312286, 0.04873983934521675, 0.057814668864011765, -0.02591891959309578, -0.028450794517993927, 0.06341990828514099, 0.002238622633740306, -0.09270721673965454, 0.042229972779750824, -0.05613706633448601, -0.022879956290125847, -0.08986390382051468, -0.07164812833070755, 0.035353973507881165, -0.03126705437898636, -0.08093684166669846, -0.0641663447022438, -0.17760230600833893, -0.03267018869519234, 0.05411795899271965, -0.0370466373860836, 0.0115069430321455, 0.0005340282805263996, -0.01784108392894268, -0.003457360900938511, 0.005673091858625412, -0.015533744357526302, -0.016714399680495262, 0.04299311339855194, -0.056258879601955414, 0.019013965502381325, 0.015535158105194569, 0.01936730369925499, -0.14138953387737274, 0.05904524028301239, -0.18193300068378448, 0.16553537547588348, -0.062277570366859436, -0.04937775060534477, -0.11255378276109695, 0.014976485632359982, -0.0076978448778390884, 0.0070076691918075085, 0.02379736863076687, 0.09548918157815933, -0.15423241257667542, 0.015571126714348793, 0.15000653266906738, -0.13282433152198792, -0.07396508008241653, 0.04926394671201706, -0.06965494900941849, 0.13174685835838318, 0.09654784947633743, 0.11303620785474777, 0.11031603813171387, -0.1400589793920517, -0.047463711351156235, 0.027027424424886703, -0.02321106381714344, 0.14695850014686584, 0.03441879525780678, -0.06456463038921356, -0.04696336388587952, 0.007691322825849056, -0.09328540414571762, -0.038463957607746124, 0.03298052400350571, -0.04116368293762207, -0.013172664679586887, -0.008503081277012825, 0.01172311045229435, -0.03521557152271271, -0.0832521989941597, -0.017843009904026985, -0.11187510937452316, 0.059477854520082474, 0.059381917119026184, -0.0666753500699997, 0.04013528302311897, -0.14045317471027374, 0.03169107809662819, 0.005501228384673595, -0.037811119109392166, -0.1812175214290619, -0.16986657679080963, 0.05247703194618225, -0.13437148928642273, 0.08147992193698883, -0.04742712527513504, 0.012591508217155933, 0.011879215948283672, -0.030259937047958374, -0.002813675906509161, -0.015036772936582565, -0.011173446662724018, -0.02685295045375824, -0.09073049575090408, -0.039710674434900284, -0.002777061890810728, 0.15935255587100983, -0.16410565376281738, 0.021915709599852562, 0.1458038091659546, 0.12820033729076385, 0.030127687379717827, -0.06431827694177628, -0.005771074444055557, -0.030596155673265457, 0.009785672649741173, -0.07048891484737396, 0.03151949867606163, 0.07377580553293228, -0.07181866466999054, 0.12171374261379242, -0.15616120398044586, -0.15174230933189392, 0.07601407170295715, 0.04561830312013626, -0.11103065311908722, -0.06230155751109123, -0.014080964028835297, -0.02122846245765686, -0.055037971585989, -0.0792987048625946, 0.11675914376974106, 0.05812808871269226, 0.07528206706047058, -0.08957485854625702, -0.07914126664400101, -0.05058193579316139, -0.022464703768491745, -0.02565535344183445, 0.04797206073999405, 0.01113258209079504, -0.23874591290950775, 0.08866848051548004, 0.0460236519575119, 0.0068457284942269325, 0.09693123400211334, -0.007760666310787201, -0.055373664945364, -0.0450078621506691, 0.08178906887769699, 0.03405319154262543, -0.0030083185993134975, 0.016058823093771935, 0.06541892886161804, 0.04562412574887276, -0.03832884132862091, 0.019168561324477196, -0.11220234632492065, 0.05407999828457832, 0.036487724632024765, -0.042039066553115845, 0.040767572820186615, 0.03013542853295803, 0.0179760679602623, 0.07357949763536453, 0.01692689210176468, 0.049126312136650085, 0.04887928068637848, -0.03292449191212654, -0.12238459289073944, 0.10159561783075333, -0.09082823246717453, -0.20814120769500732, -0.20153513550758362, -0.04895169287919998, -0.018349576741456985, 0.01777033880352974, -0.0047172196209430695, 0.0469624288380146, -0.06839475780725479, -0.10943224281072617, -0.004105767700821161, 0.0032171595375984907, -0.09411540627479553, -0.06640946865081787, 0.019856620579957962, 0.06692410260438919, -0.1255456507205963, -0.023217765614390373, 0.039523668587207794, -0.10797062516212463, -0.04377667233347893, -0.0022538225166499615, 0.039611976593732834, 0.11667106300592422, -0.01284056156873703, -0.021464761346578598, -0.0258797574788332, 0.23463140428066254, -0.07759696990251541, 0.08611337840557098, 0.15460792183876038, -0.10459230095148087, 0.07747037708759308, 0.1576576679944992, 0.04295552894473076, -0.05219394713640213, 0.04006080701947212, 0.06578342616558075, -0.009720869362354279, -0.26601603627204895, -0.049583058804273605, -0.042349595576524734, -0.06570266932249069, 0.11046791821718216, 0.0665174350142479, 0.050828881561756134, 0.05855368450284004, -0.07947184145450592, 0.022605447098612785, 0.10973367094993591, 0.11407343298196793, 0.19629743695259094, 0.013755833730101585, 0.0632113590836525, -0.0687272846698761, 0.005943480879068375, 0.04871191084384918, 0.015804369002580643, 0.07129277288913727, -0.03868281841278076, 0.1597844362258911, 0.07236078381538391, 0.06155519187450409, 0.05090731382369995, 0.020904626697301865, -0.048052169382572174, 0.04759868606925011, 0.0024510486982762814, -0.09479893743991852, -0.06459465622901917, 0.05493181198835373, -0.0026819277554750443, 0.0020422888919711113, 0.02108088694512844, -0.09317949414253235, 0.04040371999144554, 0.15870730578899384, 0.040473394095897675, -0.19242417812347412, -0.05320711433887482, 0.028206637129187584, -0.05175658315420151, -0.10889093577861786, -0.01845434494316578, 0.07521524280309677, -0.14869655668735504, 0.09348125755786896, -0.02667824737727642, 0.0844842717051506, -0.11678335070610046, -0.036636706441640854, 0.07961957901716232, 0.06855009496212006, -0.015330453403294086, 0.06561856716871262, -0.20806600153446198, 0.010572805069386959, 0.04365938529372215, 0.08144566416740417, -0.04513213783502579, 0.07437282055616379, 0.014671244658529758, -0.012058697640895844, 0.12575842440128326, -0.015501408837735653, -0.04295558109879494, -0.07224605977535248, -0.02780519798398018, -0.021740542724728584, 0.07689538598060608, -0.09893743693828583, 0.07055581361055374, -0.05491984263062477, 0.015509368851780891, -0.01590723544359207, 0.00009727832366479561, -0.12308651208877563, -0.19477546215057373, 0.03011608123779297, 0.023283235728740692, 0.054954979568719864, -0.06045123189687729, 0.009051090106368065, -0.0201958529651165, 0.19924534857273102, -0.18710799515247345, -0.08973731100559235, -0.10441490262746811, -0.028850741684436798, 0.10915831476449966, -0.05065280571579933, 0.0738116130232811, -0.03373364731669426, 0.1620384305715561, -0.02690211497247219, -0.07057598978281021, 0.017300808802247047, -0.036926474422216415, -0.15726909041404724, -0.047333188354969025, 0.14892534911632538, 0.08419054746627808, 0.03311635181307793, 0.039771419018507004, 0.05487093701958656, 0.0357479490339756, -0.083170086145401, -0.020592650398612022, 0.1748228669166565, 0.018555622547864914, 0.030566392466425896, -0.0937606617808342, -0.07592923194169998, -0.08034758269786835, 0.006516230292618275, 0.08699292689561844, 0.1341371089220047, -0.04653909429907799, 0.09148117154836655, 0.12905222177505493, -0.11641433835029602, -0.20313939452171326, 0.032558150589466095, 0.0574951209127903, -0.0015978700248524547, 0.03371909260749817, -0.1917622834444046, 0.11912202835083008, 0.06642510741949081, -0.02891610935330391, 0.028344301506876945, -0.23436252772808075, -0.10268312692642212, 0.07158087193965912, 0.01691216044127941, -0.2004426270723343, -0.0549517497420311, -0.091698557138443, -0.046469174325466156, -0.1575801968574524, 0.16590967774391174, -0.050511736422777176, 0.04818287491798401, 0.08186034858226776, 0.03984851390123367, 0.06643946468830109, -0.014951962046325207, 0.10361555218696594, -0.031254466623067856, 0.04979672282934189, -0.08412545919418335, -0.02828688733279705, 0.0793277695775032, -0.04831329733133316, 0.10220859199762344, -0.0005327465478330851, 0.03486363962292671, -0.0125382449477911, -0.07090780138969421, -0.03566610440611839, 0.08625367283821106, -0.0773504450917244, -0.06671474874019623, -0.07661246508359909, 0.06364303082227707, 0.09457288682460785, -0.026169903576374054, -0.004610799718648195, -0.0900908038020134, 0.007317356765270233, 0.18545664846897125, 0.12985563278198242, 0.1275714635848999, -0.11158256232738495, -0.023666102439165115, -0.022858820855617523, 0.05326776206493378, 0.00983407348394394, 0.06086777523159981, 0.0673639327287674, 0.011539028026163578, 0.1292133331298828, 0.0293792262673378, -0.15518273413181305, -0.024459680542349815, 0.012021190486848354, -0.0868292897939682, -0.1116965040564537, 0.005177784711122513, 0.08739035576581955, -0.09702672809362411, -0.08869931101799011, 0.129991814494133, 0.0006776969530619681, -0.024611620232462883, 0.06474131345748901, 0.01317583303898573, -0.009606100618839264, 0.06054230406880379, 0.023607756942510605, 0.02368181385099888, -0.05738343670964241, 0.0969744324684143, 0.13013926148414612, -0.09110933542251587, 0.004212880507111549, 0.057893335819244385, -0.07656170427799225, -0.03692034259438515, -0.091294065117836, 0.08051601052284241, -0.01475805975496769, -0.04848729819059372, 0.0538574755191803, -0.06839019060134888, 0.03310062363743782, 0.16406689584255219, 0.02798960544168949, 0.09025324136018753, -0.04201296344399452, 0.016702864319086075, -0.03986748307943344, 0.10364348441362381, -0.012101313099265099, 0.05494917556643486, -0.06120714172720909, 0.10112854838371277, 0.002553042722865939, -0.03768914192914963, 0.003411180106922984, -0.021239759400486946, -0.0702350065112114, -0.006178665906190872, -0.058504555374383926, 0.07380016893148422, -0.08296964317560196, 0.009924578480422497, 0.02041558176279068, 0.014543039724230766, 0.010316911153495312, 0.02932085283100605, -0.03199252486228943, -0.06242687255144119, -0.01858648844063282, 0.0866040363907814, -0.18618854880332947, -0.016392314806580544, 0.09044193476438522, -0.10808748006820679, 0.08042088150978088, -0.015413682907819748, -0.03673449531197548, -0.017861787229776382, -0.14615392684936523, -0.02932773157954216, -0.028353996574878693, 0.04107343778014183, -0.01217578910291195, -0.13969409465789795, 0.017148319631814957, -0.003807585220783949, -0.07574284821748734, -0.003672513412311673, 0.08556633442640305, -0.11221983283758163, -0.0008227374637499452, 0.08137476444244385, -0.04133834317326546, -0.06222853437066078, 0.07329171895980835, 0.11050955951213837, 0.03688809648156166, 0.12141504138708115, -0.0430208183825016, 0.0235221479088068, -0.11524990946054459, 0.0027871476486325264, 0.013524219393730164, 0.02126035839319229, -0.034172479063272476, -0.05160650610923767, 0.03502450883388519, -0.025028977543115616, 0.1476864069700241, 0.019100399687886238, -0.06911896914243698, 0.051078975200653076, 0.014616873115301132, -0.02068888396024704, 0.024134734645485878, 0.1375889778137207, 0.014276783913373947, -0.019034286960959435, 0.049093712121248245, -0.07377728074789047, -0.07027457654476166, 0.018236443400382996, 0.07906351238489151, 0.12689578533172607, 0.04775604233145714, -0.011708403937518597, 0.02676471695303917, -0.007845841348171234, -0.0432518795132637, 0.06402090936899185, -0.004159239120781422, 0.022939041256904602, -0.08006380498409271, 0.17459793388843536, 0.09183798730373383, -0.16776889562606812, 0.09704887866973877, 0.01077142171561718, -0.05271511897444725, -0.12633183598518372, -0.26421308517456055, -0.05275586619973183, 0.04061871021986008, -0.000246424344368279, -0.07809539139270782, 0.06667671352624893, 0.08279877156019211, 0.0053649344481527805, -0.020634636282920837, 0.12202191352844238, -0.1460959017276764, -0.1197565346956253, 0.017923573032021523, 0.0290564876049757, 0.04429594427347183, -0.034782133996486664, -0.011141768656671047, 0.04974847286939621, 0.09945856779813766, 0.06096616014838219, 0.06280111521482468, 0.10679057985544205, -0.009519139304757118, -0.03978496789932251, -0.01753321662545204, 0.024982867762446404, 0.00039148685755208135, -0.02079744264483452, 0.11788050830364227, 0.04822239652276039, 0.007596207316964865, -0.005986285395920277, 0.18438315391540527, -0.034576863050460815, -0.1189216747879982, -0.1300070434808731, 0.17064513266086578, -0.020999735221266747, 0.02772231586277485, 0.03058105707168579, -0.1057245060801506, 0.04798504710197449, 0.14198216795921326, 0.09726332873106003, -0.06563969701528549, 0.02763860672712326, 0.014707360416650772, 0.010554850101470947, -0.0035811865236610174, 0.052833594381809235, -0.015981897711753845, 0.22272160649299622, -0.037676405161619186, 0.0889776423573494, -0.025191402062773705, -0.032565392553806305, -0.022022008895874023, 0.1544744074344635, -0.10394475609064102, -0.003635710570961237, -0.0803387388586998, 0.06439414620399475, -0.036977026611566544, -0.2456960827112198, 0.06656693667173386, 0.02437683194875717, -0.06069887802004814, 0.021602006629109383, -0.00013647362357005477, -0.01641843467950821, 0.01891598291695118, -0.009264733642339706, -0.011635011062026024, 0.09273942559957504, 0.004774745088070631, -0.058256711810827255, -0.007532655727118254, 0.07045874744653702, -0.019817838445305824, 0.279679536819458, 0.027492143213748932, 0.09363202750682831, 0.09010935574769974, 0.03525160625576973, -0.13323697447776794, 0.045502640306949615, 0.07203670591115952, -0.07422398775815964, 0.02855224534869194, 0.06382758915424347, 0.010912344790995121, 0.12190014868974686, 0.09200415015220642, -0.08757958561182022, 0.10226824134588242, 0.021453173831105232, -0.05115402489900589, -0.1644897162914276, 0.020683666691184044, -0.12366650253534317, 0.12711909413337708, 0.16667164862155914, -0.02678842842578888, 0.010760566219687462, -0.04546394199132919, -0.019863152876496315, 0.0432874895632267, 0.04105108976364136, 0.03827759623527527, -0.09830860048532486, 0.061090435832738876, 0.058231107890605927, 0.07327034324407578, -0.16294169425964355, -0.07469751685857773, 0.034964438527822495, -0.011884395964443684, -0.0416354201734066, 0.0894426554441452, 0.024534672498703003, -0.002493103500455618, -0.016090264543890953, -0.11286362260580063, 0.0032954972703009844, 0.08455535024404526, -0.06753736734390259, -0.028990985825657845 ]
null
null
transformers
# bert-base-cased trained on TREC 6-class task ## Model description A simple base BERT model trained on the "trec" dataset. ## Intended uses & limitations #### How to use ##### Transformers ```python # Load model and tokenizer from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) # Use pipeline from transformers import pipeline model_name = "aychang/bert-base-cased-trec-coarse" nlp = pipeline("sentiment-analysis", model=model_name, tokenizer=model_name) results = nlp(["Where did the queen go?", "Why did the Queen hire 1000 ML Engineers?"]) ``` ##### AdaptNLP ```python from adaptnlp import EasySequenceClassifier model_name = "aychang/bert-base-cased-trec-coarse" texts = ["Where did the queen go?", "Why did the Queen hire 1000 ML Engineers?"] classifer = EasySequenceClassifier results = classifier.tag_text(text=texts, model_name_or_path=model_name, mini_batch_size=2) ``` #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data TREC https://huggingface.co/datasets/trec ## Training procedure Preprocessing, hardware used, hyperparameters... #### Hardware One V100 #### Hyperparameters and Training Args ```python from transformers import TrainingArguments training_args = TrainingArguments( output_dir='./models', num_train_epochs=2, per_device_train_batch_size=16, per_device_eval_batch_size=16, warmup_steps=500, weight_decay=0.01, evaluation_strategy="steps", logging_dir='./logs', save_steps=3000 ) ``` ## Eval results ``` {'epoch': 2.0, 'eval_accuracy': 0.974, 'eval_f1': array([0.98181818, 0.94444444, 1. , 0.99236641, 0.96995708, 0.98159509]), 'eval_loss': 0.138086199760437, 'eval_precision': array([0.98540146, 0.98837209, 1. , 0.98484848, 0.94166667, 0.97560976]), 'eval_recall': array([0.97826087, 0.90425532, 1. , 1. , 1. , 0.98765432]), 'eval_runtime': 1.6132, 'eval_samples_per_second': 309.943} ```
{"language": ["en"], "license": "mit", "tags": ["text-classification"], "datasets": ["trec"], "model-index": [{"name": "aychang/bert-base-cased-trec-coarse", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "trec", "type": "trec", "config": "default", "split": "test"}, "metrics": [{"type": "accuracy", "value": 0.974, "name": "Accuracy", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMTUwZTU1ZGU5YTRiMzNhNmQyMjNlY2M5YjAwN2RlMmYxODI2MjFkY2Q3NWFjZDg3Zjg5ZDk1Y2I1MTUxYjFhMCIsInZlcnNpb24iOjF9.GJkxJOFhsO4UaoHpHH1136Qj_fu9UQ9o3DThtT46hvMduswkgobl9iz6ICYQ7IdYKFbh3zRTlsZzjnAlzGqdBA"}, {"type": "precision", "value": 0.9793164100816639, "name": "Precision Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMTMxMjI3NWZhOGZkODJmYzkxYzdhZWIwMTBkZTg4YWZiNjcwNTVmM2RjYmQ3ZmNhZjM2MWQzYTUzNzFlMjQzOCIsInZlcnNpb24iOjF9.n45s1_gW040u5f2y-zfVx_5XU-J97dcuWlmaIZsJsCetcHtrjsbHut2gAcPxErl8UPTXSq1XDg5WWug4FPM8CQ"}, {"type": "precision", "value": 0.974, "name": "Precision Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNTY5ZTZiNmYzZDQzYWZiZDdlNDllZWQ4NTVjZWZlYWJkZDgyNGNhZjAzOTZjZDc0NDUwMTE3ODVlMjFjNTIxZCIsInZlcnNpb24iOjF9.4lR7MgvxxTblEV4LZGbko-ylIeFjcjNM5P21iYH6vkNkjItIfiXmKbL55_Zeab4oGJ5ytWz0rIdlpNnmmV29Cw"}, {"type": "precision", "value": 0.9746805065928548, "name": "Precision Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZDEzYmZmZDIyNDFmNzJmODQ2ODdhYTUyYzQyZjEzZTdhMjg3MTllOGFkNGRlMDFhYzI4ZGE5OTExNjk1ZTI5OSIsInZlcnNpb24iOjF9.Ti5gL3Tk9hCpriIUhB8ltdKRibSilvRZOxAlLCgAkrhg0dXGE5f4n8almCAjbRJEaPW6H6581PhuUfjgMqceBw"}, {"type": "recall", "value": 0.9783617516169679, "name": "Recall Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNWUwMGUwYmY3MWQwOTcwYjI2Yjc3Yzc1YWQ1YjU2ODY3MzAyMDdkNmM3MmFhZmMxZWFhMTUxNzZlNzViMDA0ZiIsInZlcnNpb24iOjF9.IWhPl9xS5pqEaFHKsBZj6JRtJRpQZQqJhQYW6zmtPi2F3speRsKc0iksfHkmPjm678v-wKUJ4zyGfRs-63HmBg"}, {"type": "recall", "value": 0.974, "name": "Recall Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNjlhMDY0MmI2NzBiMWY5NTcwYjZlYzE5ODg0ODk1ZTBjZDI4YmZiY2RmZWVlZGUxYzk2MDQ4NjRkMTQ4ZTEzZiIsInZlcnNpb24iOjF9.g5p5b0BqyZxb7Hk9DayRndhs5F0r44h8TXMJDaP6IoFdYzlBfEcZv7UkCu6s6laz9-F-hhZHUZii2ljtYasVAA"}, {"type": "recall", "value": 0.974, "name": "Recall Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYjJjNTE2ZWFjMGYyZGUzOWI3MDRhM2I2MTRjZGNkOWZkZDJhNzQ4OTYwOTQ2NDY5OGNjZTZhOWU2MzlhNTY5YyIsInZlcnNpb24iOjF9.JnRFkZ-v-yRhCf6di7ONcy_8Tv0rNXQir1TVw-cU9fNY1c4vKRmGaKmLGeR7TxpmKzEQtikb6mFwRwhIAhl8AA"}, {"type": "f1", "value": 0.9783635353409951, "name": "F1 Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYjM2NDY3MmUyMmEyZjg5MWZhNjllOGRlNWVkYzgyYmM5ZDBmMDdhYmY5NDAxZmYwMjA0YTkzNTI2MjU0NTRlZiIsInZlcnNpb24iOjF9.HlbHjJa-bpYPjujWODpvfLVMtCnNQMDBCYpLGokfBoXibZGKfIzXcgNdXLdJ-DkmMUriX3wVZtGcRvA2ErUeDw"}, {"type": "f1", "value": 0.974, "name": "F1 Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYjMxNDE4MTBmYzU2MTllMjlhNTcwYWJhMzRkNTE2ZGFiNmQ0ZTEyOWJhMmU2ZDliYTIzNDExYTM5MTAxYjcxNSIsInZlcnNpb24iOjF9.B7G9Gs74MosZPQ16QH2k-zrmlE8KCtIFu3BcrgObYiuqOz1aFURS3IPoOynVFLp1jnJtgQAmQRY_GDumSS-oDg"}, {"type": "f1", "value": 0.97377371266232, "name": "F1 Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZmEyNjRlYmE5M2U1OWY0OGY2YjQyN2E0NmQxNjY0NTY3N2JiZmMwOWQ1ZTMzZDcwNTdjNWYwNTRiNTljNjMxMiIsInZlcnNpb24iOjF9.VryHh8G_ZvoiSm1SZRMw4kheGWuI3rQ6GUVqm2uf-kkaSU20rYMW20-VKCtwayLcrIHJ92to6YvvW7yI0Le5DA"}, {"type": "loss", "value": 0.13812002539634705, "name": "loss", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNjk4MDQ5NGRiNTExYmE3NGU1ZmQ1YjUzMTQ4NzUwNWViYzFiODEzMjc2MDA2MzYyOGNjNjYxYzliNDM4Y2U0ZSIsInZlcnNpb24iOjF9.u68ogPOH6-_pb6ZVulzMVfHIfFlLwBeDp8H4iqgfBadjwj2h-aO0jzc4umWFWtzWespsZvnlDjklbhhgrd1vCQ"}]}]}]}
text-classification
aychang/bert-base-cased-trec-coarse
[ "transformers", "pytorch", "jax", "bert", "text-classification", "en", "dataset:trec", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #bert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
# bert-base-cased trained on TREC 6-class task ## Model description A simple base BERT model trained on the "trec" dataset. ## Intended uses & limitations #### How to use ##### Transformers ##### AdaptNLP #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data TREC URL ## Training procedure Preprocessing, hardware used, hyperparameters... #### Hardware One V100 #### Hyperparameters and Training Args ## Eval results
[ "# bert-base-cased trained on TREC 6-class task", "## Model description\n\nA simple base BERT model trained on the \"trec\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nTREC URL", "## Training procedure\n\nPreprocessing, hardware used, hyperparameters...", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "# bert-base-cased trained on TREC 6-class task", "## Model description\n\nA simple base BERT model trained on the \"trec\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nTREC URL", "## Training procedure\n\nPreprocessing, hardware used, hyperparameters...", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ 56, 16, 20, 9, 5, 5, 5, 20, 6, 15, 6, 10, 4 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n# bert-base-cased trained on TREC 6-class task## Model description\n\nA simple base BERT model trained on the \"trec\" dataset.## Intended uses & limitations#### How to use##### Transformers##### AdaptNLP#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.## Training data\n\nTREC URL## Training procedure\n\nPreprocessing, hardware used, hyperparameters...#### Hardware\nOne V100#### Hyperparameters and Training Args## Eval results" ]
[ -0.042306482791900635, 0.039718471467494965, -0.0007646337617188692, 0.09010499715805054, 0.17626409232616425, -0.007981636561453342, 0.09881993383169174, 0.0875883623957634, -0.08996979892253876, -0.03754369914531708, 0.11623118817806244, 0.1697741448879242, 0.023455839604139328, 0.10696275532245636, -0.04159223660826683, -0.25405964255332947, 0.01328955590724945, 0.04140638932585716, -0.04673474282026291, 0.12127429991960526, 0.15842010080814362, -0.11825437098741531, 0.10437395423650742, -0.02754095569252968, -0.1353091448545456, 0.022339612245559692, 0.018069662153720856, -0.05202494189143181, 0.11859090626239777, 0.04269130155444145, 0.14209075272083282, -0.0030178732704371214, 0.11476544290781021, -0.20650531351566315, 0.019610660150647163, 0.054352063685655594, 0.00643192557618022, 0.08554814755916595, 0.05442109331488609, -0.028757065534591675, 0.04825931414961815, 0.0018985702190548182, 0.10738669335842133, 0.06044328585267067, -0.11644358187913895, -0.07032091170549393, -0.052550144493579865, 0.02561090514063835, 0.10555096715688705, 0.11605291068553925, 0.012608667835593224, 0.16112782061100006, -0.15530826151371002, 0.10318514704704285, 0.16377000510692596, -0.21657635271549225, -0.06772684305906296, 0.06076806038618088, 0.1311216652393341, 0.011721653863787651, -0.09188015758991241, -0.006505573634058237, 0.06289224326610565, 0.0791369080543518, 0.11840053647756577, -0.055811863392591476, -0.06350664794445038, 0.0543554350733757, -0.13447925448417664, -0.02256934531033039, 0.18300952017307281, -0.008512054570019245, -0.04173647612333298, 0.006632855162024498, -0.037048034369945526, -0.08055460453033447, -0.02502956986427307, -0.04263235256075859, 0.008946964517235756, 0.003943596966564655, -0.07005711644887924, -0.048745062202215195, -0.09818828850984573, -0.10244380682706833, -0.04798329994082451, 0.062118954956531525, 0.020722009241580963, -0.009708673693239689, -0.0848296657204628, 0.12782008945941925, 0.07149113714694977, -0.09827590733766556, 0.041322916746139526, -0.04756958782672882, -0.0380387045443058, -0.05963534861803055, -0.027157507836818695, -0.03580407425761223, 0.006692201364785433, 0.1649412363767624, 0.0387713797390461, 0.01926451176404953, -0.008754746988415718, 0.03350893780589104, -0.04037563130259514, 0.14429742097854614, -0.10818647593259811, -0.07698054611682892, 0.020714344456791878, 0.02197529748082161, -0.05082140490412712, -0.009108456782996655, -0.13954301178455353, 0.017163610085844994, 0.14420722424983978, 0.06965992599725723, 0.0094460379332304, 0.052471790462732315, -0.08647487312555313, -0.03931482881307602, 0.02117149531841278, -0.09391745179891586, 0.016266966238617897, -0.038181181997060776, -0.03130694106221199, -0.05049968138337135, 0.03307262808084488, 0.021156921982765198, -0.06188339367508888, 0.04341081902384758, -0.0894145593047142, -0.018053563311696053, -0.15655353665351868, -0.15970851480960846, 0.008213745430111885, -0.07889480143785477, 0.02143213152885437, -0.08364111185073853, -0.2168896645307541, -0.007159044966101646, 0.0801977813243866, -0.06272624433040619, -0.04049180448055267, -0.04926326870918274, -0.09169662743806839, 0.016363870352506638, -0.023503106087446213, 0.13587482273578644, -0.03770594298839569, 0.07474673539400101, -0.00003222043233108707, 0.04735986143350601, -0.03907598927617073, 0.05619008466601372, -0.10452857613563538, 0.014023453928530216, -0.1280202865600586, 0.05883365124464035, -0.07443518191576004, 0.08045605570077896, -0.11767583340406418, -0.09344889968633652, 0.016279159113764763, 0.023941610008478165, 0.07629583030939102, 0.13881786167621613, -0.22413787245750427, -0.020779849961400032, 0.08329182118177414, -0.047241032123565674, -0.07217586785554886, 0.13101573288440704, -0.0664292573928833, 0.11873705685138702, 0.0730208307504654, 0.09464450925588608, 0.006551679689437151, -0.06538049131631851, 0.02623855322599411, 0.06110162287950516, 0.01155275572091341, -0.017948376014828682, 0.0786602571606636, 0.01705738715827465, -0.07178092747926712, 0.009733340702950954, 0.0005718379397876561, 0.016153624281287193, -0.10855217278003693, -0.07590541243553162, 0.0075058164075016975, -0.0848228707909584, 0.07741474360227585, 0.015996277332305908, 0.054020386189222336, -0.05250374972820282, -0.13084635138511658, 0.14583033323287964, 0.06644856184720993, -0.0208677276968956, 0.02661159448325634, -0.12232577800750732, 0.05784137174487114, -0.11069667339324951, -0.022143028676509857, -0.1797610968351364, 0.005694780498743057, 0.0019531266298145056, 0.10071662813425064, 0.06803495436906815, 0.10612475872039795, 0.05989634245634079, 0.011281928978860378, -0.060348402708768845, -0.010552178137004375, -0.029121432453393936, 0.04958401247859001, -0.1327192485332489, -0.22184406220912933, -0.06355935335159302, -0.06182172894477844, 0.05779721960425377, -0.17065519094467163, 0.019851095974445343, 0.031044941395521164, 0.10525520890951157, 0.017237959429621696, 0.023801565170288086, -0.05038236454129219, 0.06801813840866089, -0.05708934739232063, -0.04113428667187691, 0.07040062546730042, 0.009652084670960903, -0.012050848454236984, 0.038501862436532974, -0.07528118789196014, 0.08916378766298294, 0.14128464460372925, -0.1634635478258133, -0.10591328144073486, 0.05863108113408089, -0.021631179377436638, -0.029037561267614365, -0.03971470519900322, -0.04379992187023163, 0.17398740351200104, 0.005086052697151899, 0.13108773529529572, -0.0651412308216095, -0.04352208226919174, 0.000707903120201081, -0.030032381415367126, 0.01005676668137312, 0.1290683150291443, 0.06474069505929947, -0.1674700677394867, 0.11196935921907425, 0.08223240077495575, -0.07124095410108566, 0.08399217575788498, 0.021676814183592796, -0.046777889132499695, -0.002729990752413869, -0.0846741795539856, 0.025668106973171234, 0.127662792801857, -0.18973013758659363, -0.020191779360175133, 0.028057847172021866, 0.021566594019532204, 0.03930770978331566, -0.15745066106319427, 0.0042275795713067055, 0.015193280763924122, -0.020647624507546425, 0.011454678140580654, -0.020182061940431595, -0.011859617196023464, 0.0829349011182785, 0.053477369248867035, -0.09549196064472198, 0.0825212299823761, -0.009659110568463802, -0.09630009531974792, 0.21936936676502228, -0.1035829409956932, -0.17326945066452026, -0.10176168382167816, -0.10278316587209702, 0.006124068982899189, 0.0205338504165411, 0.042096447199583054, -0.07387889176607132, -0.07584672421216965, -0.05579408258199692, -0.029776588082313538, -0.06399080902338028, 0.001851506414823234, -0.07562090456485748, 0.015257105231285095, 0.034168243408203125, -0.11831802129745483, -0.0014694215497002006, 0.01820485107600689, -0.09995046257972717, 0.02261638082563877, -0.0307226050645113, 0.09103573858737946, 0.1472516655921936, -0.03296704962849617, 0.016969796270132065, -0.03361661359667778, 0.23715490102767944, -0.0473952554166317, -0.006166831124573946, 0.12440440058708191, -0.01960660330951214, 0.012755508534610271, 0.14468897879123688, 0.03636215627193451, -0.06351862847805023, 0.05646509677171707, 0.04668756574392319, -0.04323866590857506, -0.2480432391166687, -0.07587607204914093, -0.06374817341566086, -0.0032295058481395245, 0.08193104714155197, 0.0608002170920372, 0.032371096312999725, 0.04523857682943344, 0.026162946596741676, 0.07510626316070557, -0.036686502397060394, 0.10033649206161499, 0.12145628035068512, -0.007694919593632221, 0.1262209117412567, -0.02239230088889599, -0.05847742781043053, 0.05529145523905754, 0.0269929189234972, 0.15700678527355194, 0.00416931277140975, 0.06935741007328033, 0.06273235380649567, 0.09468001872301102, 0.03187623247504234, 0.1220339685678482, -0.020306948572397232, 0.004588333889842033, -0.03981144726276398, -0.03280221298336983, -0.041055817157030106, 0.03180351108312607, -0.011721570044755936, 0.00015467341290786862, -0.07100227475166321, -0.04046829044818878, 0.025473203510046005, 0.21713455021381378, 0.08320265263319016, -0.2976492643356323, -0.04969258978962898, 0.032210823148489, -0.06754203885793686, -0.06418164074420929, 0.036664318293333054, 0.02308267541229725, -0.12776246666908264, -0.019831718876957893, -0.05501605197787285, 0.14166083931922913, -0.05791290104389191, 0.03344022482633591, 0.045195072889328, 0.0766429752111435, -0.02102169208228588, 0.13149896264076233, -0.3809219002723694, 0.2336985021829605, 0.019974371418356895, 0.11317634582519531, -0.06654052436351776, -0.023181460797786713, 0.0057024601846933365, 0.07436691224575043, 0.10023432224988937, -0.016302963718771935, 0.11821264773607254, -0.18073172867298126, -0.06957834959030151, 0.012077250517904758, 0.02527003549039364, 0.022134291008114815, 0.08084091544151306, -0.05720428377389908, -0.002817734843119979, 0.06040876358747482, -0.0021709981374442577, -0.14646658301353455, -0.10396388173103333, -0.005977791268378496, -0.05623561888933182, 0.04942753538489342, -0.09104733914136887, -0.10940930247306824, -0.004064321983605623, 0.19046470522880554, -0.02962738461792469, -0.002303078770637512, -0.13115811347961426, 0.000055201748182298616, 0.0924249216914177, -0.04720509424805641, 0.02037477307021618, -0.022147128358483315, 0.05897578224539757, 0.029776571318507195, -0.04596919193863869, 0.13752783834934235, -0.1041298657655716, -0.11677933484315872, -0.06328283250331879, 0.02057948336005211, 0.07176637649536133, 0.0797748938202858, 0.027325840666890144, -0.00531690614297986, -0.06725975126028061, -0.09714493900537491, -0.05648912116885185, 0.025824446231126785, 0.058553848415613174, 0.06722158193588257, -0.09428281337022781, 0.007795770186930895, -0.0503450371325016, -0.019454818218946457, 0.12838637828826904, 0.1722409427165985, -0.07326438277959824, 0.05168812721967697, 0.13140009343624115, -0.09548402577638626, -0.3196447789669037, 0.029937000945210457, 0.001732685137540102, 0.005635865963995457, -0.00011385208199499175, -0.1419064700603485, 0.11391058564186096, -0.025296038016676903, -0.022122768685221672, -0.024166718125343323, -0.16245964169502258, -0.12863072752952576, 0.16547854244709015, 0.09559577703475952, 0.07439716905355453, -0.10735324770212173, -0.02022033929824829, -0.1053374856710434, -0.14201846718788147, 0.14403796195983887, -0.1447158008813858, 0.10571828484535217, 0.004406512714922428, 0.0031113247387111187, -0.013492092490196228, -0.05968089401721954, 0.07071314007043839, -0.01246162410825491, 0.07613642513751984, -0.05110914260149002, -0.09252304583787918, 0.10362697392702103, -0.004378759302198887, 0.03495865687727928, -0.029379338026046753, 0.05770878493785858, -0.11798030883073807, -0.06060558557510376, -0.06318868696689606, 0.0686928927898407, -0.029345566406846046, -0.11528193950653076, -0.04666990414261818, 0.06647012382745743, 0.02130582556128502, -0.021549077704548836, 0.04389207065105438, -0.05688054487109184, 0.07178905606269836, 0.171102374792099, 0.25020962953567505, -0.07614234834909439, 0.021924585103988647, 0.03570670634508133, -0.05372227355837822, 0.06672441959381104, -0.04281080141663551, 0.0347733274102211, 0.1140672117471695, 0.005594122689217329, 0.1463392823934555, 0.09639297425746918, -0.06010594218969345, 0.008172789588570595, 0.04798189178109169, -0.09060194343328476, -0.10710339993238449, -0.029576988890767097, 0.041312430053949356, -0.13240431249141693, 0.008844034746289253, 0.10121889412403107, -0.10317522287368774, -0.025737186893820763, -0.003160991007462144, 0.008250928483903408, -0.10066784173250198, 0.21024233102798462, 0.09844011813402176, 0.05818304792046547, -0.09143488109111786, 0.07599466294050217, 0.04315419867634773, -0.03365793451666832, 0.02798647992312908, 0.09112122654914856, -0.09262082725763321, -0.06212351843714714, 0.11565009504556656, 0.2542009651660919, -0.1341995745897293, -0.05335652083158493, -0.11925627291202545, -0.0939636304974556, 0.08339809626340866, 0.1689523607492447, 0.0901569128036499, 0.058089256286621094, -0.07537510246038437, 0.02077619358897209, -0.11211235076189041, 0.10371174663305283, 0.05842215567827225, 0.054102275520563126, -0.10715728253126144, 0.09998144209384918, -0.0010504923993721604, 0.09190394729375839, -0.06850353628396988, 0.0003635631292127073, -0.14157772064208984, 0.037386517971754074, -0.20441441237926483, 0.0027233518194407225, -0.02564222551882267, -0.003914867527782917, 0.008368123322725296, -0.07625715434551239, -0.04195794090628624, 0.011014080606400967, -0.0866519957780838, -0.0122285271063447, 0.02604127675294876, 0.03382917493581772, -0.08617950975894928, -0.006641983985900879, 0.04578515887260437, -0.035149555653333664, 0.05881201848387718, -0.011659566313028336, -0.019479352980852127, 0.047650039196014404, -0.18990802764892578, -0.00865083932876587, -0.016175204887986183, 0.024874364957213402, 0.08531536906957626, -0.06620810180902481, -0.013168588280677795, 0.027860352769494057, 0.06395045667886734, 0.011683633551001549, 0.07336924970149994, -0.10842668265104294, 0.008357558399438858, -0.0461258627474308, -0.09362854063510895, -0.045108456164598465, 0.08448570221662521, 0.08780936151742935, 0.053028933703899384, 0.16417433321475983, -0.06486594676971436, 0.040830664336681366, -0.1352347582578659, 0.015380890108644962, -0.036607977002859116, -0.05799153074622154, -0.09290650486946106, -0.061751965433359146, 0.0701313316822052, -0.03925659880042076, 0.16732420027256012, 0.07993470877408981, 0.02905280515551567, 0.04899900406599045, -0.02618340775370598, 0.026036959141492844, 0.03441665321588516, 0.19366899132728577, 0.031422488391399384, 0.007082850206643343, -0.04692225158214569, 0.07145530730485916, 0.031048739328980446, 0.1344718039035797, 0.15288230776786804, 0.134200781583786, 0.010433657094836235, 0.0628495141863823, 0.01549951545894146, -0.05975363776087761, -0.12660256028175354, -0.015882764011621475, 0.02006198838353157, 0.05983666330575943, -0.07342841476202011, 0.17639978229999542, 0.1315539926290512, -0.1377413272857666, 0.05193198099732399, -0.05672982335090637, -0.10416271537542343, -0.13840490579605103, -0.06314114481210709, -0.0813603550195694, -0.09488338232040405, -0.01221198495477438, -0.1251121312379837, 0.000158048773300834, 0.12682947516441345, 0.036676075309515, -0.02772768959403038, 0.1163395345211029, -0.012549180537462234, 0.02985108457505703, 0.024157939478754997, -0.016221098601818085, 0.030166899785399437, -0.06039951741695404, -0.024249987676739693, -0.06341319531202316, -0.00900027621537447, 0.028109479695558548, -0.02771170251071453, -0.07366050034761429, 0.020603662356734276, 0.019760649651288986, -0.08735903352499008, 0.019874228164553642, 0.02160145901143551, 0.03662578761577606, 0.19925329089164734, 0.0130991842597723, -0.017546914517879486, 0.027397597208619118, 0.2663344442844391, -0.1186366155743599, -0.15103603899478912, -0.11558640003204346, 0.2790355980396271, 0.026930347084999084, 0.039104171097278595, 0.0032267202623188496, -0.06437168270349503, 0.01055531669408083, 0.2717064321041107, 0.22001899778842926, -0.09600067883729935, -0.013912239111959934, 0.035828787833452225, -0.003626894438639283, -0.02376333251595497, 0.1161968782544136, 0.07432270795106888, 0.1521005779504776, -0.1435527354478836, -0.0031484533101320267, -0.017628280445933342, -0.04950318858027458, -0.06860236078500748, -0.018990173935890198, 0.06249648332595825, -0.011120823211967945, -0.04445216432213783, 0.06761195510625839, -0.10233201086521149, -0.1254214495420456, -0.0006143694627098739, -0.06685532629489899, -0.10611185431480408, -0.03542055934667587, 0.04481175169348717, -0.009362448006868362, 0.05812865495681763, -0.020951123908162117, 0.03672689199447632, 0.05194457992911339, 0.00508527597412467, -0.09104893356561661, -0.06109319254755974, 0.15189293026924133, 0.02590198442339897, 0.17817504703998566, -0.05517982318997383, 0.07282447069883347, 0.1438746154308319, 0.02047152630984783, -0.08977510780096054, 0.07627680897712708, 0.00987841747701168, -0.06673596799373627, 0.0811447724699974, 0.05015037953853607, -0.023498933762311935, 0.06284309178590775, -0.02544206939637661, -0.10321740806102753, 0.05684942007064819, -0.06457361578941345, -0.06369937211275101, -0.08762158453464508, 0.008025799877941608, -0.08682986348867416, 0.14244140684604645, 0.15341296792030334, -0.0482790470123291, 0.00004768256621900946, -0.08954180032014847, 0.019306764006614685, 0.013116419315338135, 0.027467813342809677, -0.06324029713869095, -0.19221457839012146, -0.005787283647805452, 0.05797430872917175, -0.008405041880905628, -0.2638043761253357, -0.03581792861223221, -0.029933910816907883, -0.033208101987838745, -0.0951169803738594, 0.08830960094928741, 0.09179744124412537, 0.05474511533975601, -0.0558929443359375, -0.03179166838526726, -0.07175492495298386, 0.1268831044435501, -0.17300358414649963, -0.09089912474155426 ]
null
null
null
# TorchScript model of bert-large-cased-whole-word-masking-finetuned-squad ## Model description A serialized torchscript model of bert-large-cased-whole-word-masking-finetuned-squad with a config.pbtxt for deployment using NVIDIA Triton Inference Server.
{"language": ["en"], "license": "mit", "tags": ["question-answering", "torchscript", "FastNN"], "datasets": ["squad"]}
question-answering
aychang/bert-large-cased-whole-word-masking-finetuned-squad
[ "question-answering", "torchscript", "FastNN", "en", "dataset:squad", "license:mit", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us
# TorchScript model of bert-large-cased-whole-word-masking-finetuned-squad ## Model description A serialized torchscript model of bert-large-cased-whole-word-masking-finetuned-squad with a URL for deployment using NVIDIA Triton Inference Server.
[ "# TorchScript model of bert-large-cased-whole-word-masking-finetuned-squad", "## Model description\n\nA serialized torchscript model of bert-large-cased-whole-word-masking-finetuned-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ "TAGS\n#question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us \n", "# TorchScript model of bert-large-cased-whole-word-masking-finetuned-squad", "## Model description\n\nA serialized torchscript model of bert-large-cased-whole-word-masking-finetuned-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ 33, 29, 51 ]
[ "passage: TAGS\n#question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us \n# TorchScript model of bert-large-cased-whole-word-masking-finetuned-squad## Model description\n\nA serialized torchscript model of bert-large-cased-whole-word-masking-finetuned-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ 0.00539175420999527, -0.22668680548667908, -0.003855516202747822, 0.13933292031288147, 0.04963967576622963, 0.04896323382854462, 0.21962666511535645, 0.019280891865491867, 0.07429719716310501, -0.001933113788254559, 0.1167193129658699, 0.03419942036271095, -0.00023775383306201547, -0.04307788237929344, 0.04311243072152138, -0.15370739996433258, 0.034333813935518265, -0.04956141859292984, -0.03574412688612938, 0.06702622771263123, 0.07473956793546677, -0.04836384952068329, 0.10047844052314758, -0.008903888054192066, -0.06506385654211044, -0.004636375233530998, -0.030011219903826714, -0.007000804413110018, 0.10093027353286743, 0.07505388557910919, 0.06291764229536057, -0.022897671908140182, 0.006608135532587767, -0.051047325134277344, 0.05501352623105049, 0.0020797974430024624, -0.06309730559587479, 0.026512375101447105, -0.015908150002360344, 0.1213405653834343, 0.18899065256118774, -0.03196320682764053, 0.02857048809528351, 0.0716804563999176, -0.0665367916226387, 0.007564947474747896, 0.010075331665575504, -0.14777125418186188, 0.03901961073279381, 0.03500739485025406, 0.001916619949042797, 0.14727605879306793, -0.20014235377311707, 0.08056236803531647, 0.009365241043269634, -0.3129826486110687, -0.13318867981433868, 0.022528858855366707, 0.1512223780155182, 0.044758886098861694, -0.03546128422021866, 0.0936838835477829, -0.02742515504360199, 0.06662795692682266, 0.03224673122167587, -0.05910564586520195, -0.2736019492149353, 0.0050287144258618355, -0.10779981315135956, 0.01022742036730051, 0.15605603158473969, 0.07712734490633011, 0.010948395356535912, -0.07159317284822464, -0.098845936357975, 0.11722587794065475, -0.05922296270728111, 0.02875116840004921, -0.0012070754310116172, -0.01802874356508255, -0.016109783202409744, -0.06499132513999939, -0.0495426245033741, -0.15459522604942322, -0.11353825777769089, 0.11055195331573486, 0.027890188619494438, 0.09556980431079865, -0.24682679772377014, -0.020660022273659706, 0.09859545528888702, -0.1000584065914154, -0.0368492491543293, -0.09916166216135025, -0.03206949681043625, 0.07079339772462845, 0.004836705047637224, -0.06648752838373184, 0.2159091681241989, 0.02269885130226612, 0.17050068080425262, 0.025990501046180725, -0.09069956839084625, 0.04932199418544769, -0.039816003292798996, 0.035738006234169006, -0.02574051171541214, -0.023138293996453285, 0.03285945579409599, -0.004527676850557327, 0.11557552218437195, -0.03159940242767334, -0.1220051497220993, -0.00015799638640601188, -0.11584613472223282, 0.05217042192816734, 0.10420704632997513, 0.09703020006418228, -0.013695372268557549, -0.02051372267305851, 0.14054259657859802, -0.030490735545754433, -0.0008549683843739331, 0.02766585722565651, 0.0026306514628231525, -0.09408652037382126, 0.0020242936443537474, -0.055110614746809006, 0.02220163680613041, -0.08497757464647293, -0.04308159649372101, -0.06607203185558319, -0.05031783878803253, -0.10615303367376328, 0.0925813838839531, -0.05340244248509407, -0.01755657233297825, -0.1454252302646637, -0.04977288842201233, 0.11781761795282364, -0.00020578976545948535, -0.01457340270280838, 0.029941344633698463, -0.13203683495521545, -0.03512927144765854, 0.05303407832980156, 0.0324600525200367, 0.1513567417860031, -0.03294292464852333, 0.030762063339352608, 0.031105035915970802, 0.12007909268140793, -0.22469547390937805, -0.030869169160723686, -0.10336808115243912, -0.004321481101214886, -0.0016776067204773426, 0.06663543730974197, -0.020301546901464462, 0.04985910281538963, -0.02778717502951622, -0.04427004233002663, -0.10213224589824677, 0.0222195852547884, 0.036461200565099716, 0.1727278232574463, -0.21363836526870728, -0.018113616853952408, 0.2587487995624542, -0.016234733164310455, -0.22640202939510345, 0.19464708864688873, -0.022548628970980644, -0.07503001391887665, 0.05441238358616829, 0.08725448697805405, 0.060307446867227554, -0.07936324179172516, 0.05099775269627571, 0.08501806110143661, -0.16826026141643524, -0.019764408469200134, 0.10854807496070862, 0.02931518480181694, -0.12087962031364441, 0.04228466749191284, 0.11183205991983414, 0.005722219590097666, -0.05529940873384476, -0.0692436471581459, -0.03369355946779251, -0.04628024250268936, 0.09652794152498245, -0.02797606959939003, -0.027325773611664772, -0.1337953358888626, -0.08030159026384354, -0.052982937544584274, 0.06111922487616539, 0.026000957936048508, 0.002689109183847904, -0.14993898570537567, -0.018713882192969322, -0.04262951388955116, 0.08846653997898102, -0.09057002514600754, -0.27021512389183044, 0.04703861474990845, 0.10960455983877182, 0.04830174520611763, 0.060618408024311066, 0.04791064187884331, 0.027765406295657158, -0.008218381553888321, 0.04879317060112953, -0.02365514636039734, -0.032214198261499405, -0.01697157882153988, -0.1668953001499176, -0.06443231552839279, -0.05834652855992317, 0.013131286017596722, -0.11814264208078384, 0.02136504091322422, -0.016324885189533234, -0.09640884399414062, -0.058455344289541245, 0.0661652535200119, 0.03055126778781414, -0.029024308547377586, -0.03985577076673508, -0.010102083906531334, 0.07889766246080399, -0.03462100028991699, -0.08248282968997955, -0.025532636791467667, 0.014681325294077396, -0.04290103167295456, 0.09081223607063293, -0.07525116205215454, 0.06599953770637512, 0.08743638545274734, -0.04487419128417969, -0.010401024483144283, -0.019341498613357544, -0.06319907307624817, 0.04176889359951019, 0.014196980744600296, 0.019814031198620796, -0.09929978847503662, 0.061887919902801514, 0.04804529249668121, -0.0015854761004447937, 0.0026703428011387587, 0.10009728372097015, 0.08353584259748459, -0.18284688889980316, -0.01781804859638214, 0.07858077436685562, 0.03068256564438343, 0.09628304094076157, 0.03874841332435608, 0.014370009303092957, -0.01628771238029003, -0.010532532818615437, 0.04361370578408241, 0.08434972912073135, -0.24001622200012207, 0.019427375867962837, 0.04493257775902748, -0.026025142520666122, -0.03487006202340126, -0.06388842314481735, -0.019063515588641167, -0.03565285727381706, -0.06775233894586563, -0.18014346063137054, 0.08183015137910843, 0.009953932836651802, 0.08982360363006592, 0.04605155065655708, -0.1383449286222458, 0.06330413371324539, -0.10069219022989273, -0.06467031687498093, 0.14303110539913177, -0.03493295982480049, -0.17659880220890045, -0.13076381385326385, -0.04318786412477493, -0.14563605189323425, 0.039349328726530075, 0.10541932284832001, -0.0707012265920639, 0.07094872742891312, -0.07804790139198303, 0.07218940556049347, -0.013925113715231419, 0.03380889818072319, -0.06497622281312943, 0.001044963370077312, -0.03539140522480011, -0.1295180469751358, -0.028324130922555923, -0.07721593976020813, -0.12838302552700043, 0.07199359685182571, -0.0805811807513237, 0.14870940148830414, 0.04467600956559181, 0.027294382452964783, -0.01293263677507639, -0.05684022605419159, 0.07421882450580597, -0.005120133049786091, 0.05647635459899902, 0.15979741513729095, 0.038947347551584244, 0.046796008944511414, 0.09039773792028427, 0.0719887763261795, -0.06304222345352173, 0.06159361079335213, -0.016127072274684906, -0.11467033624649048, -0.13959023356437683, -0.07164561748504639, -0.12933899462223053, 0.11796043813228607, 0.029216622933745384, 0.0628964826464653, -0.12092609703540802, 0.05703430995345116, -0.0046020993031561375, -0.00886517483741045, 0.013917583972215652, 0.052584704011678696, 0.12420955300331116, 0.04866637662053108, 0.07584874331951141, -0.09905799478292465, -0.039498548954725266, 0.0855548307299614, -0.04665151983499527, 0.11179399490356445, 0.0828624963760376, 0.1788083165884018, 0.16350391507148743, 0.05139829218387604, 0.06411389261484146, 0.08911486715078354, 0.01650821790099144, -0.009020361118018627, -0.008623601868748665, -0.09008890390396118, -0.010541667230427265, -0.034994833171367645, -0.008297291584312916, 0.10683257132768631, -0.019332628697156906, 0.1358485072851181, 0.10368524491786957, 0.2996146082878113, 0.0650414526462555, -0.28873300552368164, -0.024107355624437332, 0.01663558930158615, 0.020503101870417595, -0.054254740476608276, 0.04656578600406647, 0.11158382147550583, -0.07771775126457214, 0.03203019127249718, -0.004034630488604307, 0.09740573167800903, 0.08497684448957443, 0.06323093175888062, 0.040418099611997604, 0.09993242472410202, 0.029265310615301132, 0.12385140359401703, -0.2665833830833435, 0.09803827852010727, 0.01776137202978134, 0.060430869460105896, -0.09682231396436691, -0.03954773023724556, 0.027265362441539764, -0.012492438778281212, 0.1265517771244049, -0.007802353240549564, 0.1936308741569519, 0.013372700661420822, -0.09553200751543045, 0.12350683659315109, 0.083226278424263, 0.023236475884914398, 0.039382509887218475, 0.01347073819488287, -0.028995126485824585, 0.043894700706005096, 0.08367016911506653, -0.029484089463949203, -0.07881079614162445, -0.06615663319826126, 0.15988580882549286, -0.16831649839878082, -0.02962302416563034, -0.008049669675529003, -0.02035946398973465, 0.07377465814352036, -0.04402093589305878, -0.02778284065425396, -0.03253887966275215, -0.06684927642345428, 0.033171698451042175, -0.06047626584768295, 0.017316970974206924, -0.10315313190221786, 0.030024003237485886, -0.008375401608645916, -0.13563765585422516, 0.13727806508541107, -0.0943167582154274, 0.018415533006191254, -0.08627818524837494, 0.13729992508888245, 0.016117874532938004, 0.036593325436115265, 0.03484785929322243, -0.04021281749010086, -0.041081514209508896, -0.09890252351760864, -0.01348723191767931, 0.04627861827611923, -0.052191443741321564, 0.026567960157990456, -0.1083332747220993, 0.09462755173444748, 0.005353523418307304, 0.08436102420091629, 0.13958631455898285, 0.17401911318302155, -0.07632487267255783, 0.015181594528257847, 0.11720874905586243, 0.010241951793432236, -0.2079111784696579, -0.077732615172863, -0.03022710606455803, -0.01681697927415371, -0.0030528265051543713, -0.1308622658252716, 0.15274594724178314, 0.059327077120542526, 0.008328370749950409, 0.28434914350509644, -0.23138950765132904, -0.05932464823126793, 0.10054483264684677, 0.05116020143032074, 0.10559167712926865, -0.1097756028175354, -0.022122301161289215, -0.08232323080301285, -0.1661812663078308, 0.1243218258023262, -0.13878977298736572, 0.11128512769937515, -0.05009061098098755, 0.011875971220433712, -0.030452294275164604, -0.020612969994544983, 0.04295840859413147, 0.00006741348624927923, 0.06916692107915878, -0.04485594108700752, -0.06625071167945862, 0.23421765863895416, -0.0319599024951458, 0.08645982295274734, -0.16140158474445343, 0.011110457591712475, -0.05098629370331764, -0.017902741208672523, -0.10571396350860596, 0.035219188779592514, -0.035949621349573135, -0.011232402175664902, -0.050821453332901, -0.05030502751469612, -0.01304539293050766, 0.02308720536530018, 0.05383938178420067, -0.020688071846961975, -0.06335516273975372, -0.0625915452837944, 0.1840551644563675, -0.0669953003525734, -0.05983332172036171, -0.05192654952406883, -0.023261452093720436, 0.10258438438177109, -0.030511246994137764, 0.015929438173770905, 0.05095631629228592, 0.01648727059364319, -0.02835930325090885, 0.10934624075889587, 0.009589480236172676, 0.038517337292432785, 0.06019645184278488, -0.056000154465436935, -0.2598022222518921, 0.05130128189921379, 0.10495026409626007, -0.14776284992694855, 0.07488791644573212, 0.037403471767902374, 0.025997940450906754, -0.007494249381124973, 0.030872009694576263, 0.02775474451482296, -0.0655217319726944, -0.01735636778175831, 0.041646815836429596, 0.08670152723789215, -0.0967477411031723, 0.11187436431646347, -0.02244078554213047, 0.06262049823999405, -0.016325028613209724, 0.10645405948162079, -0.09209119528532028, -0.0760589987039566, -0.02043182961642742, 0.14804355800151825, 0.07411075383424759, -0.06561346352100372, -0.0006643884116783738, -0.147417813539505, 0.012999370694160461, 0.11414661258459091, 0.06141028180718422, 0.047435007989406586, -0.060326844453811646, 0.040283653885126114, -0.022040706127882004, 0.06136650592088699, -0.18182159960269928, 0.03909612074494362, 0.029696650803089142, -0.10016521066427231, -0.01776152476668358, 0.14715199172496796, -0.07294332981109619, -0.038540374487638474, -0.08065368235111237, 0.046714719384908676, -0.15815851092338562, 0.04069308191537857, -0.013270186260342598, 0.004816344939172268, -0.0491098016500473, 0.02284763567149639, -0.05327186733484268, -0.012200050055980682, 0.025835173204541206, -0.028922734782099724, 0.017654351890087128, 0.07284873723983765, -0.1562378704547882, -0.031235864385962486, 0.021859508007764816, -0.04706491902470589, 0.07729018479585648, 0.07624447345733643, -0.05722307041287422, 0.014053956605494022, -0.1280285269021988, -0.07628892362117767, -0.005953322630375624, 0.07360467314720154, 0.08559178560972214, 0.0850609689950943, -0.009843500331044197, 0.01104672159999609, 0.05329728126525879, -0.02146233804523945, 0.02238398604094982, -0.05273408442735672, 0.06659521162509918, -0.02115468680858612, -0.03671128302812576, -0.03173145651817322, -0.04771832376718521, 0.019046861678361893, 0.07390553504228592, 0.13933099806308746, -0.03406478837132454, 0.02031308226287365, -0.11972375959157944, -0.004604165907949209, 0.005653261207044125, -0.08251982927322388, -0.022994251921772957, -0.08137782663106918, 0.06389530003070831, -0.06458290666341782, 0.10555600374937057, -0.020604072138667107, -0.027635570615530014, -0.021328384056687355, 0.08572210371494293, -0.024668602272868156, -0.04435233771800995, 0.16432158648967743, 0.021514147520065308, -0.05589000880718231, 0.0709671825170517, -0.023395011201500893, 0.10255437344312668, 0.20933111011981964, 0.23413075506687164, 0.1440040022134781, 0.11032677441835403, 0.09432122111320496, 0.17409224808216095, 0.10092440992593765, -0.07055148482322693, -0.1485528200864792, 0.0432254783809185, 0.056068114936351776, -0.04834742844104767, 0.04532290995121002, 0.04470868781208992, 0.014270537532866001, 0.022329572588205338, -0.01631556637585163, -0.037208687514066696, -0.06706351041793823, -0.11794242262840271, -0.07557746022939682, 0.027290211990475655, -0.005083595402538776, -0.08556997030973434, -0.09005279839038849, 0.16392070055007935, 0.058103516697883606, -0.03488465026021004, 0.21439963579177856, -0.002893167082220316, 0.07648103684186935, 0.08309675753116608, -0.03441089391708374, 0.0037608107086271048, 0.11444156616926193, -0.007679999805986881, 0.007582665421068668, 0.01670648157596588, 0.01478566974401474, 0.04280969873070717, -0.032639119774103165, 0.047923531383275986, -0.10005613416433334, -0.0844963788986206, -0.045995187014341354, 0.040357474237680435, -0.020816609263420105, 0.020672617480158806, 0.11033886671066284, -0.0577421560883522, 0.04910654202103615, 0.046535737812519073, -0.009688199497759342, -0.17004773020744324, 0.016610588878393173, 0.10412877053022385, -0.14688843488693237, 0.010405167937278748, -0.10047906637191772, -0.016830511391162872, -0.09046617150306702, 0.21590059995651245, 0.30217066407203674, -0.16207556426525116, -0.0017538709798827767, -0.029315495863556862, 0.014467108994722366, -0.007722174748778343, 0.1461239904165268, 0.017532091587781906, 0.11231231689453125, -0.027759350836277008, -0.04611575976014137, -0.03714176267385483, 0.048385147005319595, -0.07965194433927536, -0.0825689360499382, 0.05356841906905174, 0.010696575976908207, -0.04679039865732193, -0.0014719339087605476, -0.07992851734161377, 0.04502451792359352, -0.0128385154530406, 0.031359024345874786, -0.045378364622592926, -0.04316502809524536, -0.01711014099419117, 0.053666453808546066, 0.13060757517814636, -0.09805049002170563, 0.014203627593815327, -0.04987554997205734, -0.02612139657139778, -0.12066468596458435, -0.0883919969201088, 0.09917647391557693, -0.06707978248596191, 0.1834573894739151, -0.04659082368016243, 0.09977280348539352, 0.09508228302001953, 0.0019261683337390423, -0.0723726898431778, 0.09679971635341644, 0.03566101938486099, 0.04406704008579254, 0.03534403443336487, -0.11592835932970047, -0.012141212821006775, -0.03050142340362072, 0.055204663425683975, 0.08278992027044296, -0.010516962967813015, 0.0021841181442141533, -0.016186898574233055, -0.12403680384159088, 0.038276366889476776, -0.06635575741529465, 0.07317284494638443, 0.12553516030311584, -0.0518542155623436, -0.06636211276054382, -0.02718227542936802, 0.0075663188472390175, 0.040814727544784546, -0.004624513443559408, -0.11141104251146317, -0.06297649443149567, -0.04147540405392647, 0.021241996437311172, 0.014527825638651848, -0.3065709173679352, -0.04736078530550003, -0.11858712881803513, 0.028279917314648628, 0.01880805939435959, -0.03971419483423233, 0.1612056940793991, 0.051307689398527145, -0.04198308661580086, -0.1526225507259369, 0.0004264266462996602, 0.13572201132774353, -0.13207940757274628, -0.11827687919139862 ]
null
null
transformers
# TREC 6-class Task: distilbert-base-cased ## Model description A simple base distilBERT model trained on the "trec" dataset. ## Intended uses & limitations #### How to use ##### Transformers ```python # Load model and tokenizer from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) # Use pipeline from transformers import pipeline model_name = "aychang/distilbert-base-cased-trec-coarse" nlp = pipeline("sentiment-analysis", model=model_name, tokenizer=model_name) results = nlp(["Where did the queen go?", "Why did the Queen hire 1000 ML Engineers?"]) ``` ##### AdaptNLP ```python from adaptnlp import EasySequenceClassifier model_name = "aychang/distilbert-base-cased-trec-coarse" texts = ["Where did the queen go?", "Why did the Queen hire 1000 ML Engineers?"] classifer = EasySequenceClassifier results = classifier.tag_text(text=texts, model_name_or_path=model_name, mini_batch_size=2) ``` #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data TREC https://huggingface.co/datasets/trec ## Training procedure Preprocessing, hardware used, hyperparameters... #### Hardware One V100 #### Hyperparameters and Training Args ```python from transformers import TrainingArguments training_args = TrainingArguments( output_dir='./models', overwrite_output_dir=False, num_train_epochs=2, per_device_train_batch_size=16, per_device_eval_batch_size=16, warmup_steps=500, weight_decay=0.01, evaluation_strategy="steps", logging_dir='./logs', fp16=False, eval_steps=500, save_steps=300000 ) ``` ## Eval results ``` {'epoch': 2.0, 'eval_accuracy': 0.97, 'eval_f1': array([0.98220641, 0.91620112, 1. , 0.97709924, 0.98678414, 0.97560976]), 'eval_loss': 0.14275787770748138, 'eval_precision': array([0.96503497, 0.96470588, 1. , 0.96969697, 0.98245614, 0.96385542]), 'eval_recall': array([1. , 0.87234043, 1. , 0.98461538, 0.99115044, 0.98765432]), 'eval_runtime': 0.9731, 'eval_samples_per_second': 513.798} ```
{"language": ["en"], "license": "mit", "tags": ["text-classification"], "datasets": ["trec"], "model-index": [{"name": "aychang/distilbert-base-cased-trec-coarse", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "trec", "type": "trec", "config": "default", "split": "test"}, "metrics": [{"type": "accuracy", "value": 0.97, "name": "Accuracy", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZGNmZTQ1Mjk3YTQ0NTdiZmY2NGM2NDM2Yzc2OTI4NGNiZDg4MmViN2I0ZGZiYWJlMTg1ZDU0MTc2ZTg1NjcwZiIsInZlcnNpb24iOjF9.4x_Ze9S5MbAeIHZ4p1EFmWev8RLkAIYWKqouAzYOxTNqdfFN0HnqULiM19EMP42v658vl_fR3-Ig0xG45DioCA"}, {"type": "precision", "value": 0.9742915631870833, "name": "Precision Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMjA2MWVjMDc3MDYyY2M3NzY4NGNhY2JlNzJjMGQzZDUzZjE3ZWI1MjVmMzc4ODM2ZTQ4YmRhOTVkZDU0MzJiNiIsInZlcnNpb24iOjF9.EfmXJ6w5_7dK6ys03hpADP9h_sWuPAHgxpltUtCkJP4Ys_Gh8Ak4pGS149zt5AdP_zkvsWlXwAvx5BDMEoB2AA"}, {"type": "precision", "value": 0.97, "name": "Precision Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNDVjOGFjM2RkMDMxZTFiMzE1ZDM4OTRjMzkwOWE2NTJmMmUwMDdiZDg5ZjExYmFmZjg2Y2Y5NzcxZWVkODkwZSIsInZlcnNpb24iOjF9.BtO7DqJsUhSXE-_tJZJOPPd421VmZ3KR9-KkrhJkLNenoV2Xd6Pu6i5y6HZQhFB-9WfEhU9cCsIPQ1ioZ7dyDA"}, {"type": "precision", "value": 0.9699546283251607, "name": "Precision Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGQ0Mzc2MTE2YjkwNGY1MDEzNWQwYmNlZDMzZjBmNWM0ODExYjM1OTQyZGJkNjI2OTA5MDczZjFmOGM5MmMzMyIsInZlcnNpb24iOjF9.fGi2qNpOjWd1ci3p_E1p80nOqabiKiQqpQIxtk5aWxe_Nzqh3XiOCBF8vswCRvX8qTKdCc2ZEJ4s8dZMeltfCA"}, {"type": "recall", "value": 0.972626762268805, "name": "Recall Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMjQwMWZiYjIyMGVhN2M1ZDE5M2EzZmQ1ODRlYzE0MzJhZmU3ZTM1MmIyNTg5ZjBlMDcyMmQ0NmYzZjFmMmM4NSIsInZlcnNpb24iOjF9.SYDxsRw0xoQuQhei0YBdUbBxG891gqLafVFLdPMCJtQIktqCTrPW0sMKtis7GA-FEbNQVu8lp92znvlryNiFCw"}, {"type": "recall", "value": 0.97, "name": "Recall Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMjQ0MjczYjFhZDdiMjdkMWVlZTAzYWU0ODVhNjkxN2I1N2Y1Y2IyOTNlYWQxM2UxODIyNDZhZDM3MWIwMTgzZCIsInZlcnNpb24iOjF9.C5cfDTz_H4Y7nEO4Eq_XFy92CSbo3IBuL5n8wBKkTuB6hSgctTHOdOJzV8gWyMJ9gRcNqxp_yVU4BEB_I_0KAA"}, {"type": "recall", "value": 0.97, "name": "Recall Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNDZmYWM3OWExZWI1ZjRiZjczYWQwOWI5NWQzNDNkODcyMjBhMmVkYjY0MGZjYzlhNWQ0Y2MyMjc3OWEyZjY4NCIsInZlcnNpb24iOjF9.65WM5ihNfbKOCNZ6apX7iVAC2Ge_cwz9Xwa5oJHFq3Ci97eBFqK-qtADdB_SFRcSQUoNodaBeIhNfe0hVddxCA"}, {"type": "f1", "value": 0.9729834427867218, "name": "F1 Macro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYWQyZGZmYjU4NjE4M2YzMTUxOWVkYjU0YTFmYzE3MmQ2NjhmNDY1MGRmNGQ1MWZjYjM1Mzg5Y2RmNTk5YmZiMSIsInZlcnNpb24iOjF9.WIF-fmV0SZ6-lcg3Rz6TjbVl7nLvy_ftDi8PPhDIP1V61jgR1AcjLFeEgeZLxSFMdmU9yqG2DWYubF0luK0jCg"}, {"type": "f1", "value": 0.97, "name": "F1 Micro", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMDM0NDY0YzI2ZTBjYWVmZmVkOTI4ODkzM2RhNWM2ZjkwYTU3N2FjNjA4NjUwYWVjODNhMGEwMzdhYmE2YmIwYyIsInZlcnNpb24iOjF9.sihEhcsOeg8dvpuGgC-KCp1PsRNyguAif2uTBv5ELtRnM5KmMaHzRqpdpdc88Dj_DeuY6Y6qPQJt_dGk2q1rDQ"}, {"type": "f1", "value": 0.9694196751375908, "name": "F1 Weighted", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMTQ5ZjdiM2NiNDNkZTY5ZjNjNWUzZmI1MzgwMjhhNDEzMTEzZjFiNDhmZDllYmI0NjIwYjY0ZjcxM2M0ODE3NSIsInZlcnNpb24iOjF9.x4oR_PL0ALHYl-s4S7cPNPm4asSX3s3h30m-TKe7wpyZs0x6jwOqF-Tb1kgd4IMLl23pzsezmh72e_PmBFpRCg"}, {"type": "loss", "value": 0.14272506535053253, "name": "loss", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiODU3NGFiMzIxYWI4NzYxMzUxZGE5ZTZkYTlkN2U5MTI1NzA5NTBiNGM3Y2Q5YmVmZjU0MmU5MjJlZThkZTllMCIsInZlcnNpb24iOjF9.3QeWbECpJ0MHV5gC0_ES6PpwplLsCHPKuToErB1MSG69xNWVyMjKu1-1YEWZOU6dGfwKGh_HvwucY5kC9qwWBQ"}]}]}]}
text-classification
aychang/distilbert-base-cased-trec-coarse
[ "transformers", "pytorch", "distilbert", "text-classification", "en", "dataset:trec", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #distilbert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
# TREC 6-class Task: distilbert-base-cased ## Model description A simple base distilBERT model trained on the "trec" dataset. ## Intended uses & limitations #### How to use ##### Transformers ##### AdaptNLP #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data TREC URL ## Training procedure Preprocessing, hardware used, hyperparameters... #### Hardware One V100 #### Hyperparameters and Training Args ## Eval results
[ "# TREC 6-class Task: distilbert-base-cased", "## Model description\n\nA simple base distilBERT model trained on the \"trec\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nTREC URL", "## Training procedure\n\nPreprocessing, hardware used, hyperparameters...", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ "TAGS\n#transformers #pytorch #distilbert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "# TREC 6-class Task: distilbert-base-cased", "## Model description\n\nA simple base distilBERT model trained on the \"trec\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nTREC URL", "## Training procedure\n\nPreprocessing, hardware used, hyperparameters...", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ 55, 16, 22, 9, 5, 5, 5, 20, 6, 15, 6, 10, 4 ]
[ "passage: TAGS\n#transformers #pytorch #distilbert #text-classification #en #dataset-trec #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n# TREC 6-class Task: distilbert-base-cased## Model description\n\nA simple base distilBERT model trained on the \"trec\" dataset.## Intended uses & limitations#### How to use##### Transformers##### AdaptNLP#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.## Training data\n\nTREC URL## Training procedure\n\nPreprocessing, hardware used, hyperparameters...#### Hardware\nOne V100#### Hyperparameters and Training Args## Eval results" ]
[ -0.05427967384457588, 0.015722500160336494, -0.0009245880646631122, 0.08824535459280014, 0.18027552962303162, 0.01509677805006504, 0.12258879840373993, 0.07990962266921997, -0.0790855884552002, -0.03062337078154087, 0.10461676865816116, 0.18650265038013458, 0.023496853187680244, 0.11970707774162292, -0.045831818133592606, -0.26709020137786865, 0.01514724362641573, 0.023050542920827866, -0.034245822578668594, 0.1327408403158188, 0.16867993772029877, -0.10593067109584808, 0.10082412511110306, -0.029874417930841446, -0.14412373304367065, -0.0010079317726194859, 0.02255718596279621, -0.053271062672138214, 0.11776119470596313, 0.03248246759176254, 0.13130001723766327, 0.0026545117143541574, 0.1140175461769104, -0.2170119285583496, 0.020970460027456284, 0.05926140397787094, 0.004029375966638327, 0.07428277283906937, 0.04321259260177612, -0.04956211522221565, 0.08874251693487167, -0.01476860512048006, 0.1111912801861763, 0.06198383495211601, -0.11789332330226898, -0.06811072677373886, -0.06509844213724136, 0.04401542991399765, 0.09320621192455292, 0.12629447877407074, 0.017397822812199593, 0.1597595065832138, -0.15892937779426575, 0.10476186871528625, 0.13999679684638977, -0.18278959393501282, -0.06308826804161072, 0.075089231133461, 0.12139420211315155, 0.018397673964500427, -0.08602495491504669, -0.009090571664273739, 0.058821581304073334, 0.08737020194530487, 0.10294145345687866, -0.052884772419929504, -0.059660498052835464, 0.03682604059576988, -0.12508073449134827, -0.02610776014626026, 0.2045547068119049, -0.02272394299507141, -0.024346880614757538, -0.02709376811981201, -0.05211912840604782, -0.08454663306474686, -0.014032755047082901, -0.0457107238471508, 0.0005017623188905418, 0.005768822506070137, -0.04112733528017998, -0.04122541472315788, -0.09486344456672668, -0.096492700278759, -0.06692682206630707, 0.060116883367300034, 0.023071210831403732, 0.0004306823539081961, -0.08647938072681427, 0.1315034180879593, 0.10018341988325119, -0.1066347137093544, 0.027793018147349358, -0.06999967247247696, -0.05250910669565201, -0.05845999717712402, -0.026699863374233246, -0.020982274785637856, 0.01484833937138319, 0.211665078997612, -0.0019646501168608665, 0.019081640988588333, 0.02418305166065693, 0.03339317440986633, -0.028669200837612152, 0.13568131625652313, -0.10116568952798843, -0.08220402151346207, 0.0066678486764431, 0.020628850907087326, -0.03624677285552025, -0.01539369486272335, -0.12539620697498322, 0.0013110707513988018, 0.1373165100812912, 0.07375787943601608, 0.034975551068782806, 0.059067502617836, -0.08213519304990768, -0.05823611095547676, 0.049202047288417816, -0.10282536596059799, 0.008731979876756668, -0.04623420163989067, -0.03703823685646057, -0.020325494930148125, 0.02950255014002323, 0.019818726927042007, -0.0679621696472168, 0.05558367073535919, -0.08600161969661713, -0.009266190230846405, -0.1554909199476242, -0.1418517678976059, 0.01058361865580082, -0.09405449032783508, 0.016113508492708206, -0.08430436253547668, -0.2445155680179596, -0.024838363751769066, 0.07350200414657593, -0.05910627543926239, -0.03579018637537956, -0.03213714808225632, -0.08542105555534363, 0.016121728345751762, -0.02282293513417244, 0.08890065550804138, -0.04872363060712814, 0.065929114818573, -0.004201827570796013, 0.04438115283846855, -0.04363972321152687, 0.052520185708999634, -0.10886035114526749, 0.021506451070308685, -0.12284215539693832, 0.07260521501302719, -0.06723778694868088, 0.10671304911375046, -0.10150325298309326, -0.10457361489534378, -0.014690935611724854, 0.014220521785318851, 0.056599728763103485, 0.1310238540172577, -0.24000076949596405, -0.015318945050239563, 0.0787125676870346, -0.08011627942323685, -0.09865808486938477, 0.12057091295719147, -0.060741398483514786, 0.12016040086746216, 0.06703197956085205, 0.09227462112903595, 0.05400187522172928, -0.059995491057634354, 0.014156303368508816, 0.056469231843948364, -0.019174905493855476, -0.0133155956864357, 0.06373509019613266, 0.021421242505311966, -0.019996123388409615, 0.018591932952404022, 0.01924959570169449, 0.01383102498948574, -0.10415390878915787, -0.08159299194812775, -0.015574374236166477, -0.07453900575637817, 0.07090174406766891, 0.032777316868305206, 0.036101650446653366, -0.04662429913878441, -0.10944687575101852, 0.14039532840251923, 0.08322543650865555, -0.011689665727317333, 0.02222355827689171, -0.11983364075422287, 0.05873008072376251, -0.07397959381341934, -0.02329554408788681, -0.19343224167823792, -0.013348935171961784, -0.0031998371705412865, 0.11016406863927841, 0.06411555409431458, 0.08563011139631271, 0.05141777545213699, 0.0019402934703975916, -0.055470891296863556, -0.010841792449355125, -0.026205917820334435, 0.04994809627532959, -0.11075836420059204, -0.21856115758419037, -0.056281015276908875, -0.05181822553277016, 0.08702563494443893, -0.15390902757644653, 0.02309049293398857, 0.03010033816099167, 0.11694897711277008, 0.006708104629069567, 0.03095138631761074, -0.05809133127331734, 0.06610593944787979, -0.06229695677757263, -0.0551532506942749, 0.05796978622674942, 0.002280097920447588, -0.01339059416204691, 0.026770757511258125, -0.09389436990022659, 0.0962705984711647, 0.15096549689769745, -0.1746557503938675, -0.09682291001081467, 0.04259556904435158, -0.034020520746707916, -0.03637370467185974, -0.03365990146994591, -0.025849398225545883, 0.12810136377811432, 0.0030181698966771364, 0.10633964836597443, -0.061180587857961655, -0.04022350162267685, -0.008921501226723194, -0.026663780212402344, 0.010693574324250221, 0.10615789890289307, 0.07952850311994553, -0.1606467217206955, 0.11191172152757645, 0.07701181620359421, -0.05687520280480385, 0.08791331946849823, 0.009598120115697384, -0.04212281480431557, -0.005646699108183384, -0.08547189831733704, 0.018830973654985428, 0.13059501349925995, -0.16300925612449646, -0.0032094051130115986, 0.033900391310453415, 0.027722224593162537, 0.0590268038213253, -0.15434381365776062, 0.006966275628656149, 0.0095216054469347, -0.024488933384418488, 0.012451575137674809, -0.0037169724237173796, 0.007483774330466986, 0.07924798130989075, 0.03602196276187897, -0.06633071601390839, 0.07654169201850891, -0.01605231687426567, -0.10144376754760742, 0.2132726013660431, -0.12189024686813354, -0.19297020137310028, -0.09944253414869308, -0.040341656655073166, 0.000819026492536068, 0.01415235549211502, 0.03249070420861244, -0.06036119535565376, -0.0756058394908905, -0.05394865944981575, -0.022139299660921097, -0.052012696862220764, -0.005450433120131493, -0.07001190632581711, 0.03334859386086464, 0.022121943533420563, -0.12085838615894318, -0.0025495521258562803, 0.020847706124186516, -0.08508238196372986, 0.03993496298789978, -0.03120565414428711, 0.09407615661621094, 0.17077156901359558, -0.01543886587023735, 0.014976778998970985, -0.03532908111810684, 0.24149329960346222, -0.07130003720521927, 0.011476940475404263, 0.16444244980812073, 0.001396776526235044, 0.01183000486344099, 0.15100157260894775, 0.02732923813164234, -0.07035120576620102, 0.06253289431333542, 0.04262208193540573, -0.04884735122323036, -0.2484802007675171, -0.1058955192565918, -0.07589984685182571, -0.012608791701495647, 0.06471740454435349, 0.06662608683109283, 0.08786608278751373, 0.053425975143909454, 0.0012316509382799268, 0.0584569089114666, -0.01881648786365986, 0.09063167870044708, 0.15576905012130737, 0.002792702754959464, 0.12815552949905396, -0.036060694605112076, -0.03463013097643852, 0.0639939233660698, 0.020536402240395546, 0.17910528182983398, -0.007930267602205276, 0.07053080946207047, 0.07400086522102356, 0.10152031481266022, 0.022940419614315033, 0.11152429133653641, -0.0027895462699234486, 0.015939539298415184, -0.022630883380770683, -0.027041753754019737, -0.05031292513012886, 0.042319681495428085, -0.019676364958286285, -0.007834574207663536, -0.05789611488580704, 0.012282801792025566, 0.05517077445983887, 0.2001340687274933, 0.06668656319379807, -0.30231162905693054, -0.07070353627204895, 0.030069230124354362, -0.05238141492009163, -0.05395970866084099, 0.03449515625834465, 0.011798390187323093, -0.13262058794498444, 0.0066116563975811005, -0.0545053593814373, 0.1394091546535492, -0.0642433911561966, 0.02791752852499485, 0.08183173835277557, 0.08577056974172592, -0.009197692386806011, 0.12319186329841614, -0.38850775361061096, 0.21006423234939575, 0.013571304269134998, 0.11571551859378815, -0.08428552746772766, -0.01548728346824646, 0.002325759269297123, 0.1032148152589798, 0.07951321452856064, -0.011331788264214993, 0.0721270889043808, -0.14359676837921143, -0.044741228222846985, -0.0004598814412020147, 0.04666417837142944, 0.03269306570291519, 0.07621637731790543, -0.06981758028268814, 0.015102590434253216, 0.06499597430229187, -0.03725980594754219, -0.1642967164516449, -0.12255235761404037, 0.013059010729193687, -0.05999978259205818, 0.014802634716033936, -0.09604089707136154, -0.09601009637117386, 0.006017561536282301, 0.16568800806999207, -0.06282616406679153, -0.02837366797029972, -0.14535237848758698, -0.003795838914811611, 0.1258898675441742, -0.051467739045619965, 0.030690306797623634, -0.012475063093006611, 0.04035082086920738, 0.02601437270641327, -0.06270487606525421, 0.12294788658618927, -0.10656541585922241, -0.13967007398605347, -0.043981775641441345, 0.03019552119076252, 0.06186971813440323, 0.06360062956809998, 0.014966868795454502, -0.005031880922615528, -0.06261438131332397, -0.11985723674297333, -0.06816845387220383, 0.0575563982129097, 0.04453689232468605, 0.08378025144338608, -0.0764986053109169, 0.013796190731227398, -0.04161043092608452, -0.020850276574492455, 0.10458695143461227, 0.15012119710445404, -0.0784381702542305, 0.06377047300338745, 0.08677699416875839, -0.09311801195144653, -0.31019288301467896, 0.030684366822242737, -0.015828952193260193, -0.0007848346722312272, 0.009263631887733936, -0.12350526452064514, 0.11030448228120804, -0.02195342816412449, -0.01895308494567871, 0.005241620820015669, -0.22581620514392853, -0.14101412892341614, 0.14476026594638824, 0.06575636565685272, 0.03941395506262779, -0.10729504376649857, -0.023100081831216812, -0.11032884567975998, -0.17198842763900757, 0.128675177693367, -0.1523299515247345, 0.09827198088169098, -0.005740809254348278, 0.005157439038157463, -0.018191851675510406, -0.062003690749406815, 0.08231069892644882, 0.00877310335636139, 0.09197058528661728, -0.07341660559177399, -0.07154438644647598, 0.12729622423648834, -0.002233132254332304, 0.04628528654575348, -0.023841766640543938, 0.07370263338088989, -0.09826620668172836, -0.06996475905179977, -0.06415902078151703, 0.07675136625766754, -0.03990646451711655, -0.11693789064884186, -0.06340918689966202, 0.0761885792016983, 0.02488255314528942, -0.03399917483329773, 0.021317338570952415, -0.044217053800821304, 0.0756053775548935, 0.15034158527851105, 0.2433532476425171, -0.08203703165054321, 0.02304146997630596, 0.024648992344737053, -0.0337817445397377, 0.05557617172598839, -0.04032764211297035, 0.035090215504169464, 0.12478529661893845, 0.004215634893625975, 0.13826382160186768, 0.09605827182531357, -0.04657337814569473, 0.0030391195323318243, 0.05355951935052872, -0.08573099225759506, -0.14157317578792572, -0.01880762353539467, 0.0019446497317403555, -0.1138821691274643, 0.020692890509963036, 0.10174418240785599, -0.0892946645617485, -0.013262463733553886, -0.0032219053246080875, 0.02501961588859558, -0.09369103610515594, 0.22017470002174377, 0.09733571112155914, 0.05290434509515762, -0.09730291366577148, 0.08523500710725784, 0.04499555379152298, -0.043397437781095505, 0.010509721003472805, 0.10462570935487747, -0.10726543515920639, -0.06897460669279099, 0.0892377495765686, 0.1885402649641037, -0.1530323028564453, -0.06976880133152008, -0.1289544254541397, -0.11380337923765182, 0.07773958891630173, 0.15160603821277618, 0.09144765883684158, 0.07701360434293747, -0.0656939297914505, 0.016573922708630562, -0.11353977024555206, 0.1016751378774643, 0.06720355153083801, 0.049302197992801666, -0.1322811245918274, 0.099110446870327, 0.003566241357475519, 0.09203540533781052, -0.07142128795385361, 0.0054085273295640945, -0.12737837433815002, 0.034311242401599884, -0.2060500830411911, -0.00030303001403808594, -0.0551842525601387, 0.005139456130564213, 0.01734614185988903, -0.07627242058515549, -0.04556865617632866, 0.019692935049533844, -0.0688762441277504, -0.020775925368070602, 0.014561887830495834, 0.035775959491729736, -0.0999562218785286, -0.0004394659190438688, 0.04538954049348831, -0.029939938336610794, 0.07483220845460892, 0.001428981893695891, -0.02675367332994938, 0.03792817145586014, -0.18681563436985016, -0.001039114547893405, -0.007420445792376995, 0.03774894401431084, 0.07395947724580765, -0.06179565191268921, -0.010457775555551052, 0.042496614158153534, 0.053494539111852646, 0.022436272352933884, 0.05141596496105194, -0.11462553590536118, 0.016801370307803154, -0.0639110878109932, -0.08435189723968506, -0.047968070954084396, 0.07729014754295349, 0.09763096272945404, 0.04248674586415291, 0.17284244298934937, -0.06831023097038269, 0.04625464230775833, -0.15592193603515625, 0.009414478205144405, -0.031628042459487915, -0.039290305227041245, -0.07316884398460388, -0.05073119327425957, 0.0752745121717453, -0.03116803988814354, 0.19777195155620575, 0.07888496667146683, 0.030513351783156395, 0.05602467805147171, 0.011514068581163883, 0.002462281147018075, 0.03898591548204422, 0.18071416020393372, 0.03080451302230358, 0.01382013875991106, -0.04692430421710014, 0.05072228983044624, 0.024507731199264526, 0.08970553427934647, 0.18094711005687714, 0.15661297738552094, 0.008748377673327923, 0.05365031585097313, 0.024870753288269043, -0.067623570561409, -0.11109030246734619, -0.004591523669660091, -0.00027068148483522236, 0.043010082095861435, -0.06843601167201996, 0.11790981143712997, 0.14141304790973663, -0.13536277413368225, 0.058610230684280396, -0.053751878440380096, -0.10642894357442856, -0.13397760689258575, -0.0663730651140213, -0.07585978507995605, -0.1137404590845108, -0.013699617236852646, -0.1237296462059021, -0.013454846106469631, 0.12341836094856262, 0.03441156819462776, -0.04760979861021042, 0.12829241156578064, -0.00705495523288846, 0.026582226157188416, 0.012981663458049297, -0.013963790610432625, 0.030149858444929123, -0.05452537164092064, -0.0297794621437788, -0.07179244607686996, -0.0030071022920310497, 0.03028886765241623, -0.03208712488412857, -0.0851990357041359, 0.021470382809638977, 0.008051896467804909, -0.07312581688165665, 0.024246273562312126, 0.024835975840687752, 0.009607337415218353, 0.18791402876377106, 0.01991915889084339, -0.016445482149720192, 0.028539495542645454, 0.2615737020969391, -0.10648505389690399, -0.20754452049732208, -0.12883709371089935, 0.25511378049850464, 0.031299758702516556, 0.020729640498757362, 0.012028751894831657, -0.0710618644952774, 0.006878909654915333, 0.24165166914463043, 0.26277774572372437, -0.09976308047771454, -0.012807881459593773, 0.02672220580279827, -0.002398729557171464, -0.023846037685871124, 0.11361595243215561, 0.06684716045856476, 0.17936860024929047, -0.11837329715490341, -0.0007952505256980658, -0.02087637037038803, -0.04915761202573776, -0.05911298096179962, -0.014680788852274418, 0.05054372921586037, -0.008349857293069363, -0.03489626199007034, 0.08874727040529251, -0.12592223286628723, -0.07775694131851196, 0.04105484113097191, -0.06732208281755447, -0.10227140784263611, -0.037514828145504, 0.06591308116912842, -0.02542666718363762, 0.04728492721915245, -0.018355049192905426, 0.013808888383209705, 0.06534380465745926, 0.004783510696142912, -0.11367562413215637, -0.03654180094599724, 0.1393337994813919, 0.03373923525214195, 0.1582101434469223, -0.07385646551847458, 0.08284077793359756, 0.14492549002170563, 0.036156002432107925, -0.09107045829296112, 0.07622743397951126, 0.021444043144583702, -0.04935585334897041, 0.08578898012638092, 0.03437269851565361, -0.012750142253935337, 0.03670209273695946, -0.0011409485014155507, -0.12078797072172165, 0.05101893097162247, -0.049322109669446945, -0.05345267802476883, -0.10651488602161407, 0.004479869734495878, -0.0821581482887268, 0.1484202742576599, 0.1406283676624298, -0.04644530639052391, 0.015441715717315674, -0.08266929537057877, 0.01961214281618595, -0.0014679738087579608, 0.0468582883477211, -0.060288675129413605, -0.19729100167751312, 0.0008638479048386216, 0.04804044961929321, -0.009302983060479164, -0.20952564477920532, -0.04879879951477051, -0.023338234052062035, -0.053882893174886703, -0.08683235198259354, 0.10857568681240082, 0.0919693112373352, 0.05789158493280411, -0.05356569588184357, -0.018952222540974617, -0.07095309346914291, 0.13617417216300964, -0.1869436800479889, -0.08289331197738647 ]
null
null
null
# TorchScript model of distilbert-squad ## Model description A serialized torchscript model of distilbert-squad with a config.pbtxt for deployment using NVIDIA Triton Inference Server.
{"language": ["en"], "license": "mit", "tags": ["question-answering", "torchscript", "FastNN"], "datasets": ["squad"]}
question-answering
aychang/distilbert-squad
[ "question-answering", "torchscript", "FastNN", "en", "dataset:squad", "license:mit", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us
# TorchScript model of distilbert-squad ## Model description A serialized torchscript model of distilbert-squad with a URL for deployment using NVIDIA Triton Inference Server.
[ "# TorchScript model of distilbert-squad", "## Model description\n\nA serialized torchscript model of distilbert-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ "TAGS\n#question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us \n", "# TorchScript model of distilbert-squad", "## Model description\n\nA serialized torchscript model of distilbert-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ 33, 12, 34 ]
[ "passage: TAGS\n#question-answering #torchscript #FastNN #en #dataset-squad #license-mit #region-us \n# TorchScript model of distilbert-squad## Model description\n\nA serialized torchscript model of distilbert-squad with a URL for deployment using NVIDIA Triton Inference Server." ]
[ -0.027889307588338852, -0.14914438128471375, -0.003942803014069796, 0.10985821485519409, 0.07756263762712479, 0.09827492386102676, 0.26733341813087463, 0.001415330683812499, 0.12559135258197784, -0.01405874639749527, 0.10653166472911835, 0.10988076776266098, -0.00759201729670167, -0.10254894942045212, 0.015343695878982544, -0.1481126993894577, 0.020422181114554405, -0.04069698229432106, 0.0026219207793474197, 0.04528820887207985, 0.08797299116849899, -0.0551891028881073, 0.09426706284284592, -0.04094265028834343, -0.021215027198195457, -0.026250746101140976, -0.01626678556203842, 0.03673005849123001, 0.12700749933719635, 0.04458673670887947, 0.0704936534166336, 0.0036950234789401293, 0.0633058249950409, -0.09942507743835449, 0.04684080556035042, -0.0053033242002129555, -0.052076179534196854, 0.02963295392692089, -0.028959311544895172, 0.041717492043972015, 0.2212745100259781, 0.04217188432812691, 0.06957817077636719, 0.07767601311206818, -0.11795523017644882, 0.058790694922208786, -0.057501308619976044, -0.15258604288101196, 0.10763350129127502, 0.04119757190346718, 0.018216446042060852, 0.13790804147720337, -0.22520025074481964, 0.07040542364120483, -0.000495858839713037, -0.2428891360759735, -0.1200680285692215, 0.08682557195425034, 0.14109766483306885, 0.06205059960484505, -0.04062287136912346, 0.07797878235578537, 0.008663566783070564, 0.0678730309009552, 0.0039327023550868034, -0.0788978561758995, -0.24429269134998322, 0.02545166015625, -0.11120811104774475, 0.005221793428063393, 0.24988147616386414, 0.05754034221172333, 0.03872857615351677, 0.001964468276128173, -0.08967317640781403, 0.11832791566848755, -0.05159271880984306, -0.017065295949578285, -0.01780875213444233, 0.0005907799350097775, 0.048996634781360626, -0.039589766412973404, -0.05203981697559357, -0.14993003010749817, -0.09580746293067932, 0.019986592233181, 0.0239845123142004, 0.1069134846329689, -0.21271276473999023, 0.023322448134422302, 0.09667891263961792, -0.07522305101156235, -0.02757449634373188, -0.12391816079616547, -0.09278751909732819, 0.02122734859585762, 0.03147042542695999, -0.09321665018796921, 0.19820988178253174, 0.16519996523857117, 0.17734992504119873, 0.007729832082986832, -0.0619075708091259, 0.02591615915298462, -0.012673218734562397, 0.02747909165918827, 0.007782758213579655, -0.029263334348797798, -0.0010485287057235837, -0.01839521713554859, 0.04975775256752968, -0.034620560705661774, -0.11922359466552734, 0.010884481482207775, -0.07881925255060196, 0.08962413668632507, 0.09338584542274475, 0.07400088012218475, -0.0371408686041832, -0.026118997484445572, 0.13993427157402039, -0.014257977716624737, -0.0055084978230297565, 0.020467810332775116, -0.020119432359933853, -0.02331658825278282, -0.006671091541647911, -0.031543292105197906, 0.036905352026224136, -0.048687275499105453, -0.059109918773174286, -0.0552382692694664, -0.03961630538105965, -0.0768112987279892, 0.061968546360731125, -0.040438927710056305, -0.011526994407176971, -0.1330243945121765, -0.10367561876773834, 0.07510703802108765, 0.02726217359304428, -0.028537122532725334, 0.04825372248888016, -0.13687078654766083, -0.016899986192584038, 0.061287783086299896, 0.02514719031751156, 0.041426897048950195, -0.05148962140083313, 0.01835932768881321, 0.015790339559316635, 0.07289814203977585, -0.20981624722480774, -0.023590683937072754, -0.11382915824651718, 0.01301457080990076, 0.04203501343727112, 0.0434814915060997, -0.05942594259977341, 0.08161895722150803, -0.008259885013103485, -0.021990474313497543, -0.05855171009898186, -0.005448483396321535, 0.004333359654992819, 0.20203495025634766, -0.25907155871391296, 0.013849114067852497, 0.16639749705791473, -0.04731087386608124, -0.22954365611076355, 0.1202569380402565, -0.021503936499357224, -0.08637307584285736, 0.04341107979416847, 0.06889091432094574, 0.1176941841840744, -0.14950907230377197, 0.06223328784108162, 0.07823122292757034, -0.18709301948547363, -0.03387461230158806, 0.10180731862783432, 0.037075236439704895, -0.02932809107005596, 0.062334589660167694, 0.06165323406457901, 0.028196459636092186, -0.08296862989664078, -0.07395559549331665, -0.042695097625255585, -0.04298393055796623, 0.1166907474398613, 0.016605406999588013, -0.04032538831233978, -0.09097098559141159, -0.03732563555240631, 0.013541297987103462, 0.09077901393175125, 0.0723632276058197, -0.053476374596357346, -0.12021823972463608, 0.0390101782977581, -0.01778559200465679, 0.04205593466758728, -0.08552159368991852, -0.3221578299999237, 0.035871583968400955, 0.15806257724761963, 0.02508602850139141, 0.16727106273174286, 0.04989689961075783, 0.04623156040906906, 0.009587679989635944, 0.01770806312561035, -0.018277958035469055, 0.017811566591262817, 0.0018575998255982995, -0.1774948686361313, -0.06802334636449814, -0.0746445432305336, 0.04086367413401604, -0.15649983286857605, 0.020505700260400772, -0.04051139950752258, 0.012621779926121235, -0.06880717724561691, 0.06982049345970154, 0.00047622036072425544, -0.021443581208586693, -0.08532922714948654, -0.05016595497727394, 0.07333844155073166, -0.03814268112182617, -0.032291434705257416, -0.04631362482905388, 0.052004631608724594, -0.08551381528377533, 0.0906950831413269, -0.08824964612722397, 0.058916449546813965, 0.05775708332657814, -0.06529112905263901, -0.012555383145809174, -0.005288323853164911, -0.061622943729162216, -0.05074097588658333, 0.03901336342096329, -0.012280705384910107, -0.039410561323165894, 0.05122768133878708, 0.029351569712162018, -0.022656844928860664, 0.010149921290576458, 0.04930594563484192, 0.21000251173973083, -0.1124168112874031, 0.003018580377101898, 0.0691012516617775, 0.0677013024687767, 0.062291838228702545, -0.006429985165596008, 0.019345467910170555, -0.03499189764261246, -0.008667519316077232, 0.033578403294086456, 0.1490645706653595, -0.21081626415252686, 0.03726503252983093, 0.026149960234761238, -0.009571416303515434, -0.04556792974472046, -0.07792650163173676, -0.045596878975629807, -0.045885443687438965, -0.06135174259543419, -0.09141208231449127, 0.0928737074136734, 0.010528848506510258, 0.07857735455036163, -0.001196402939967811, -0.1428970843553543, 0.03551797196269035, -0.07323159277439117, -0.06838800013065338, 0.1445114016532898, -0.07247284799814224, -0.07274540513753891, -0.09590169042348862, -0.018963448703289032, -0.1307857185602188, 0.048096686601638794, 0.1013648733496666, -0.08758191019296646, 0.0553717277944088, -0.02816202864050865, 0.047300487756729126, -0.03908037021756172, 0.02451782301068306, -0.07543279230594635, 0.016211586073040962, -0.08239626884460449, -0.1344844549894333, -0.048914771527051926, -0.06597312539815903, -0.11553486436605453, 0.07713450491428375, -0.10032027214765549, 0.12363897264003754, 0.11320462822914124, 0.048388298600912094, -0.0091200340539217, -0.052000414580106735, 0.1148662343621254, 0.01259118877351284, 0.1067759320139885, 0.1579531729221344, 0.053090911358594894, 0.05126446112990379, 0.10629002004861832, 0.060312237590551376, -0.08048456907272339, 0.050835996866226196, 0.003357997862622142, -0.10816080123186111, -0.1810975819826126, -0.10293522477149963, -0.1320161521434784, 0.09934429824352264, 0.024914460256695747, 0.04660210385918617, 0.017298586666584015, 0.07431143522262573, -0.0017819397617131472, -0.041945043951272964, 0.017601730301976204, 0.05178888142108917, 0.15181037783622742, 0.01793760061264038, 0.06031063571572304, -0.09495092928409576, -0.006151981186121702, 0.10022691637277603, -0.05334167927503586, 0.16142764687538147, 0.0976865142583847, 0.058340538293123245, 0.16573180258274078, 0.06628860533237457, 0.026342228055000305, 0.0913398489356041, 0.038157444447278976, -0.014602376148104668, -0.019498290494084358, -0.0735672116279602, -0.04133189469575882, -0.0017893726471811533, -0.004651803523302078, 0.07794343680143356, -0.026060672476887703, 0.16020531952381134, 0.09612993150949478, 0.2190648466348648, 0.06875687092542648, -0.2942046523094177, -0.04875684902071953, 0.0177653506398201, 0.0408751554787159, -0.026349538937211037, 0.04127894341945648, 0.04845034331083298, -0.08504603058099747, 0.013557670637965202, -0.016700157895684242, 0.10447683930397034, 0.048380687832832336, 0.03799006715416908, 0.09462505578994751, 0.07932423800230026, 0.03972797840833664, 0.11259595304727554, -0.27969253063201904, 0.08170703798532486, 0.006685651373118162, 0.07431568205356598, -0.07125955820083618, -0.04038757085800171, 0.0023470777086913586, 0.07545740157365799, 0.0821516141295433, 0.02546025812625885, 0.09497834742069244, -0.002685546875, -0.033899419009685516, 0.07893380522727966, 0.05374522879719734, 0.03306468203663826, 0.030324053019285202, -0.02139025367796421, 0.004620095249265432, 0.03529173880815506, 0.008711738511919975, -0.09273449331521988, -0.08228333294391632, -0.02526657097041607, 0.07311493158340454, -0.10417214781045914, -0.04780649021267891, 0.004500406328588724, 0.07605693489313126, 0.00928011629730463, -0.08403957635164261, -0.037703998386859894, -0.066202811896801, -0.06004784256219864, 0.0889400839805603, -0.04645718261599541, 0.032420940697193146, -0.10715928673744202, -0.013591162860393524, -0.023698706179857254, -0.11751537024974823, 0.05722106248140335, -0.11694413423538208, -0.0048236786387860775, -0.12600822746753693, 0.161687433719635, 0.02676304802298546, 0.013496463187038898, 0.04603077098727226, -0.01850847899913788, -0.00433716457337141, -0.13008937239646912, -0.007882297970354557, 0.05406082794070244, -0.06896912306547165, -0.0394744798541069, -0.0927734524011612, 0.11591952294111252, -0.024600330740213394, 0.05983886122703552, 0.06500855833292007, 0.16103526949882507, -0.0851956233382225, 0.020435955375432968, 0.04965545982122421, -0.04242228716611862, -0.12201641499996185, -0.0830598995089531, -0.05068087577819824, -0.07266132533550262, 0.03437947481870651, -0.10998854041099548, 0.13810154795646667, 0.05917748808860779, 0.019182996824383736, 0.2651402950286865, -0.3071798086166382, -0.045645926147699356, 0.06949969381093979, 0.05007902905344963, 0.08404622226953506, -0.12227574735879898, -0.04627039283514023, -0.04923293739557266, -0.2849061191082001, 0.10025911033153534, -0.06248382106423378, 0.06910745054483414, -0.06521739065647125, 0.03406199440360069, -0.02397925592958927, -0.015828188508749008, 0.07612598687410355, 0.07077737897634506, 0.08476198464632034, -0.06685949116945267, -0.09644433110952377, 0.28764674067497253, -0.03672589734196663, 0.06285294145345688, -0.09947343170642853, 0.04683978855609894, -0.14216290414333344, 0.017186269164085388, -0.07044307887554169, 0.052787475287914276, -0.05788528174161911, -0.0436953604221344, -0.049507174640893936, -0.031113039702177048, -0.016936810687184334, -0.013598957099020481, 0.03407660499215126, 0.04459155723452568, 0.0019326067995280027, -0.09213624894618988, 0.11279568076133728, -0.06395068019628525, -0.07371529191732407, -0.07917121797800064, 0.015968244522809982, 0.04721983149647713, -0.03455646336078644, -0.01111211534589529, 0.1176077276468277, 0.028826311230659485, 0.016167104244232178, 0.0989251509308815, 0.013105416670441628, 0.05226370692253113, 0.06003883481025696, -0.09877852350473404, -0.2698546350002289, 0.07599049806594849, 0.06431609392166138, -0.06743664294481277, 0.10784145444631577, 0.049538761377334595, 0.017013737931847572, 0.008814679458737373, 0.036855071783065796, 0.029199447482824326, -0.07205529510974884, 0.020897137001156807, 0.0563935786485672, 0.08865150064229965, -0.06218693405389786, 0.09669630974531174, -0.00043041579192504287, 0.07476874440908432, -0.06028084084391594, 0.10348675400018692, -0.10853453725576401, -0.09357154369354248, -0.008711682632565498, 0.14021849632263184, -0.03698762133717537, -0.034682657569646835, -0.04088040068745613, -0.11038181930780411, 0.003907287493348122, 0.09073478728532791, 0.038257699459791183, 0.08266133069992065, -0.05304335430264473, 0.0402517095208168, -0.027686260640621185, 0.05460579693317413, -0.13930590450763702, 0.047479353845119476, -0.029772667214274406, -0.13098040223121643, -0.008857352659106255, 0.1091194599866867, -0.07681558281183243, -0.0631130114197731, -0.10867589712142944, 0.04599132761359215, -0.1921614408493042, -0.012389488518238068, -0.09336350113153458, 0.007424678187817335, -0.021850768476724625, -0.05827546864748001, -0.03801920637488365, 0.000539436936378479, 0.04946963116526604, 0.0022508767433464527, 0.02630329690873623, 0.10359220206737518, -0.1830979585647583, -0.038395680487155914, 0.01585540361702442, -0.04008976370096207, 0.05789346992969513, 0.09521769732236862, -0.02641345001757145, 0.04598580300807953, -0.1256261169910431, -0.04039902240037918, 0.0038227641489356756, 0.0833447128534317, 0.099972665309906, 0.06401890516281128, -0.034891385585069656, 0.0276276133954525, 0.05518737807869911, -0.01624138653278351, -0.026528434827923775, -0.05667949840426445, 0.03826853632926941, 0.00011192740930709988, -0.06766592711210251, -0.03722639009356499, -0.0322914682328701, 0.10265406966209412, 0.06932089477777481, 0.15374000370502472, -0.028458306565880775, 0.08247997611761093, -0.1449730098247528, -0.023451989516615868, 0.005785513669252396, -0.04314953833818436, -0.000010217052476946265, -0.09869121015071869, 0.025071945041418076, -0.05803168937563896, 0.13136065006256104, 0.0037670223973691463, 0.04073946923017502, -0.009098236449062824, 0.11218181997537613, -0.0037437279243022203, -0.02223430760204792, 0.14542509615421295, 0.03215621039271355, -0.03482351452112198, -0.03327074646949768, 0.004220261704176664, 0.041558463126420975, 0.049140386283397675, 0.22006581723690033, 0.1415327489376068, 0.05695213004946709, 0.025611963123083115, 0.17701356112957, 0.06756468862295151, -0.07349739968776703, -0.17346957325935364, 0.09838613867759705, 0.021505113691091537, -0.053839653730392456, 0.09711716324090958, 0.07451430708169937, -0.0166253000497818, 0.023097869008779526, -0.021680474281311035, -0.0499352291226387, -0.04089956730604172, -0.09848998486995697, -0.029872193932533264, -0.06249123439192772, 0.003916263114660978, -0.08568252623081207, -0.06257881224155426, 0.12993060052394867, 0.0658489540219307, -0.05564494431018829, 0.15682771801948547, 0.11736540496349335, 0.07547730952501297, -0.001192597090266645, -0.020710013806819916, -0.033486507833004, 0.051444508135318756, 0.04140128567814827, -0.023469867184758186, -0.006656466983258724, -0.0282585471868515, 0.02142932638525963, -0.0496722012758255, 0.01847897283732891, -0.08169816434383392, -0.07521301507949829, 0.002780809300020337, 0.03194032981991768, -0.12832878530025482, 0.02852451428771019, 0.0911361500620842, -0.040018126368522644, 0.06387530267238617, 0.06110098212957382, 0.01819922775030136, -0.2255008965730667, -0.012928864918649197, 0.13141290843486786, -0.12470601499080658, 0.022790858522057533, -0.09040902554988861, -0.003535177791491151, -0.05528469383716583, 0.14754721522331238, 0.25613781809806824, -0.11456030607223511, -0.01878749020397663, -0.050364743918180466, 0.008927075192332268, -0.023735739290714264, 0.11795197427272797, -0.018894607201218605, 0.11464615911245346, -0.008961032144725323, -0.03523901849985123, -0.07394787669181824, 0.021658089011907578, -0.02866204082965851, -0.08246757835149765, 0.05538560077548027, 0.03742469474673271, -0.04243497923016548, 0.04668344184756279, -0.07751528173685074, 0.07700135558843613, 0.0076846168376505375, 0.04753430187702179, -0.03066280670464039, -0.061722200363874435, 0.01566058024764061, 0.0138796865940094, 0.12302201241254807, -0.09471753239631653, -0.04835763946175575, 0.051997747272253036, -0.01921292394399643, -0.15495169162750244, -0.10542451590299606, 0.10121522843837738, -0.05158786475658417, 0.1566932499408722, -0.09292979538440704, 0.10221754759550095, 0.0914207324385643, 0.03688950836658478, -0.08606330305337906, 0.05879398062825203, 0.07558655738830566, 0.05330095440149307, 0.007511400617659092, -0.1617179661989212, 0.007075711153447628, 0.005286698695272207, 0.05735253170132637, 0.01343591045588255, 0.02869938127696514, 0.05875319615006447, -0.05470499396324158, -0.12461341172456741, 0.05688195675611496, -0.04499955475330353, 0.094942606985569, 0.06464900821447372, -0.05684368312358856, -0.040492862462997437, -0.047138866037130356, 0.01523598562926054, 0.039646733552217484, 0.027703534811735153, -0.11594025045633316, -0.08547275513410568, -0.030610481277108192, 0.04598120227456093, 0.010304123163223267, -0.16633203625679016, -0.06331318616867065, -0.14803779125213623, 0.012253683060407639, 0.0651986375451088, 0.008978914469480515, 0.1508541703224182, 0.06418892741203308, -0.05119936913251877, -0.18827277421951294, -0.026899566873908043, 0.15957245230674744, -0.12329021841287613, -0.07662899047136307 ]
null
null
null
# TorchScript model of faster-rcnn ## Model description A serialized torchscript model of [faster-rcnn](https://pytorch.org/vision/stable/models.html#faster-r-cnn) with a config.pbtxt for deployment using NVIDIA Triton Inference Server.
{"language": ["en"], "license": "mit", "tags": ["object-detection", "torchscript", "FastNN"], "datasets": ["coco"]}
object-detection
aychang/fasterrcnn-resnet50-cpu
[ "object-detection", "torchscript", "FastNN", "en", "dataset:coco", "license:mit", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #object-detection #torchscript #FastNN #en #dataset-coco #license-mit #region-us
# TorchScript model of faster-rcnn ## Model description A serialized torchscript model of faster-rcnn with a URL for deployment using NVIDIA Triton Inference Server.
[ "# TorchScript model of faster-rcnn", "## Model description\n\nA serialized torchscript model of faster-rcnn with a URL for deployment using NVIDIA Triton Inference Server." ]
[ "TAGS\n#object-detection #torchscript #FastNN #en #dataset-coco #license-mit #region-us \n", "# TorchScript model of faster-rcnn", "## Model description\n\nA serialized torchscript model of faster-rcnn with a URL for deployment using NVIDIA Triton Inference Server." ]
[ 32, 11, 33 ]
[ "passage: TAGS\n#object-detection #torchscript #FastNN #en #dataset-coco #license-mit #region-us \n# TorchScript model of faster-rcnn## Model description\n\nA serialized torchscript model of faster-rcnn with a URL for deployment using NVIDIA Triton Inference Server." ]
[ -0.0022977953776717186, -0.09428861737251282, -0.0034128765109926462, 0.12409916520118713, 0.03179033100605011, 0.023235918954014778, 0.0858667865395546, 0.03513389825820923, 0.051206931471824646, 0.0022325559984892607, 0.15303246676921844, 0.21949857473373413, -0.055671997368335724, -0.07294122874736786, 0.023022418841719627, -0.17709043622016907, 0.07629887014627457, -0.0692356675863266, 0.04982716217637062, 0.07480145990848541, 0.08044596016407013, -0.02425803430378437, 0.126075878739357, -0.037210263311862946, -0.06329876184463501, -0.029434887692332268, -0.041958145797252655, -0.04124605655670166, 0.07676149159669876, 0.002894978504627943, 0.03776905685663223, 0.013978849165141582, 0.0975530743598938, -0.09517829120159149, 0.046989381313323975, 0.01752632111310959, -0.029279641807079315, 0.06777140498161316, -0.009458490647375584, 0.05639561265707016, 0.36267250776290894, 0.06508330255746841, 0.03232741355895996, 0.061968207359313965, -0.045737430453300476, 0.05544291436672211, 0.023799454793334007, -0.0877060815691948, -0.0671427845954895, 0.05242904648184776, 0.04449295997619629, 0.16661368310451508, -0.2179742306470871, 0.07226768881082535, -0.002898523351177573, -0.2556236684322357, -0.13298840820789337, 0.1641075313091278, 0.18071497976779938, 0.10639245063066483, -0.032188933342695236, 0.10627119243144989, -0.043406225740909576, 0.05471073463559151, 0.06025402620434761, 0.0036061087157577276, -0.246182382106781, 0.07772128283977509, -0.12013526260852814, 0.005790361203253269, 0.08239151537418365, 0.08299097418785095, 0.07840665429830551, -0.013241641223430634, -0.1659964621067047, -0.035612232983112335, -0.06074005737900734, 0.004207605961710215, -0.011826487258076668, 0.04122539982199669, -0.05800711736083031, -0.043979693204164505, -0.008417895063757896, -0.12022756785154343, -0.08477061241865158, 0.10240370035171509, 0.05173984542489052, 0.16112393140792847, -0.09568872302770615, 0.04367667809128761, 0.24662111699581146, -0.09854820370674133, -0.061814289540052414, -0.11895520985126495, -0.04046522453427315, 0.04320620372891426, 0.04305502027273178, 0.08150267601013184, 0.17040115594863892, 0.06440674513578415, 0.20485180616378784, 0.011415028013288975, -0.06943966448307037, 0.04920664802193642, -0.12573084235191345, 0.056299708783626556, -0.1098402887582779, 0.015400050207972527, 0.017693260684609413, -0.01760968007147312, 0.07394587993621826, -0.0406106561422348, -0.12652304768562317, 0.007549452129751444, -0.09268251806497574, 0.08844786882400513, 0.007755345199257135, 0.09382835775613785, 0.03570282459259033, -0.027122298255562782, 0.18685726821422577, 0.016006391495466232, -0.02159358188509941, 0.07452189177274704, -0.10974659025669098, 0.011289923451840878, 0.00774882547557354, -0.06187940016388893, 0.01498046051710844, -0.10109544545412064, -0.04519132152199745, -0.045504502952098846, -0.036205463111400604, -0.07635969668626785, 0.05529945716261864, 0.0512707494199276, -0.007700825110077858, -0.20254787802696228, -0.09950913488864899, 0.10439144819974899, 0.05078727751970291, -0.0488080158829689, 0.033049535006284714, -0.1472596973180771, -0.09732364863157272, 0.07751462608575821, 0.042923759669065475, 0.10348895937204361, -0.09500432014465332, 0.01226448267698288, -0.02337259240448475, 0.10236421972513199, -0.09106317907571793, -0.05917179957032204, -0.12321370095014572, -0.04869655519723892, 0.12098667770624161, 0.05846560001373291, -0.09942825883626938, 0.08451955020427704, -0.05377747863531113, 0.0029650337528437376, -0.1850394457578659, 0.00021635906887240708, -0.05321888253092766, 0.14330962300300598, -0.17639504373073578, 0.034627482295036316, 0.2075406312942505, 0.0329011008143425, -0.1263103485107422, 0.1072913184762001, -0.03315374627709389, -0.10447052121162415, 0.030519891530275345, -0.024008501321077347, 0.02765710838139057, -0.17380192875862122, 0.10764061659574509, 0.15471521019935608, -0.2542114853858948, -0.05382680147886276, 0.10925550758838654, 0.1058148741722107, -0.12018877267837524, 0.04443344101309776, 0.13155728578567505, 0.09107048064470291, -0.10398325324058533, -0.06915826350450516, -0.04603084921836853, -0.07641104608774185, 0.0657256543636322, -0.012139827013015747, -0.026642227545380592, -0.048670459538698196, -0.13277685642242432, -0.1529269814491272, 0.1509212851524353, 0.044562794268131256, -0.06528807431459427, -0.11209893971681595, -0.033102720975875854, -0.07787560671567917, 0.009182806126773357, -0.0627676397562027, -0.22913414239883423, 0.08260329812765121, 0.0776529386639595, 0.006076175719499588, 0.0033263955265283585, 0.0589044950902462, 0.07798264920711517, 0.07325238734483719, 0.05776005610823631, 0.02554175816476345, 0.013293862342834473, -0.0077054426074028015, -0.0998757928609848, -0.028365734964609146, -0.03405648097395897, 0.10231438279151917, -0.21249304711818695, 0.015142152085900307, -0.03777310997247696, -0.013066206127405167, -0.0577399805188179, 0.07130119204521179, -0.004404483828693628, -0.07045280933380127, -0.06602180004119873, -0.07457147538661957, 0.04568729177117348, -0.002973278285935521, 0.0019116114126518369, 0.15883773565292358, 0.012254222296178341, -0.009855174459517002, 0.1022815927863121, -0.07069529592990875, 0.03789033368229866, 0.13420376181602478, -0.07690414041280746, -0.008992382325232029, -0.026951409876346588, 0.005584070924669504, -0.026609724387526512, 0.02221645601093769, 0.07733683288097382, -0.06318820267915726, -0.009180533699691296, 0.059354040771722794, 0.0004484999808482826, 0.011711533181369305, 0.10834863781929016, 0.24703553318977356, -0.17904487252235413, -0.0036337177734822035, -0.023861346766352654, 0.029517564922571182, 0.03355768695473671, 0.05489302799105644, -0.03945736214518547, 0.013799168169498444, -0.0057722050696611404, 0.03642519190907478, 0.03916950151324272, -0.06157742068171501, 0.03301673009991646, 0.01055951602756977, -0.0409814678132534, 0.018634578213095665, -0.07754284143447876, -0.027009617537260056, -0.007063694763928652, -0.06722693145275116, -0.024546919390559196, 0.0775584727525711, -0.004989733919501305, 0.07293232530355453, 0.009290676563978195, -0.25664570927619934, 0.004780594725161791, -0.045710477977991104, -0.01786591298878193, 0.13788405060768127, -0.008522659540176392, -0.0840533897280693, -0.13342764973640442, -0.08533192425966263, -0.040451303124427795, 0.027652421966195107, 0.06662876904010773, -0.09004781395196915, 0.032383717596530914, -0.07736378908157349, -0.030993618071079254, -0.07658688724040985, -0.016548825427889824, -0.04706971347332001, 0.06713152676820755, -0.11088792979717255, -0.1289110630750656, -0.04208198934793472, -0.0959622859954834, -0.08014235645532608, 0.01740826852619648, -0.15959486365318298, 0.16900353133678436, 0.16867320239543915, 0.02309485524892807, -0.047573741525411606, 0.03393423557281494, 0.025937262922525406, 0.010157197713851929, 0.061815809458494186, 0.1277027428150177, 0.0701526626944542, 0.03996172547340393, 0.011002830229699612, 0.07536870986223221, -0.03976975381374359, 0.02441389113664627, -0.013604513369500637, -0.09581408649682999, -0.14408335089683533, -0.1280776411294937, -0.11144039034843445, 0.024427803233265877, 0.05873753875494003, 0.11687318980693817, 0.016045454889535904, 0.049035269767045975, -0.0015973522095009685, -0.007106112781912088, 0.06617620587348938, 0.055593833327293396, 0.0701468288898468, 0.08526016026735306, 0.06285907328128815, -0.0969533622264862, -0.07138921320438385, 0.08654169738292694, -0.018388474360108376, 0.16513501107692719, 0.04820562154054642, 0.1551961600780487, 0.10077042132616043, 0.0643458217382431, 0.11213283985853195, 0.018088070675730705, 0.07168760150671005, 0.005476348102092743, -0.021087253466248512, -0.09065046161413193, 0.017305022105574608, -0.01471015065908432, -0.08585440367460251, 0.027141347527503967, 0.06548648327589035, 0.23872244358062744, 0.10009806603193283, 0.21189551055431366, 0.008462220430374146, -0.46346309781074524, -0.05237415060400963, -0.042998962104320526, 0.029260456562042236, -0.032623257488012314, 0.0755392387509346, 0.04227617755532265, -0.052396390587091446, -0.06031056493520737, -0.023123381659388542, 0.08010514825582504, -0.018268629908561707, 0.014238071627914906, 0.13021551072597504, 0.13551922142505646, 0.06389139592647552, 0.11787707358598709, -0.1335785686969757, 0.09264208376407623, 0.005835504736751318, 0.0671282634139061, -0.08743253350257874, -0.04068635031580925, 0.08002885431051254, 0.015628211200237274, 0.10826162993907928, -0.006029113195836544, -0.028517333790659904, -0.049310341477394104, -0.16971619427204132, 0.06450169533491135, 0.08062630891799927, 0.05267892777919769, 0.02929646521806717, -0.03004300966858864, -0.06403130292892456, 0.033441800624132156, -0.05741983652114868, -0.07389184087514877, -0.11378169059753418, -0.06539160758256912, 0.11857059597969055, -0.21343807876110077, -0.011210610158741474, -0.001827474683523178, 0.0805523693561554, 0.04441291093826294, 0.021641097962856293, 0.021136919036507607, -0.05201392248272896, -0.0926687940955162, 0.10751481354236603, -0.04937158524990082, 0.023579763248562813, -0.12619037926197052, 0.06110360473394394, 0.005019727163016796, -0.17183765769004822, 0.09877026081085205, -0.1014106422662735, 0.023186231032013893, -0.10711268335580826, 0.12079343944787979, 0.03496874123811722, -0.005672533996403217, 0.024096915498375893, -0.02852701023221016, -0.05439865589141846, -0.08662595599889755, -0.05159929767251015, 0.10820263624191284, -0.16638287901878357, -0.054874833673238754, -0.10514538735151291, 0.09615316987037659, -0.05601953715085983, 0.08057466894388199, 0.075246162712574, 0.09765014052391052, -0.09626679867506027, 0.016084784641861916, 0.08405160903930664, -0.017642471939325333, -0.1855447143316269, -0.12046505510807037, -0.06350207328796387, -0.07738446444272995, -0.11242660135030746, -0.11249254643917084, 0.175043985247612, 0.10324070602655411, -0.011335108429193497, 0.26542824506759644, -0.3761550486087799, -0.007668356876820326, 0.07518132776021957, 0.003190716030076146, 0.05496266484260559, -0.045525990426540375, -0.05906955152750015, -0.057949408888816833, -0.2861299216747284, 0.09866160899400711, -0.12603674829006195, 0.09374227374792099, -0.06898002326488495, -0.018927328288555145, -0.0038555199280381203, -0.039240214973688126, 0.032887157052755356, 0.009833965450525284, 0.06788771599531174, -0.05432341992855072, -0.0724225714802742, 0.1729498952627182, -0.05371605232357979, 0.1132950633764267, -0.06877053529024124, 0.07245539873838425, -0.03135765343904495, -0.017061257734894753, -0.0747499018907547, 0.017559872940182686, -0.0015955640701577067, -0.008074932731688023, -0.08780013024806976, -0.044489528983831406, 0.030938822776079178, -0.0039842380210757256, 0.06991413235664368, 0.08211872726678848, -0.0744553878903389, -0.03328273445367813, -0.0280458927154541, -0.04673795774579048, 0.039840202778577805, -0.10200484097003937, 0.009816506877541542, 0.0819869115948677, -0.11793074011802673, -0.015176634304225445, 0.09062188118696213, -0.049053944647312164, -0.03680496662855148, 0.1202019676566124, -0.025057625025510788, 0.023479914292693138, 0.03762304037809372, -0.04149871692061424, -0.08686355501413345, 0.000009091548236028757, 0.015437058173120022, 0.06197729334235191, 0.005763061810284853, 0.0419221855700016, -0.08042087405920029, 0.019829384982585907, 0.011349636130034924, 0.0348745621740818, -0.03370560705661774, -0.010297959670424461, 0.08022410422563553, 0.06256583333015442, -0.06572817265987396, 0.14211370050907135, 0.013870327733457088, 0.1439940631389618, -0.08559969067573547, 0.1478753685951233, -0.150639608502388, -0.06994004547595978, -0.01951787993311882, 0.12357578426599503, -0.08256153762340546, -0.025319311767816544, -0.023155026137828827, -0.051952920854091644, 0.06999190896749496, 0.1256020963191986, 0.04455868527293205, 0.05244975537061691, -0.06597308814525604, 0.038483332842588425, -0.08709731698036194, 0.05705047398805618, -0.16190935671329498, 0.032078735530376434, -0.03821532428264618, -0.1367437243461609, 0.0017591255018487573, 0.11648356914520264, -0.08263909071683884, -0.05872621014714241, -0.05955059081315994, 0.0548161156475544, -0.13837125897407532, 0.05821232125163078, -0.04483838751912117, 0.027297381311655045, -0.028874682262539864, 0.028749795630574226, -0.06367824971675873, 0.04551863297820091, 0.022909048944711685, -0.01212528720498085, -0.02696714922785759, 0.05448771268129349, -0.1407100260257721, -0.02724633738398552, -0.007726552430540323, -0.05650349706411362, 0.019904091954231262, 0.11175724118947983, -0.05030852183699608, 0.05964105948805809, -0.2361985445022583, -0.07985979318618774, 0.0626351460814476, 0.10966436564922333, 0.04849620908498764, 0.1328597366809845, 0.043669380247592926, -0.0055266800336539745, 0.023801341652870178, -0.0667395070195198, -0.012500856071710587, -0.05214681103825569, -0.0560501292347908, -0.050197359174489975, -0.0263958852738142, -0.0362967811524868, 0.01757081039249897, 0.06116470322012901, 0.0211383905261755, 0.12395094335079193, -0.02196415886282921, -0.012185567058622837, -0.09697138518095016, -0.030429139733314514, 0.005719145759940147, -0.09551799297332764, -0.01728363148868084, -0.05171683803200722, 0.059013642370700836, -0.08700810372829437, 0.22232405841350555, 0.026225198060274124, -0.010037804953753948, -0.029331840574741364, 0.05288178101181984, -0.08024756610393524, -0.013368360698223114, 0.1356087028980255, 0.12258755415678024, -0.003804667154327035, 0.04350489005446434, -0.007685364224016666, 0.14137466251850128, 0.06694298982620239, 0.09451920539140701, 0.18587392568588257, 0.17944400012493134, 0.10117010772228241, 0.20150209963321686, 0.0725945234298706, -0.09721589088439941, -0.13081665337085724, -0.020723087713122368, 0.04153714329004288, -0.04141048714518547, 0.00496893934905529, 0.13687743246555328, 0.0724097341299057, -0.05826753377914429, -0.021408459171652794, -0.005070619285106659, -0.12966635823249817, -0.10681167989969254, -0.08382423222064972, -0.050947435200214386, 0.03160042688250542, -0.014424851164221764, -0.017840122804045677, 0.20129722356796265, 0.0692000463604927, -0.003451879369094968, 0.07353267818689346, 0.12802934646606445, 0.03908417373895645, 0.03927072510123253, -0.05364086106419563, -0.0641583651304245, 0.09402411431074142, 0.007099136710166931, 0.0015592542476952076, -0.04557124152779579, 0.02730800211429596, 0.009887426160275936, -0.018840063363313675, 0.07487562298774719, -0.1249571293592453, -0.09261737018823624, 0.010232379660010338, 0.011711125262081623, -0.15274493396282196, -0.061428457498550415, 0.1387757658958435, -0.044546399265527725, 0.06877212971448898, 0.18181857466697693, 0.006844930350780487, -0.16696052253246307, -0.00010572456812951714, 0.14746646583080292, -0.02701001614332199, 0.0593993179500103, -0.09214475750923157, 0.025919467210769653, -0.08636460453271866, 0.2187294214963913, 0.22732222080230713, -0.14060638844966888, -0.033766571432352066, 0.042079221457242966, 0.0037804557941854, -0.007693347986787558, 0.15805435180664062, 0.0812404602766037, 0.1169295385479927, -0.04226655885577202, -0.11869757622480392, -0.056864503771066666, 0.022022685036063194, -0.06967075169086456, -0.039089541882276535, 0.009704504162073135, 0.0024695892352610826, -0.08811414241790771, 0.009723124094307423, -0.18093223869800568, 0.11912130564451218, -0.03761771321296692, 0.06486514955759048, -0.029812943190336227, -0.08144526928663254, -0.04617336019873619, 0.007118046283721924, 0.09125857800245285, -0.09625185281038284, -0.08207210153341293, -0.004650377202779055, 0.011523631401360035, -0.22973552346229553, -0.04639267921447754, -0.012435387820005417, 0.008193020708858967, 0.26787275075912476, -0.03868022561073303, 0.03368736803531647, 0.08823423087596893, 0.00408873101696372, -0.1247636154294014, 0.1535627394914627, 0.02422899752855301, -0.018488848581910133, 0.034218449145555496, -0.10849765688180923, -0.041281793266534805, 0.032370004802942276, 0.05035136640071869, 0.07142870873212814, 0.01504445355385542, 0.0003992572892457247, -0.016715995967388153, -0.0725940391421318, -0.06400305032730103, 0.03870759904384613, 0.04677949845790863, 0.09419836848974228, -0.01081803534179926, -0.010583572089672089, -0.07433384656906128, -0.0019911208655685186, 0.02374543435871601, -0.043880313634872437, -0.0750778317451477, -0.04292530566453934, -0.08861885964870453, 0.02582240104675293, 0.08945216983556747, -0.1478663682937622, -0.036361582577228546, -0.14315000176429749, -0.027423372492194176, 0.05708751454949379, 0.026358401402831078, 0.16761092841625214, 0.05897541344165802, -0.03682854026556015, -0.09904112666845322, 0.024389998987317085, 0.11099518835544586, -0.16021846234798431, -0.09730710089206696 ]
null
null
transformers
# IMDB Sentiment Task: roberta-base ## Model description A simple base roBERTa model trained on the "imdb" dataset. ## Intended uses & limitations #### How to use ##### Transformers ```python # Load model and tokenizer from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) # Use pipeline from transformers import pipeline model_name = "aychang/roberta-base-imdb" nlp = pipeline("sentiment-analysis", model=model_name, tokenizer=model_name) results = nlp(["I didn't really like it because it was so terrible.", "I love how easy it is to watch and get good results."]) ``` ##### AdaptNLP ```python from adaptnlp import EasySequenceClassifier model_name = "aychang/roberta-base-imdb" texts = ["I didn't really like it because it was so terrible.", "I love how easy it is to watch and get good results."] classifer = EasySequenceClassifier results = classifier.tag_text(text=texts, model_name_or_path=model_name, mini_batch_size=2) ``` #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data IMDB https://huggingface.co/datasets/imdb ## Training procedure #### Hardware One V100 #### Hyperparameters and Training Args ```python from transformers import TrainingArguments training_args = TrainingArguments( output_dir='./models', overwrite_output_dir=False, num_train_epochs=2, per_device_train_batch_size=8, per_device_eval_batch_size=8, warmup_steps=500, weight_decay=0.01, evaluation_strategy="steps", logging_dir='./logs', fp16=False, eval_steps=800, save_steps=300000 ) ``` ## Eval results ``` {'epoch': 2.0, 'eval_accuracy': 0.94668, 'eval_f1': array([0.94603457, 0.94731017]), 'eval_loss': 0.2578844428062439, 'eval_precision': array([0.95762642, 0.93624502]), 'eval_recall': array([0.93472, 0.95864]), 'eval_runtime': 244.7522, 'eval_samples_per_second': 102.144} ```
{"language": ["en"], "license": "mit", "tags": ["text-classification"], "datasets": ["imdb"]}
text-classification
aychang/roberta-base-imdb
[ "transformers", "pytorch", "jax", "roberta", "text-classification", "en", "dataset:imdb", "license:mit", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #roberta #text-classification #en #dataset-imdb #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us
# IMDB Sentiment Task: roberta-base ## Model description A simple base roBERTa model trained on the "imdb" dataset. ## Intended uses & limitations #### How to use ##### Transformers ##### AdaptNLP #### Limitations and bias This is minimal language model trained on a benchmark dataset. ## Training data IMDB URL ## Training procedure #### Hardware One V100 #### Hyperparameters and Training Args ## Eval results
[ "# IMDB Sentiment Task: roberta-base", "## Model description\n\nA simple base roBERTa model trained on the \"imdb\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nIMDB URL", "## Training procedure", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #text-classification #en #dataset-imdb #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# IMDB Sentiment Task: roberta-base", "## Model description\n\nA simple base roBERTa model trained on the \"imdb\" dataset.", "## Intended uses & limitations", "#### How to use", "##### Transformers", "##### AdaptNLP", "#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.", "## Training data\n\nIMDB URL", "## Training procedure", "#### Hardware\nOne V100", "#### Hyperparameters and Training Args", "## Eval results" ]
[ 57, 12, 21, 9, 5, 5, 5, 20, 6, 3, 6, 10, 4 ]
[ "passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #en #dataset-imdb #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us \n# IMDB Sentiment Task: roberta-base## Model description\n\nA simple base roBERTa model trained on the \"imdb\" dataset.## Intended uses & limitations#### How to use##### Transformers##### AdaptNLP#### Limitations and bias\n\nThis is minimal language model trained on a benchmark dataset.## Training data\n\nIMDB URL## Training procedure#### Hardware\nOne V100#### Hyperparameters and Training Args## Eval results" ]
[ -0.08453140407800674, 0.07960882037878036, -0.0009968917584046721, 0.06482906639575958, 0.16175708174705505, 0.02732716128230095, 0.16545242071151733, 0.13388845324516296, -0.005205670837312937, -0.013927618972957134, 0.07558269053697586, 0.1376921832561493, 0.043163374066352844, 0.12276215106248856, -0.06805366277694702, -0.24523748457431793, 0.03002762794494629, 0.014407459646463394, -0.0022818376310169697, 0.09968800097703934, 0.15930871665477753, -0.08220062404870987, 0.11442545056343079, -0.026760337874293327, -0.14646033942699432, -0.008239454589784145, 0.02830779179930687, -0.07772763818502426, 0.1022157371044159, 0.05679858475923538, 0.09015648066997528, 0.03138213977217674, 0.09413614869117737, -0.18378901481628418, 0.035009752959012985, 0.01001645065844059, -0.005787334404885769, 0.07644074410200119, 0.039933234453201294, -0.05278012529015541, 0.18187682330608368, 0.027533460408449173, 0.07671403139829636, 0.03370857983827591, -0.13368314504623413, -0.13796082139015198, -0.05251625180244446, -0.002770480699837208, 0.049962203949689865, 0.13799631595611572, -0.03335327282547951, 0.13527990877628326, -0.13839557766914368, 0.08470562845468521, 0.1471472829580307, -0.18933290243148804, -0.025754578411579132, 0.12710855901241302, 0.11100689321756363, 0.038135118782520294, -0.10049968957901001, 0.011696385219693184, 0.041778143495321274, 0.03911019489169121, 0.11284946650266647, -0.046936433762311935, -0.05216684937477112, 0.03856130689382553, -0.11866693198680878, -0.05339599400758743, 0.23770077526569366, 0.01846553385257721, -0.06324826180934906, -0.054543521255254745, 0.00034945947118103504, -0.07329238951206207, -0.06382372975349426, -0.001609224476851523, 0.005085606127977371, -0.0015408055623993278, -0.08525167405605316, -0.008487459272146225, -0.07996510714292526, -0.08008454740047455, -0.04417865723371506, 0.07651909440755844, 0.008651189506053925, 0.013005794025957584, -0.08668775856494904, 0.09750771522521973, 0.0043327463790774345, -0.10391788929700851, 0.0048165833577513695, -0.05584027245640755, 0.006077078636735678, -0.04612387716770172, -0.033710237592458725, 0.03127053752541542, 0.004826480057090521, 0.2363612800836563, 0.03024255484342575, 0.013572121039032936, 0.036433521658182144, 0.07171308249235153, 0.021662671118974686, 0.18059013783931732, -0.09037063270807266, -0.10207623243331909, 0.053472936153411865, 0.042374420911073685, -0.032738763839006424, -0.03206030651926994, -0.1334027647972107, -0.03963025286793709, 0.13013142347335815, 0.05363915488123894, 0.0252142995595932, 0.08802379667758942, -0.10465527325868607, -0.02096313051879406, 0.048814352601766586, -0.07157136499881744, 0.024087639525532722, -0.025417083874344826, -0.09830445051193237, -0.03290293738245964, 0.0644569918513298, 0.030056079849600792, -0.020625025033950806, 0.028377247974276543, -0.07570912688970566, 0.015828967094421387, -0.11952041834592819, -0.15034008026123047, 0.04633471742272377, -0.036651112139225006, 0.067084401845932, -0.1263476014137268, -0.17905981838703156, -0.020956778898835182, 0.07576531916856766, -0.0943748950958252, -0.04258246347308159, -0.059237539768218994, -0.054292723536491394, 0.04649752005934715, -0.03603208065032959, 0.11849047243595123, -0.0710647851228714, 0.03590486943721771, 0.03000224381685257, 0.09188590198755264, -0.023479850962758064, 0.060606375336647034, -0.11494027823209763, 0.0068353572860360146, -0.18859994411468506, 0.05350390076637268, -0.099215067923069, 0.06680739670991898, -0.08209573477506638, -0.059768110513687134, -0.00882417056709528, 0.053618527948856354, 0.016484811902046204, 0.17991051077842712, -0.1656726598739624, -0.046294257044792175, 0.0437699593603611, -0.07813991606235504, -0.14007703959941864, 0.12870270013809204, -0.07328718900680542, 0.13017913699150085, 0.024204133078455925, 0.11944631487131119, -0.034217510372400284, -0.10064923018217087, -0.02988896146416664, 0.11150315403938293, -0.043916307389736176, -0.0345773920416832, 0.11422659456729889, 0.06729216128587723, 0.000506020849570632, -0.0006179760675877333, 0.03688378632068634, 0.06892965734004974, -0.09387378394603729, -0.06861367076635361, 0.017048709094524384, -0.06729674339294434, 0.0113679813221097, 0.050142064690589905, 0.06686197966337204, -0.029051685705780983, -0.09582386910915375, 0.005518863443285227, 0.0821930319070816, 0.013971012085676193, 0.026120001450181007, -0.12672628462314606, 0.08572264760732651, -0.09796221554279327, -0.019717823714017868, -0.1934785544872284, 0.013202746398746967, -0.0413736030459404, 0.13874340057373047, 0.08007272332906723, 0.08987712115049362, 0.05883142352104187, -0.020367542281746864, -0.07243610173463821, -0.00001783788866305258, 0.06514544785022736, 0.04721841588616371, -0.07623466104269028, -0.241641566157341, -0.04405347630381584, -0.09803581982851028, 0.1701335310935974, -0.19334933161735535, 0.005345792509615421, -0.008779716677963734, 0.05937676131725311, 0.018859701231122017, 0.0018052068771794438, 0.012031566351652145, 0.029159942641854286, -0.07065721601247787, -0.06596207618713379, 0.06332046538591385, 0.019784899428486824, -0.053675759583711624, 0.10740496963262558, -0.12425994873046875, 0.1286243200302124, 0.1605803221464157, -0.14692135155200958, -0.08243564516305923, 0.02640838921070099, -0.0065419296734035015, -0.01052877027541399, -0.015648584812879562, 0.00676416140049696, 0.12328648567199707, -0.0013349649962037802, 0.13865159451961517, -0.05349777638912201, 0.000673533184453845, -0.011466064490377903, -0.08970220386981964, -0.03753521665930748, 0.10947389900684357, 0.07046417146921158, -0.1700267642736435, 0.13712120056152344, 0.060369133949279785, -0.058445677161216736, 0.11960650235414505, 0.0188708808273077, -0.02206411585211754, -0.004059709142893553, -0.0758700966835022, 0.02712933160364628, 0.12619516253471375, -0.1602097600698471, -0.03158001974225044, 0.06101192161440849, 0.01869376003742218, 0.0386582687497139, -0.12481249868869781, -0.04623749479651451, 0.024604547768831253, -0.008099522441625595, -0.07659726589918137, 0.04500315710902214, 0.007632594555616379, 0.1060425192117691, 0.03917565196752548, -0.0986599326133728, 0.05750032514333725, -0.024385347962379456, -0.11108428984880447, 0.18878285586833954, -0.06612244993448257, -0.144540473818779, -0.10285633057355881, -0.11404616385698318, 0.03514464572072029, 0.02816179394721985, 0.016674986109137535, -0.062482208013534546, -0.040532294660806656, -0.06747932732105255, 0.017274267971515656, -0.10449202358722687, -0.04544343054294586, -0.06891952455043793, 0.034036312252283096, 0.03255314752459526, -0.11663908511400223, -0.026077015325427055, 0.03371225669980049, -0.05657577887177467, 0.08439350873231888, -0.0670146718621254, 0.0921512171626091, 0.12206018716096878, 0.001964608207345009, 0.023332742974162102, -0.0420399010181427, 0.2954738736152649, -0.0905696302652359, -0.0009088972001336515, 0.1647799164056778, 0.001558099640533328, 0.002508108736947179, 0.16675803065299988, 0.007547701243311167, -0.07406014204025269, 0.019041193649172783, -0.0035422765649855137, -0.07190433889627457, -0.20895668864250183, -0.056588590145111084, -0.07173144817352295, 0.027792315930128098, 0.0592992790043354, 0.027163470163941383, 0.0738680362701416, 0.08440229296684265, 0.016717350110411644, 0.09099355340003967, 0.005210543051362038, 0.1022217869758606, 0.051815617829561234, 0.023507244884967804, 0.12378474324941635, -0.04118962585926056, -0.03823202848434448, 0.05899866297841072, -0.01357941422611475, 0.17991676926612854, -0.006288208533078432, 0.07078070193529129, 0.03738372027873993, 0.10220294445753098, 0.050442975014448166, 0.103845976293087, -0.01260087639093399, -0.04334527626633644, -0.060269393026828766, -0.0071700806729495525, -0.06258732825517654, 0.0772247388958931, -0.003613291773945093, -0.05029952898621559, -0.08437100797891617, -0.010352044366300106, 0.014946868643164635, 0.07896753400564194, 0.07148867845535278, -0.31171169877052307, -0.03406986594200134, 0.0534227080643177, -0.0164219718426466, -0.07824867218732834, 0.004486390855163336, 0.00306913279928267, -0.12488926202058792, 0.0028711750637739897, -0.05032695457339287, 0.12686577439308167, -0.08227667957544327, 0.010969712398946285, -0.052240755409002304, 0.05309716612100601, -0.028846971690654755, 0.13436667621135712, -0.3700766861438751, 0.2128361165523529, -0.0009081137832254171, 0.08930022269487381, -0.08801736682653427, -0.037103284150362015, 0.06787507981061935, 0.10024214535951614, 0.12160015106201172, 0.006837377790361643, -0.04293830692768097, -0.1872846633195877, -0.003953889943659306, 0.006317364051938057, 0.02768007665872574, 0.020533205941319466, 0.08525598794221878, -0.03638952970504761, 0.013261630199849606, 0.059983424842357635, 0.05388505384325981, -0.11450041830539703, -0.11571717262268066, -0.004380489233881235, -0.005675703287124634, 0.025709832087159157, -0.054746173322200775, -0.06531241536140442, -0.0413876548409462, 0.11384324729442596, 0.037622593343257904, -0.004087361041456461, -0.12889733910560608, 0.06035679206252098, 0.08242771774530411, -0.0663529485464096, 0.025236858054995537, -0.028581839054822922, 0.0857822448015213, 0.007123873103410006, -0.07712889462709427, 0.12984147667884827, -0.12055832147598267, -0.09423229843378067, -0.0417022742331028, -0.018301306292414665, 0.010599439032375813, 0.08786293119192123, 0.015357782132923603, 0.02865493670105934, -0.0815148577094078, -0.08065575361251831, 0.0013709038030356169, -0.025665294378995895, 0.06712585687637329, 0.060811832547187805, -0.043227121233940125, 0.02269105799496174, -0.05583054944872856, -0.06464558094739914, 0.13451966643333435, 0.2766440510749817, -0.07030798494815826, -0.010208574123680592, 0.06139938160777092, -0.08458246290683746, -0.2990873157978058, 0.014294624328613281, -0.038668982684612274, 0.015184695832431316, 0.07020781189203262, -0.14332415163516998, 0.054709065705537796, -0.04550151899456978, -0.03674684464931488, 0.008159864693880081, -0.12014644593000412, -0.12328729033470154, 0.16727468371391296, 0.09179548174142838, 0.16984987258911133, -0.10481145977973938, 0.004494301043450832, -0.0818740651011467, -0.12633781135082245, 0.08593031018972397, -0.10651220381259918, 0.1209108978509903, 0.0017789674457162619, 0.07619965076446533, -0.0002735412563197315, -0.06165371090173721, 0.11685343086719513, -0.007971000857651234, 0.11770473420619965, -0.09000667184591293, -0.10466943681240082, 0.08184056729078293, -0.021608496084809303, 0.09635280817747116, -0.11468929052352905, 0.024771980941295624, -0.14031660556793213, -0.0570773147046566, -0.06835225969552994, 0.03466123715043068, -0.013782050460577011, -0.07004977762699127, -0.10349927097558975, 0.09683682769536972, -0.0005300024640746415, -0.035189542919397354, 0.1211283877491951, -0.0689147487282753, 0.05203885957598686, 0.15985597670078278, 0.18102626502513885, -0.02846267819404602, 0.057276129722595215, 0.03245551139116287, -0.03638138249516487, 0.05243893340229988, -0.11701638996601105, 0.010240631178021431, 0.07664059102535248, -0.02331460826098919, 0.18464623391628265, 0.08548842370510101, -0.06624113768339157, 0.0219001192599535, 0.09248367697000504, -0.031818851828575134, -0.11497241258621216, -0.006762235425412655, 0.013982152566313744, -0.07047934830188751, -0.06633861362934113, 0.06452449411153793, -0.10404570400714874, 0.005409251898527145, -0.02503570169210434, 0.007070400286465883, -0.0749407485127449, 0.13029395043849945, 0.10759662836790085, 0.05993486940860748, -0.11285997927188873, 0.11893825232982635, 0.03130560368299484, -0.03636588156223297, 0.05192815884947777, 0.046216391026973724, -0.0871567353606224, -0.09776363521814346, 0.05599083751440048, 0.3015536963939667, -0.10767941921949387, -0.03253687545657158, -0.1425795555114746, -0.11415529996156693, 0.05459222570061684, 0.08129260689020157, 0.08256257325410843, 0.036881223320961, -0.066825270652771, 0.007686007767915726, -0.11809059977531433, 0.08443909883499146, 0.0900282934308052, 0.020272163674235344, -0.13552232086658478, 0.06584485620260239, -0.006130721420049667, 0.04609532281756401, -0.0783875584602356, -0.003403802402317524, -0.11993589252233505, 0.04306920990347862, -0.19473305344581604, -0.004675224889069796, -0.08856122940778732, -0.004127290565520525, -0.0003291275061201304, -0.07093410938978195, -0.05072799324989319, -0.019017614424228668, -0.11239905655384064, 0.0018613016000017524, -0.01955207996070385, 0.04704319313168526, -0.06092173978686333, -0.04188378155231476, 0.038073740899562836, -0.03661932423710823, 0.05894539877772331, 0.08195029944181442, -0.01358750369399786, 0.023453421890735626, -0.18682098388671875, -0.027036521583795547, 0.025460360571742058, -0.007129869423806667, 0.06696329265832901, -0.09660536795854568, 0.003924400545656681, 0.010266902856528759, 0.061039019376039505, 0.020123643800616264, 0.03640568256378174, -0.09032764285802841, 0.006701187696307898, -0.02847401797771454, -0.07973361015319824, -0.07003258913755417, 0.12139227241277695, 0.11549285054206848, 0.04165611043572426, 0.1889927089214325, -0.062128446996212006, 0.0274500772356987, -0.16159562766551971, 0.028988832607865334, -0.026157096028327942, -0.08376116305589676, -0.04452790692448616, -0.10310477763414383, 0.06789656728506088, -0.024162108078598976, 0.24066677689552307, 0.07489804178476334, -0.03742160275578499, 0.026952160522341728, 0.037557754665613174, 0.07752262055873871, -0.0022595617920160294, 0.14078466594219208, 0.04784074425697327, -0.020742110908031464, -0.04103482887148857, 0.08092443645000458, 0.08640246838331223, 0.1318352371454239, 0.10953117161989212, 0.15945710241794586, 0.03194357082247734, 0.10132727026939392, -0.021273193880915642, -0.08110769093036652, -0.05156683549284935, -0.015272283926606178, 0.004141254350543022, 0.038539063185453415, -0.05510025471448898, 0.1358027458190918, 0.18504776060581207, -0.10150181502103806, 0.08503440767526627, -0.10245216637849808, -0.07687602937221527, -0.12800978124141693, -0.1375148594379425, -0.07810468226671219, -0.121250219643116, 0.02756406180560589, -0.12240549176931381, 0.012354585342109203, 0.11163955926895142, 0.04483262822031975, -0.0415603332221508, 0.09309131652116776, -0.02914082631468773, -0.031653113663196564, 0.07101940363645554, -0.030364932492375374, -0.017618900164961815, -0.044272180646657944, 0.006570818834006786, 0.00045976106775924563, -0.0626339465379715, 0.03832579776644707, -0.013291413895785809, -0.0861351415514946, 0.05397212877869606, 0.0014714376302435994, -0.13836264610290527, -0.01301450002938509, 0.03905309736728668, 0.045975491404533386, 0.21814750134944916, -0.01788499392569065, 0.039692483842372894, 0.017834005877375603, 0.28828251361846924, -0.0976007729768753, -0.11369277536869049, -0.13107341527938843, 0.20892129838466644, -0.01621754840016365, 0.018500708043575287, -0.03986060619354248, -0.05880863219499588, 0.0049831438809633255, 0.21308420598506927, 0.25826922059059143, -0.1349402517080307, -0.026388585567474365, -0.0019428617088124156, -0.0007079258211888373, -0.04028378427028656, 0.09305600821971893, 0.05337178707122803, 0.10961639881134033, -0.13263796269893646, 0.03695956617593765, -0.05093240737915039, -0.03857826441526413, 0.0017355810850858688, 0.059004563838243484, 0.06874153763055801, -0.0005634132539853454, -0.0837637260556221, 0.04593830928206444, -0.14490148425102234, -0.1029038280248642, -0.003088759956881404, -0.09452810883522034, -0.10905652493238449, -0.05062350630760193, 0.017017381265759468, 0.05392476171255112, 0.08256633579730988, -0.01898333989083767, 0.034642815589904785, -0.03833045810461044, 0.011154642328619957, -0.09175567328929901, -0.04928351938724518, 0.12865082919597626, 0.09684337675571442, 0.1218547597527504, -0.0656997337937355, 0.08645906299352646, 0.14597035944461823, 0.009906010702252388, -0.1164279505610466, 0.09916888922452927, -0.009501388296484947, -0.010827441699802876, 0.055867768824100494, 0.12874507904052734, -0.03004678711295128, 0.05447406321763992, 0.03019150346517563, -0.09019674360752106, 0.06933905929327011, -0.10898610949516296, -0.10006751120090485, -0.08184222877025604, -0.01180946920067072, -0.06590425223112106, 0.1373310685157776, 0.1431296169757843, -0.03851299732923508, 0.0015708698192611337, -0.06281501799821854, 0.04258940368890762, 0.088868647813797, 0.011111646890640259, -0.06929527968168259, -0.12604597210884094, -0.01820666342973709, 0.1003594696521759, -0.004428479820489883, -0.27094268798828125, -0.010192450135946274, -0.11164478957653046, -0.06471869349479675, -0.054776597768068314, 0.08642784506082535, 0.095160111784935, 0.05943844094872475, -0.05283369868993759, -0.09679809957742691, -0.008197816088795662, 0.15145862102508545, -0.16295936703681946, -0.06520017236471176 ]
null
null
transformers
# My Awesome Model
{"tags": ["conversational"]}
text-generation
aydin/DialoGPT-medium-michael
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# My Awesome Model
[ "# My Awesome Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# My Awesome Model" ]
[ 51, 4 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# My Awesome Model" ]
[ -0.05259015038609505, 0.05521034821867943, -0.005910294596105814, 0.017722278833389282, 0.15250112116336823, 0.02286236733198166, 0.07657632976770401, 0.09513414651155472, -0.025391526520252228, -0.047348517924547195, 0.15119488537311554, 0.19781284034252167, -0.020334534347057343, 0.101333387196064, -0.04688440263271332, -0.3143521845340729, 0.06439975649118423, 0.05463787540793419, -0.015605635941028595, 0.12023304402828217, 0.09468326717615128, -0.0530015267431736, 0.08742043375968933, -0.012155864387750626, -0.1293085366487503, -0.0027921805158257484, -0.002384399762377143, -0.10180269181728363, 0.11194873601198196, 0.033712033182382584, 0.05166437849402428, 0.0182647667825222, -0.05843055993318558, -0.139859139919281, 0.03845210000872612, -0.015005595050752163, -0.05602653697133064, 0.05648263916373253, 0.059830192476511, -0.07164353132247925, 0.1669619083404541, 0.13275989890098572, -0.04237370565533638, 0.056127581745386124, -0.17620700597763062, 0.017941240221261978, 0.01800798624753952, 0.019184142351150513, 0.05306641012430191, 0.10830496996641159, -0.03932326287031174, 0.09217294305562973, -0.11410652846097946, 0.08313368260860443, 0.07800983637571335, -0.29151955246925354, -0.025312699377536774, 0.10440942645072937, 0.06437138468027115, 0.048375632613897324, -0.013386772945523262, 0.0621674507856369, 0.02149512618780136, 0.008602659218013287, 0.02225899137556553, -0.06727100163698196, -0.05789240449666977, 0.032748885452747345, -0.0967593789100647, -0.03634428232908249, 0.19753605127334595, -0.024647634476423264, 0.053590498864650726, -0.06265407055616379, -0.11300963163375854, -0.039751436561346054, -0.050429005175828934, -0.029761891812086105, -0.05090925097465515, 0.09489558637142181, 0.004352911841124296, -0.09534718841314316, -0.13405443727970123, -0.01370926946401596, -0.1618979275226593, 0.15892250835895538, 0.012579603120684624, 0.046201955527067184, -0.19210097193717957, 0.11465331166982651, -0.03857925534248352, -0.08259090781211853, 0.030513519421219826, -0.12010065466165543, 0.03160654753446579, -0.008132083341479301, -0.019599268212914467, -0.049325279891490936, 0.061037879437208176, 0.08101806789636612, 0.018783701583743095, 0.005755073390901089, 0.018167443573474884, 0.05343452841043472, 0.05891622602939606, 0.10033947974443436, -0.02891627699136734, -0.0625043511390686, 0.0025436533614993095, -0.12051084637641907, -0.01122665498405695, -0.05357983708381653, -0.18095199763774872, 0.002246231772005558, 0.02455340512096882, 0.05192234739661217, 0.011778532527387142, 0.09955989569425583, -0.028496338054537773, -0.026898741722106934, 0.06898727267980576, 0.002862759632989764, -0.015707949176430702, -0.005368964280933142, -0.010934269987046719, 0.11485416442155838, -0.023099146783351898, 0.04774846136569977, -0.12022071331739426, 0.020393015816807747, -0.07851235568523407, -0.0019349842332303524, -0.06214260309934616, -0.04864754155278206, -0.0019346009939908981, -0.06985589861869812, 0.021118074655532837, -0.14833110570907593, -0.17990200221538544, -0.005064866971224546, 0.021302316337823868, -0.052403319627046585, -0.09162671118974686, -0.0982397273182869, -0.02586611732840538, 0.03574685752391815, -0.05873546749353409, 0.013170980848371983, -0.06884536147117615, 0.06542801111936569, 0.0029820678755640984, 0.05682007595896721, -0.14085575938224792, 0.08719147741794586, -0.12582023441791534, -0.023288866505026817, -0.061977192759513855, 0.1109607070684433, 0.024780582636594772, 0.1267160177230835, 0.004311583004891872, -0.0033308975398540497, -0.08729329705238342, 0.08271238207817078, -0.04243258014321327, 0.22770646214485168, -0.10479787737131119, -0.08809807151556015, 0.2632525563240051, -0.05423165112733841, -0.16432519257068634, 0.10179096460342407, -0.014350244775414467, 0.12198644131422043, 0.13850919902324677, 0.16080057621002197, 0.007628654129803181, 0.03313867375254631, 0.10115300863981247, 0.08631709218025208, -0.08573295921087265, -0.0611947737634182, 0.023627014830708504, -0.011463395319879055, -0.10670105367898941, 0.046802595257759094, 0.04794782027602196, 0.08188598603010178, -0.04982871189713478, -0.028600862249732018, -0.01972118206322193, -0.044152840971946716, 0.05264130234718323, 0.007675500120967627, 0.13217447698116302, -0.03674980252981186, -0.03692879155278206, -0.023745311424136162, 0.01699630729854107, -0.03115241602063179, 0.007061392068862915, -0.05687357112765312, 0.11091547459363937, -0.03406180441379547, 0.051789235323667526, -0.16953988373279572, -0.04873261600732803, -0.02087729424238205, 0.1402055323123932, 0.04973345249891281, 0.1329866498708725, 0.06287940591573715, -0.010758201591670513, 0.00859389640390873, 0.007998145185410976, 0.13181665539741516, 0.007865442894399166, -0.07660657912492752, -0.047718439251184464, 0.09176599979400635, -0.05973208695650101, 0.06147782504558563, -0.098741315305233, -0.004747362341731787, -0.01433002483099699, 0.08674649894237518, 0.006352655589580536, 0.029382232576608658, -0.006192679051309824, 0.003654100699350238, -0.06161240115761757, 0.017873648554086685, 0.12492607533931732, -0.01421504095196724, -0.07439801841974258, 0.22084392607212067, -0.15798072516918182, 0.18006981909275055, 0.18165533244609833, -0.3081994652748108, 0.024602634832262993, -0.08860466629266739, -0.036338552832603455, 0.03426366671919823, 0.0491504967212677, -0.034147560596466064, 0.16587987542152405, -0.016766328364610672, 0.201018825173378, -0.03547777235507965, -0.01287798210978508, -0.010399105958640575, -0.03656993433833122, -0.010632630437612534, 0.09065473079681396, 0.15122920274734497, -0.1677125245332718, 0.18270380795001984, 0.1660280078649521, 0.06873020529747009, 0.17776396870613098, 0.034313347190618515, -0.006856906693428755, 0.07112615555524826, -0.022670727223157883, -0.07675548642873764, -0.049287427216768265, -0.26302891969680786, -0.027947327122092247, 0.06471601128578186, 0.04510856419801712, 0.11924877762794495, -0.10971947014331818, -0.037208184599876404, 0.010892451740801334, -0.013165894895792007, 0.02132410928606987, 0.09682225435972214, 0.01171150617301464, 0.11804302036762238, -0.021027036011219025, -0.05209195241332054, 0.0898953229188919, 0.02727191150188446, -0.0787680521607399, 0.19168277084827423, -0.10074768215417862, -0.3233809769153595, -0.11354339867830276, -0.18166927993297577, -0.017843691632151604, 0.05878754332661629, 0.08049646019935608, -0.09228580445051193, -0.02625267766416073, -0.01639235019683838, 0.0758359357714653, -0.09145816415548325, -0.015880629420280457, -0.09367848187685013, 0.034986745566129684, -0.10827737301588058, -0.07011983543634415, -0.05141967162489891, -0.03368452936410904, -0.04457031562924385, 0.13157756626605988, -0.12242637574672699, 0.06396433711051941, 0.2076517641544342, 0.06227295100688934, 0.05622440204024315, -0.0229496993124485, 0.23288212716579437, -0.10842552781105042, 0.02383521944284439, 0.1717897206544876, -0.03566030040383339, 0.0727933868765831, 0.13435456156730652, 0.006721907295286655, -0.08144525438547134, 0.03465581312775612, -0.04592517390847206, -0.08630958944559097, -0.20441576838493347, -0.14156180620193481, -0.12814727425575256, 0.07913564145565033, 0.03285396471619606, 0.05478321388363838, 0.15024253726005554, 0.11386489123106003, 0.007987297140061855, 0.00976672861725092, -0.006888182368129492, 0.05438044294714928, 0.17482298612594604, -0.05838097631931305, 0.10041683167219162, -0.037591226398944855, -0.1924494504928589, 0.08022978901863098, 0.04309763014316559, 0.08280511945486069, 0.07474655658006668, 0.0856199786067009, 0.013537914492189884, 0.03723837807774544, 0.10897084325551987, 0.1165735274553299, 0.031679023057222366, -0.038079675287008286, -0.04882059991359711, -0.026300756260752678, -0.03285675123333931, 0.05745977535843849, 0.07790146768093109, -0.1608346849679947, -0.06348084658384323, -0.06350091099739075, 0.07662643492221832, 0.09017108380794525, 0.11811108142137527, -0.21219493448734283, 0.01579318381845951, 0.092556893825531, -0.0494147390127182, -0.1304239183664322, 0.07402537018060684, -0.00466050673276186, -0.1397053301334381, 0.037663187831640244, -0.014095795340836048, 0.1359514445066452, -0.0778401643037796, 0.10336452722549438, -0.08307972550392151, -0.06147889420390129, 0.03632286190986633, 0.1355396956205368, -0.30774354934692383, 0.2137020230293274, -0.022472934797406197, -0.05296783149242401, -0.10508129745721817, -0.011727629229426384, 0.020913105458021164, 0.09079049527645111, 0.10090240091085434, -0.0025442070327699184, 0.0061299679800868034, -0.0345483273267746, -0.053218815475702286, 0.024456629529595375, 0.07957815378904343, -0.08542889356613159, 0.0017540202243253589, -0.02361489273607731, -0.004407065454870462, -0.032844748347997665, -0.01189463958144188, -0.011617658659815788, -0.16786961257457733, 0.06556065380573273, -0.002625665394589305, 0.11129079759120941, 0.03491498529911041, 0.0024013579823076725, -0.1009332686662674, 0.19977013766765594, 0.01796281896531582, -0.08052749931812286, -0.08830537647008896, -0.03254766762256622, 0.03660419583320618, -0.06121435388922691, 0.027481911703944206, -0.06916457414627075, 0.033381566405296326, -0.06441576033830643, -0.18325145542621613, 0.1268530637025833, -0.10945470631122589, -0.03609596937894821, -0.04321056231856346, 0.18323224782943726, -0.00929707009345293, -0.0011623724130913615, 0.05866571143269539, 0.0032208464108407497, -0.1347510665655136, -0.10740556567907333, 0.020214511081576347, -0.015275230631232262, 0.009142245166003704, 0.05559912323951721, -0.009665844030678272, 0.00045268211397342384, -0.039558928459882736, -0.023234419524669647, 0.32348164916038513, 0.10732097923755646, -0.04944206401705742, 0.17007054388523102, 0.13087597489356995, -0.0827672928571701, -0.30699312686920166, -0.10971353948116302, -0.10529600828886032, -0.026918673887848854, -0.037983208894729614, -0.19617970287799835, 0.09504909813404083, -0.03528566658496857, -0.022136637941002846, 0.11253651231527328, -0.2759084105491638, -0.0770430713891983, 0.1826775223016739, 0.003314757253974676, 0.3998824954032898, -0.10265109688043594, -0.08777514100074768, -0.06741699576377869, -0.1120782196521759, 0.2033512443304062, -0.05560711398720741, 0.08663415163755417, -0.00517998356372118, 0.15513743460178375, 0.055607251822948456, -0.02176513522863388, 0.08932057023048401, -0.005811662413179874, -0.0546204075217247, -0.1219351515173912, -0.03444604203104973, -0.009159418754279613, 0.007239421829581261, 0.03589896112680435, -0.04242607578635216, 0.01279151439666748, -0.1399589478969574, -0.045490626245737076, -0.0764620453119278, 0.024699507281184196, 0.021008269861340523, -0.0652410089969635, -0.01643640361726284, -0.03945036977529526, -0.012804778292775154, 0.03164318576455116, 0.15236099064350128, -0.06478006392717361, 0.1476556956768036, 0.04904455319046974, 0.15412139892578125, -0.14745712280273438, -0.02258288487792015, -0.06896031647920609, -0.05498642474412918, 0.04900865629315376, -0.10053684562444687, 0.050061121582984924, 0.1202658861875534, -0.0742902010679245, 0.0987328365445137, 0.0922594666481018, -0.01938629150390625, 0.0012483424507081509, 0.1226617842912674, -0.2489612102508545, -0.07742628455162048, -0.10509459674358368, 0.013337249867618084, 0.10138551890850067, 0.06995654851198196, 0.17304721474647522, -0.0037713919300585985, -0.036284226924180984, -0.0064643872901797295, 0.025414984673261642, -0.03540204465389252, 0.05724727362394333, -0.002706433180719614, 0.016663886606693268, -0.15213344991207123, 0.060368724167346954, -0.00024176653823815286, -0.1438901126384735, -0.013603870756924152, 0.16073721647262573, -0.11208858340978622, -0.15145981311798096, -0.007263668347150087, 0.13685113191604614, -0.13171035051345825, -0.03302847594022751, -0.03708777576684952, -0.170182466506958, 0.07439173012971878, 0.1024777740240097, 0.08549231290817261, 0.08025266975164413, -0.06620611250400543, -0.00807863101363182, -0.011656313203275204, -0.026087598875164986, 0.031810320913791656, -0.023377234116196632, -0.09044221043586731, 0.03872343525290489, -0.026654237881302834, 0.13591371476650238, -0.09607382118701935, -0.09331836551427841, -0.135749951004982, 0.039314381778240204, -0.12405620515346527, -0.08138058334589005, -0.12200927734375, -0.0591500885784626, 0.00224387738853693, -0.0001289021165575832, -0.035674065351486206, -0.06687422841787338, -0.13582271337509155, 0.04366770386695862, -0.04484611004590988, 0.0013091047294437885, -0.040241483598947525, 0.04561002552509308, 0.06766383349895477, -0.03493715822696686, 0.13722217082977295, 0.11722734570503235, -0.07864081114530563, 0.08946478366851807, -0.16657429933547974, -0.0683990865945816, 0.08854512125253677, 0.008173754438757896, 0.06165994703769684, 0.06743349134922028, 0.033807408064603806, 0.06109451875090599, 0.04151686280965805, 0.03488299250602722, 0.01739438995718956, -0.09271225333213806, 0.015541021712124348, 0.022296719253063202, -0.1294609159231186, -0.04801803454756737, -0.029226921498775482, 0.00939185917377472, 0.008117396384477615, 0.11003357172012329, -0.0426274873316288, 0.09439733624458313, -0.05888751894235611, 0.036728594452142715, 0.016222506761550903, -0.16461637616157532, -0.020102784037590027, -0.11915475130081177, 0.028684545308351517, -0.0033096212428063154, 0.25625869631767273, 0.06346847862005234, 0.020517030730843544, 0.01250078622251749, 0.08567021042108536, 0.07241600006818771, 0.02562166005373001, 0.1956365555524826, 0.10854171961545944, -0.05020022392272949, -0.12334850430488586, 0.09686340391635895, 0.034720368683338165, 0.06432123482227325, 0.13385434448719025, -0.026959087699651718, 0.002498799469321966, 0.11019360274076462, 0.011678861454129219, 0.04961980879306793, -0.09859088063240051, -0.16400282084941864, -0.00994415208697319, 0.061864156275987625, -0.04559077322483063, 0.12240655720233917, 0.11382720619440079, -0.020697353407740593, 0.03180128335952759, -0.010503606870770454, -0.05694027617573738, -0.16998925805091858, -0.1630837321281433, -0.08357038348913193, -0.11794789135456085, -0.0027763545513153076, -0.11386270076036453, 0.013879159465432167, 0.06452289968729019, 0.0604364387691021, -0.09019444137811661, 0.08891061693429947, 0.0687386617064476, -0.11843101680278778, 0.08828350901603699, -0.033263903111219406, 0.07249268144369125, 0.0015160300536081195, 0.003872724948450923, -0.13800905644893646, 0.032393742352724075, -0.008493867702782154, 0.04159298539161682, -0.09244006127119064, 0.022458361461758614, -0.11297028511762619, -0.07659684121608734, -0.07971972227096558, 0.05093973129987717, -0.03541257977485657, 0.1390930563211441, 0.001295371213927865, -0.035233911126852036, 0.024190181866288185, 0.22729112207889557, -0.06350252777338028, -0.030667411163449287, -0.0618741400539875, 0.21414142847061157, 0.024466563016176224, 0.10703565180301666, -0.016775688156485558, 0.019240234047174454, -0.0764411985874176, 0.3689337372779846, 0.344390869140625, -0.1225387305021286, -0.0015968306688591838, 0.031062176451086998, 0.036916591227054596, 0.11621878296136856, 0.12602226436138153, 0.057955991476774216, 0.2995031177997589, -0.08396036922931671, -0.002026971662417054, -0.02688612788915634, -0.03624163940548897, -0.04409930482506752, 0.10547586530447006, 0.06835740804672241, -0.03330419585108757, -0.027012333273887634, 0.1376710683107376, -0.2966996431350708, 0.12323499470949173, -0.15714547038078308, -0.1487535685300827, -0.06873904913663864, -0.005042468197643757, 0.08589684963226318, 0.04748665541410446, 0.1069009080529213, -0.019124338403344154, -0.08203735202550888, 0.05766449123620987, 0.0320524163544178, -0.22844897210597992, 0.011852608993649483, 0.08361081779003143, -0.06153005734086037, 0.011767351068556309, -0.017906347289681435, 0.038472190499305725, 0.07790610194206238, 0.025976579636335373, -0.032770540565252304, 0.06325861811637878, -0.005814229138195515, -0.05033424496650696, 0.04302205145359039, 0.05059972032904625, 0.017107632011175156, -0.1511564701795578, 0.07320158183574677, -0.1762860119342804, 0.0566408596932888, -0.005331212189048529, -0.04948166385293007, 0.000018263708625454456, 0.01998119056224823, -0.06808236241340637, 0.05880929157137871, 0.0952666699886322, -0.012173139490187168, -0.002317852806299925, -0.056667573750019073, 0.007662574760615826, -0.0679154172539711, -0.0747012197971344, -0.10497893393039703, -0.1338900774717331, -0.11392296850681305, 0.10846775025129318, -0.011928223073482513, -0.19833622872829437, 0.02906924858689308, -0.11258108913898468, 0.04933213070034981, -0.13360801339149475, 0.08599711954593658, 0.1282832771539688, 0.021543797105550766, -0.01265349704772234, 0.04020093381404877, 0.01591683179140091, 0.08550478518009186, -0.09200563281774521, -0.10515180230140686 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilgpt2-imdb This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the [imdb](https://www.kaggle.com/lakshmi25npathi/imdb-dataset-of-50k-movie-reviews) dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "distilgpt2-imdb", "results": []}]}
text-generation
aypan17/distilgpt2-imdb
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# distilgpt2-imdb This model is a fine-tuned version of distilgpt2 on the imdb dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# distilgpt2-imdb\n\nThis model is a fine-tuned version of distilgpt2 on the imdb dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# distilgpt2-imdb\n\nThis model is a fine-tuned version of distilgpt2 on the imdb dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 62, 30, 6, 12, 8, 3, 90, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# distilgpt2-imdb\n\nThis model is a fine-tuned version of distilgpt2 on the imdb dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.09858903288841248, 0.05785796418786049, -0.0016759686404839158, 0.07638204842805862, 0.1661156713962555, 0.028764519840478897, 0.11697712540626526, 0.13459038734436035, -0.1260804682970047, 0.024657925590872765, 0.06175536662340164, 0.08885443955659866, 0.04200506582856178, 0.11280221492052078, -0.027482978999614716, -0.2789631485939026, 0.006080750375986099, 0.01060863770544529, -0.12730713188648224, 0.11110199987888336, 0.13765420019626617, -0.08961024135351181, 0.07629252970218658, 0.026148220524191856, -0.22317451238632202, 0.007762861903756857, -0.003616663161665201, -0.0714050829410553, 0.11467859894037247, 0.02556046098470688, 0.09645067900419235, 0.004318326711654663, 0.1255718171596527, -0.17132088541984558, 0.002472326159477234, 0.09761068224906921, 0.051932357251644135, 0.0940508171916008, 0.03165893256664276, 0.007616893853992224, 0.15129579603672028, -0.08529099076986313, 0.086956687271595, 0.04759401082992554, -0.09122305363416672, -0.20600354671478271, -0.10146544873714447, 0.06411925703287125, 0.08518527448177338, 0.10710041224956512, 0.011214286088943481, 0.11026581376791, -0.08097023516893387, 0.06939317286014557, 0.1902957260608673, -0.25003495812416077, -0.06754618883132935, 0.08665456622838974, 0.04098639264702797, 0.07308656722307205, -0.08446083217859268, 0.00010358238068874925, 0.05389844626188278, 0.05788040533661842, 0.10423678904771805, -0.011981451883912086, -0.09274543076753616, 0.0010909109842032194, -0.13973109424114227, -0.024047326296567917, 0.21208778023719788, 0.02065255492925644, -0.06348469853401184, -0.07349787652492523, -0.050447531044483185, -0.07849359512329102, -0.025700759142637253, -0.06065182015299797, 0.04000496119260788, -0.01750757545232773, -0.0570085346698761, -0.08656328171491623, -0.09232372790575027, -0.09303735196590424, -0.022290701046586037, 0.11839748919010162, 0.04370318725705147, 0.02131793648004532, -0.0599498525261879, 0.120857372879982, -0.025231389328837395, -0.11145853996276855, -0.039507027715444565, -0.023723069578409195, -0.02386138029396534, -0.06259296834468842, -0.06307874619960785, -0.030758850276470184, 0.003436499973759055, 0.17112325131893158, -0.11601484566926956, 0.0626913532614708, 0.004630196373909712, 0.016375809907913208, -0.05629345774650574, 0.18213777244091034, -0.03804527595639229, -0.006153792608529329, 0.022496500983834267, 0.055318307131528854, -0.0019084322266280651, -0.0035060965456068516, -0.1097400039434433, -0.026017818599939346, 0.07904733717441559, 0.046982720494270325, -0.05643411725759506, 0.05197548866271973, -0.03470836952328682, -0.03910870850086212, 0.043376583606004715, -0.1084960401058197, 0.02339017018675804, -0.0268019400537014, -0.09730008989572525, 0.0044136191718280315, 0.036919474601745605, 0.012618201784789562, -0.032106682658195496, 0.05852264538407326, -0.07900620996952057, 0.013482492417097092, -0.11129628866910934, -0.07560646533966064, 0.012504979968070984, -0.04214425012469292, -0.0038477114867419004, -0.10103508830070496, -0.2284734845161438, -0.024292126297950745, 0.04871261119842529, -0.05418778210878372, -0.03346506133675575, -0.020307699218392372, -0.07822181284427643, 0.004640864674001932, -0.014334098435938358, 0.12815584242343903, -0.045312318950891495, 0.07720351219177246, 0.03625058755278587, 0.036069728434085846, -0.03612469136714935, 0.04694606736302376, -0.08228203654289246, 0.013400385156273842, -0.17734982073307037, 0.08670655637979507, -0.0802396610379219, 0.03295941650867462, -0.11218179762363434, -0.12106585502624512, -0.006202100310474634, -0.011583098210394382, 0.06608022004365921, 0.12601198256015778, -0.16301003098487854, -0.04999200254678726, 0.1395997852087021, -0.0852084755897522, -0.10456337034702301, 0.08840490132570267, -0.028668439015746117, 0.06930460035800934, 0.07738775759935379, 0.14055418968200684, 0.10008776187896729, -0.11609754711389542, -0.0030129062943160534, 0.038561802357435226, 0.04892962425947189, -0.0012217171024531126, 0.05872704088687897, 0.002732086693868041, 0.03572160750627518, 0.03450882434844971, -0.0656299814581871, 0.027972904965281487, -0.08360275626182556, -0.0724719911813736, -0.0708797350525856, -0.08625425398349762, 0.035139694809913635, 0.04293852671980858, 0.06434209644794464, -0.07400116324424744, -0.10354652255773544, 0.15522237122058868, 0.13937342166900635, -0.037179041653871536, 0.022677313536405563, -0.09739532321691513, 0.06236105412244797, -0.012482229620218277, -0.003565215738490224, -0.21477752923965454, -0.09696010500192642, 0.010336385108530521, -0.0005082995048724115, 0.04018985107541084, 0.012062632478773594, 0.07781243324279785, 0.06537812948226929, -0.06704828143119812, -0.014188379980623722, -0.09192032366991043, -0.01179648470133543, -0.09016291797161102, -0.1848476529121399, -0.043163374066352844, -0.02573903650045395, 0.17802396416664124, -0.20704661309719086, 0.04797475039958954, 0.0026701854076236486, 0.12950457632541656, 0.02056916616857052, -0.040997009724378586, -0.03173195570707321, 0.038470782339572906, -0.03706660866737366, -0.1024349182844162, 0.04339444264769554, 0.021644435822963715, -0.07037932425737381, -0.04531540721654892, -0.08809106051921844, 0.0794629231095314, 0.110692597925663, -0.024189593270421028, -0.08073407411575317, -0.005211568903177977, -0.07376683503389359, -0.01737161912024021, -0.05879068002104759, -0.0031856652349233627, 0.16455553472042084, -0.009479211643338203, 0.13863079249858856, -0.06767954677343369, -0.07415666431188583, 0.0023065467830747366, -0.018440496176481247, 0.01877313293516636, 0.048094116151332855, 0.07315982133150101, -0.050499994307756424, 0.10725953429937363, 0.1072065606713295, -0.06600731611251831, 0.13471761345863342, -0.05630604550242424, -0.0729367658495903, 0.005340665113180876, 0.005430768709629774, -0.004114406183362007, 0.09294676035642624, -0.1186944991350174, 0.0014988090842962265, 0.025583676993846893, 0.05264275148510933, 0.05298891291022301, -0.20458413660526276, -0.005778568331152201, 0.017913244664669037, -0.04964283108711243, -0.0323483981192112, 0.00041288358625024557, 0.004732784815132618, 0.09398148953914642, 0.027448248118162155, 0.0010803418699651957, 0.0309017114341259, 0.01599869318306446, -0.07823710888624191, 0.19792193174362183, -0.1355850100517273, -0.1600751280784607, -0.12424884736537933, 0.01872633397579193, -0.05739332363009453, 0.004636415280401707, 0.020951969549059868, -0.13360387086868286, -0.03921135887503624, -0.053843360394239426, 0.033084720373153687, -0.04004566743969917, 0.007301324047148228, 0.05551104247570038, 0.018359124660491943, 0.0884355902671814, -0.1519910991191864, -0.0015909802168607712, -0.009300175122916698, -0.11050185561180115, 0.005816521588712931, 0.027176277711987495, 0.08486627042293549, 0.15572024881839752, -0.011628041043877602, 0.0249748807400465, -0.02898854948580265, 0.2389543503522873, -0.09024159610271454, -0.0027395335491746664, 0.18097952008247375, 0.03673423081636429, 0.03638061508536339, 0.06832782179117203, 0.03649292513728142, -0.11776566505432129, 0.031092943623661995, 0.058934587985277176, -0.03194545581936836, -0.26192396879196167, -0.03966150060296059, -0.04369085654616356, -0.05374288931488991, 0.05512528121471405, 0.05171217769384384, 0.07249206304550171, 0.06373238563537598, -0.016199326142668724, 0.09634954482316971, 0.014184301719069481, 0.09016081690788269, 0.11721807718276978, 0.027226539328694344, 0.11222367733716965, -0.039784323424100876, -0.01159890927374363, 0.08077120780944824, -0.015176305547356606, 0.25910642743110657, -0.03547263145446777, 0.06839050352573395, 0.05493205413222313, 0.11481158435344696, -0.02736646495759487, 0.048973582684993744, -0.004397945944219828, -0.0194136630743742, -0.0020633183885365725, -0.049809910356998444, -0.03082447499036789, 0.026421552523970604, -0.08136401325464249, 0.03709899261593819, -0.09444393962621689, 0.04655194282531738, 0.0513240247964859, 0.23587727546691895, 0.0233656894415617, -0.31426170468330383, -0.1277366578578949, 0.018089573830366135, -0.015620251186192036, -0.07717759162187576, 0.016722308471798897, 0.09962084889411926, -0.123314268887043, 0.059556640684604645, -0.0626780241727829, 0.09923739731311798, -0.007166202645748854, 0.013789070770144463, 0.09088417142629623, 0.1612122505903244, 0.007433204911649227, 0.08787576854228973, -0.29299819469451904, 0.18034523725509644, 0.0182258952409029, 0.12627746164798737, -0.06489893049001694, 0.053155072033405304, 0.0329606793820858, 0.14214597642421722, 0.0568285696208477, 0.0033610856626182795, -0.025041382759809494, -0.15828070044517517, -0.013183826580643654, 0.02935333177447319, 0.12042862921953201, -0.013918448239564896, 0.06949502974748611, -0.06349541991949081, 0.024922415614128113, 0.07208245992660522, -0.11258739978075027, -0.17005163431167603, -0.10753372311592102, 0.015235505998134613, -0.003990781959146261, 0.026664171367883682, -0.09386421740055084, -0.08176392316818237, -0.02539128251373768, 0.17213867604732513, 0.008285654708743095, -0.05910921469330788, -0.14617669582366943, 0.10868789255619049, 0.1077231839299202, -0.05679267272353172, 0.028174644336104393, 0.013691004365682602, 0.13507823646068573, 0.05036560073494911, -0.09362855553627014, 0.05383368954062462, -0.09417077898979187, -0.18422894179821014, -0.026041295379400253, 0.10220614075660706, 0.04426535218954086, 0.05241383984684944, 0.0035758246667683125, 0.013933994807302952, 0.0018576119327917695, -0.11413206160068512, 0.012760121375322342, 0.06924131512641907, 0.08072899281978607, 0.045379236340522766, -0.07602836936712265, 0.05224355682730675, -0.02985234558582306, -0.028189288452267647, 0.1195119172334671, 0.18082799017429352, -0.07517927885055542, 0.05281863734126091, 0.03711242228746414, -0.08851515501737595, -0.1834740936756134, 0.10612508654594421, 0.09951096773147583, 0.0008189094951376319, 0.017205091193318367, -0.22954581677913666, 0.11925765872001648, 0.1150461807847023, -0.024794260039925575, 0.10208515077829361, -0.3302255868911743, -0.1363033950328827, 0.06027410551905632, 0.10437522828578949, 0.03135448694229126, -0.13854558765888214, -0.0319560170173645, -0.05004071816802025, -0.1383286565542221, 0.092313252389431, -0.1301286518573761, 0.09407304972410202, -0.009275688789784908, 0.09506095200777054, 0.01085137203335762, -0.030286569148302078, 0.1433597058057785, 0.02039538323879242, 0.0868891254067421, -0.05979950726032257, 0.054426390677690506, 0.11062534153461456, -0.04769987240433693, 0.045789193361997604, -0.0313132107257843, 0.07610573619604111, -0.11088120937347412, -0.018653767183423042, -0.0851203203201294, 0.06789319962263107, -0.041021473705768585, -0.08238191157579422, -0.0549163781106472, 0.04727434366941452, 0.020870229229331017, -0.038850001990795135, 0.09934530407190323, 0.01993190310895443, 0.13195617496967316, 0.05320129543542862, 0.09852191805839539, -0.07296030223369598, -0.05477581173181534, 0.0071131023578345776, -0.004975863266736269, 0.06249861791729927, -0.18942177295684814, 0.026788143441081047, 0.10393096506595612, 0.03437317535281181, 0.14178289473056793, 0.07236164063215256, -0.053518202155828476, 0.022629408165812492, 0.06284020841121674, -0.10888165235519409, -0.12996822595596313, -0.02550327591598034, -0.02941744029521942, -0.09848428517580032, 0.07647223770618439, 0.10609755665063858, -0.07449926435947418, -0.007879724726080894, -0.007203335408121347, 0.014985314570367336, -0.0633728876709938, 0.19197456538677216, 0.04877396672964096, 0.0541086308658123, -0.09807475656270981, 0.1180155798792839, 0.04256163537502289, -0.0485188364982605, 0.034029051661491394, 0.07185391336679459, -0.10403644293546677, -0.028846781700849533, 0.06211164966225624, 0.10909353196620941, -0.06682632118463516, -0.056794922798871994, -0.11299679428339005, -0.11123297363519669, 0.05141117423772812, 0.0849190354347229, 0.06910556554794312, -0.0009122257470153272, -0.053702328354120255, 0.06693114340305328, -0.15555831789970398, 0.07332441210746765, 0.024568190798163414, 0.07966706901788712, -0.17270222306251526, 0.14898288249969482, 0.020350053906440735, 0.03476342186331749, -0.029703965410590172, 0.01309957168996334, -0.08362329751253128, -0.0027939938008785248, -0.19530053436756134, -0.054194774478673935, -0.030832435935735703, 0.006903049070388079, -0.017994163557887077, -0.05835110694169998, -0.0452561229467392, 0.059112489223480225, -0.08663538098335266, -0.053003814071416855, 0.004974282346665859, 0.04586217552423477, -0.1303148865699768, 0.008287191390991211, 0.01568974368274212, -0.09357545524835587, 0.08074646443128586, 0.0698692798614502, 0.02485791966319084, 0.08166243880987167, -0.09058645367622375, -0.01620231568813324, 0.06445866823196411, 0.026800861582159996, 0.06325189024209976, -0.06696026027202606, 0.009463977068662643, -0.016360245645046234, 0.0579470656812191, 0.019817324355244637, 0.046296365559101105, -0.11024749279022217, -0.01226230151951313, -0.06791158765554428, -0.04538530856370926, -0.07672753185033798, 0.06331415474414825, 0.11130911856889725, 0.056982576847076416, 0.17167210578918457, -0.07823360711336136, 0.0057304962538182735, -0.2038513720035553, -0.023153146728873253, -0.008694268763065338, -0.049521107226610184, -0.08413095772266388, -0.04796339198946953, 0.07452008873224258, -0.055201709270477295, 0.12085383385419846, -0.0030957001727074385, 0.09230342507362366, 0.050649844110012054, -0.021654069423675537, -0.013755358755588531, -0.0004660722042899579, 0.18357770144939423, 0.07158172875642776, -0.0043118675239384174, 0.07013621926307678, 0.030970530584454536, 0.09288378059864044, 0.03648405522108078, 0.1986832171678543, 0.12311689555644989, -0.07407746464014053, 0.08048587292432785, 0.051050782203674316, -0.0874195396900177, -0.16983434557914734, 0.08406822383403778, -0.036271315068006516, 0.09604719281196594, -0.06226057559251785, 0.15255115926265717, 0.11885920912027359, -0.16152323782444, 0.039599817246198654, -0.0684669092297554, -0.09494925290346146, -0.1253032386302948, -0.013727550394833088, -0.0656827986240387, -0.1895054578781128, 0.01808467134833336, -0.14057806134223938, 0.043967682868242264, 0.08240476995706558, 0.015981029719114304, -0.0023030193988233805, 0.15088310837745667, -0.020456254482269287, -0.008986072614789009, 0.04565316066145897, -0.0174648966640234, -0.007870620116591454, -0.06491600722074509, -0.075875423848629, 0.02714884653687477, 0.006856916937977076, 0.08045381307601929, -0.04766488075256348, -0.045456789433956146, 0.020193975418806076, -0.02468813583254814, -0.059168729931116104, 0.018822049722075462, 0.04546481743454933, 0.023920129984617233, 0.030607696622610092, 0.028980085626244545, -0.016444137319922447, -0.018139872699975967, 0.3001803159713745, -0.09491780400276184, -0.11818544566631317, -0.15646401047706604, 0.24555085599422455, 0.043097127228975296, -0.020200137048959732, 0.05852754786610603, -0.10668305307626724, -0.04311938211321831, 0.18622547388076782, 0.1574314534664154, -0.05602242052555084, -0.040475085377693176, -0.008146592415869236, -0.022415142506361008, -0.07582755386829376, 0.15682949125766754, 0.13629743456840515, 0.05247962474822998, -0.05196953937411308, -0.021978827193379402, -0.012755722738802433, -0.02700147032737732, -0.05284252017736435, 0.09110863506793976, 0.002796601038426161, -0.0009037382551468909, -0.018245765939354897, 0.07622123509645462, -0.024980833753943443, -0.1558547168970108, 0.031636033207178116, -0.1317138969898224, -0.16180871427059174, -0.020642124116420746, 0.053233202546834946, -0.028246844187378883, 0.07139837741851807, -0.006538618355989456, -0.01058143563568592, 0.1298072338104248, -0.020373040810227394, -0.058482181280851364, -0.10433226823806763, 0.11985468864440918, -0.04832172393798828, 0.1933044195175171, -0.02994736284017563, 0.06542479991912842, 0.11628127098083496, 0.020627647638320923, -0.1305508315563202, 0.034078117460012436, 0.03656959906220436, -0.06870060414075851, 0.012461625039577484, 0.13705027103424072, -0.04719744250178337, 0.0525529682636261, 0.04448704048991203, -0.14626444876194, -0.01257328037172556, -0.05013145133852959, -0.05620745196938515, -0.08936797082424164, -0.023172147572040558, -0.08484786748886108, 0.13140398263931274, 0.21443787217140198, -0.027679188176989555, 0.022594105452299118, -0.1063927486538887, 0.04815575107932091, 0.06454648822546005, 0.07583298534154892, -0.05116401985287666, -0.22753530740737915, -0.006247157230973244, 0.06553887575864792, -0.013408983126282692, -0.20494484901428223, -0.07139565050601959, 0.03967006877064705, -0.05227487161755562, -0.060868166387081146, 0.11208280175924301, 0.08677364885807037, 0.04204433038830757, -0.03751572221517563, -0.09046133607625961, -0.06653536111116409, 0.16237486898899078, -0.16142331063747406, -0.046705640852451324 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt2-med-imdb This model is a fine-tuned version of [gpt2-medium](https://huggingface.co/gpt2-medium) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "model-index": [{"name": "gpt2-med-imdb", "results": []}]}
text-generation
aypan17/gpt2-med-imdb
[ "transformers", "pytorch", "gpt2", "text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# gpt2-med-imdb This model is a fine-tuned version of gpt2-medium on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# gpt2-med-imdb\n\nThis model is a fine-tuned version of gpt2-medium on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# gpt2-med-imdb\n\nThis model is a fine-tuned version of gpt2-medium on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0", "### Training results", "### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 54, 31, 6, 12, 8, 3, 90, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# gpt2-med-imdb\n\nThis model is a fine-tuned version of gpt2-medium on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.17.0.dev0\n- Pytorch 1.10.2\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.11191119998693466, 0.10264981538057327, -0.002660356694832444, 0.07640868425369263, 0.172215074300766, 0.03170574828982353, 0.10284847766160965, 0.13874346017837524, -0.13081935048103333, 0.05207156389951706, 0.10994459688663483, 0.08905108273029327, 0.05286567285656929, 0.1266162097454071, -0.020493952557444572, -0.27640533447265625, 0.01229099277406931, 0.028435014188289642, -0.0575854666531086, 0.10903692990541458, 0.11520075798034668, -0.09608583897352219, 0.07221675664186478, 0.02928522787988186, -0.19467467069625854, -0.008235380984842777, -0.006022591609507799, -0.1072470024228096, 0.09958421438932419, -0.004271273966878653, 0.07286336272954941, 0.0008397144847549498, 0.0944480150938034, -0.1319681704044342, -0.0036807176657021046, 0.07771075516939163, 0.04945167526602745, 0.10121696442365646, 0.06647831946611404, 0.007820459082722664, 0.12545470893383026, -0.10948041081428528, 0.08012038469314575, 0.0360519103705883, -0.09494050592184067, -0.22395047545433044, -0.09195854514837265, 0.0705661028623581, 0.056754302233457565, 0.10828225314617157, 0.004381902981549501, 0.1619616150856018, -0.07058919966220856, 0.06445018202066422, 0.21288473904132843, -0.27921465039253235, -0.04801319167017937, 0.05133965611457825, 0.05597876012325287, 0.04643716290593147, -0.08020930737257004, 0.0008856815402396023, 0.03959408774971962, 0.062487632036209106, 0.11826437711715698, 0.0024236789904534817, -0.02499713934957981, -0.006882001645863056, -0.14487454295158386, -0.05427651107311249, 0.19864967465400696, -0.005885920487344265, -0.06394990533590317, -0.10490656644105911, -0.05595278739929199, -0.1117561087012291, -0.010944627225399017, -0.03237651288509369, 0.01346652116626501, -0.04136353358626366, -0.06753017753362656, -0.07024724781513214, -0.07023974508047104, -0.08144919574260712, -0.019206324592232704, 0.15417449176311493, 0.06812836974859238, 0.033260125666856766, -0.05372045189142227, 0.13502353429794312, -0.030275847762823105, -0.11701757460832596, -0.025667324662208557, -0.025752658024430275, -0.003195946104824543, -0.03868618234992027, -0.03806805983185768, -0.03912922739982605, -0.008295231498777866, 0.15757940709590912, -0.12511441111564636, 0.05428900569677353, 0.023313192650675774, 0.015849387273192406, -0.021464891731739044, 0.17016902565956116, -0.042081575840711594, 0.016982296481728554, 0.03554750606417656, 0.0662144348025322, 0.015494401566684246, -0.017445895820856094, -0.10365194827318192, -0.040517646819353104, 0.10802260041236877, 0.06586812436580658, -0.032556891441345215, 0.05039481818675995, -0.028101937845349312, -0.04666324332356453, 0.06773249059915543, -0.10925298184156418, 0.021070415154099464, -0.028785040602087975, -0.10000073164701462, -0.0007975311600603163, 0.02617049217224121, -0.006540224887430668, -0.08152256906032562, 0.06583645939826965, -0.08198485523462296, -0.014170038513839245, -0.0869794636964798, -0.06416770070791245, 0.006371587980538607, -0.06980878859758377, -0.009543930180370808, -0.09970983862876892, -0.22067536413669586, -0.002877670805901289, 0.00794321671128273, -0.05938306823372841, -0.05477575212717056, -0.0006438805139623582, -0.07978446781635284, 0.02565019391477108, -0.02548939175903797, 0.11628539115190506, -0.03929826617240906, 0.0885976105928421, 0.06643026322126389, 0.034968044608831406, 0.013019321486353874, 0.051452234387397766, -0.08096221834421158, 0.019382499158382416, -0.14793796837329865, 0.08422063291072845, -0.08001524955034256, 0.000033255382732022554, -0.09936938434839249, -0.1303543746471405, 0.03849523141980171, -0.026994602754712105, 0.08948855847120285, 0.1386815905570984, -0.08039848506450653, -0.06277962774038315, 0.16087080538272858, -0.06801403313875198, -0.08305463939905167, 0.09631731361150742, -0.016765331849455833, -0.0020647277124226093, 0.065761499106884, 0.13611489534378052, 0.0708712786436081, -0.11074913293123245, -0.024235311895608902, 0.026232335716485977, 0.031192326918244362, -0.03088979609310627, 0.08240146934986115, -0.003173647215589881, 0.0026206346228718758, 0.04030853882431984, -0.04968005046248436, 0.03499015048146248, -0.10683699697256088, -0.06521079689264297, -0.07126719504594803, -0.07758348435163498, 0.06656067073345184, 0.031107572838664055, 0.07188744843006134, -0.07741918414831161, -0.11869306862354279, 0.08644042909145355, 0.12686358392238617, -0.04023792967200279, 0.016390535980463028, -0.08414659649133682, 0.08789227157831192, -0.07061030715703964, -0.0051418780349195, -0.2057807296514511, -0.09118737280368805, 0.04196761175990105, -0.025405865162611008, 0.02284034714102745, -0.0440024696290493, 0.05826401710510254, 0.09462016820907593, -0.043132930994033813, -0.04035915434360504, -0.08905679732561111, -0.029036948457360268, -0.09847112745046616, -0.14856471121311188, -0.04771515727043152, -0.02256820909678936, 0.1051667332649231, -0.1663494110107422, 0.03479688987135887, 0.008887850679457188, 0.1410468965768814, 0.007854352705180645, -0.049706727266311646, 0.0032818445470184088, 0.043967168778181076, -0.021202851086854935, -0.10454466193914413, 0.047935087233781815, 0.015239118598401546, -0.07589693367481232, -0.03435925394296646, -0.12049706280231476, 0.0878714844584465, 0.11339959502220154, 0.000053743337048217654, -0.07848988473415375, 0.0005143709713593125, -0.08084157109260559, -0.032050974667072296, -0.06050361320376396, -0.005058014765381813, 0.20235082507133484, 0.0023043551482260227, 0.1664983481168747, -0.07188227027654648, -0.06494829803705215, 0.02409900538623333, 0.0018804338760674, -0.010258171707391739, 0.06311165541410446, 0.06844925880432129, -0.05240486189723015, 0.1125764548778534, 0.08426763862371445, -0.0447564460337162, 0.1506222039461136, -0.05359470471739769, -0.08623521775007248, -0.01275326032191515, -0.007389978505671024, 0.0005965207237750292, 0.08960440009832382, -0.12098601460456848, -0.01438936498016119, 0.050907302647829056, 0.04013398662209511, 0.05945160984992981, -0.1822965443134308, -0.0034701137337833643, 0.012050902470946312, -0.03619205579161644, -0.01117788441479206, 0.020650383085012436, 0.019946275278925896, 0.0972333699464798, 0.04433317855000496, 0.011872539296746254, 0.05335637181997299, 0.0234326533973217, -0.06903797388076782, 0.19306613504886627, -0.1267007440328598, -0.19479399919509888, -0.12810736894607544, 0.03885833919048309, -0.0840357169508934, 0.004510818514972925, 0.02689480036497116, -0.11210855841636658, -0.04900015890598297, -0.04007556289434433, 0.05851522460579872, -0.08250726014375687, 0.014242316596210003, 0.03688427805900574, -0.007063193712383509, 0.07580019533634186, -0.14255543053150177, -0.010959168896079063, -0.007259485777467489, -0.10536389797925949, 0.023288702592253685, 0.05514933913946152, 0.08421962708234787, 0.12427026033401489, -0.01407247968018055, 0.033080555498600006, -0.035090453922748566, 0.24791781604290009, -0.0858449786901474, -0.03506903350353241, 0.17381668090820312, 0.04577458277344704, 0.05129384249448776, 0.027306415140628815, 0.03031163476407528, -0.0918402448296547, 0.011547173373401165, 0.02837841771543026, -0.03522346913814545, -0.25739845633506775, -0.055283185094594955, -0.031982552260160446, -0.024009713903069496, 0.0854242667555809, 0.04884788393974304, 0.053438812494277954, 0.07343216240406036, -0.032904691994190216, 0.10729905217885971, -0.04703325405716896, 0.11119529604911804, 0.13003873825073242, 0.024080874398350716, 0.10551168024539948, -0.03041910007596016, -0.040951479226350784, 0.07820373773574829, 0.003030981170013547, 0.24670882523059845, -0.050455473363399506, 0.10749655961990356, 0.03174756467342377, 0.13636204600334167, -0.007666442077606916, 0.04671667143702507, 0.021016035228967667, -0.000577795144636184, 0.009186563082039356, -0.05318877100944519, -0.04889082536101341, -0.002274406375363469, -0.04543156549334526, 0.04241137579083443, -0.13043293356895447, 0.028672153130173683, 0.028672896325588226, 0.22843065857887268, 0.014664452522993088, -0.3282126486301422, -0.11622441560029984, 0.008110825903713703, -0.013211571611464024, -0.09925799071788788, 0.02156480774283409, 0.06910929083824158, -0.1530565619468689, 0.048688407987356186, -0.07161242514848709, 0.08906976878643036, -0.06588031351566315, 0.009656853042542934, 0.051484614610672, 0.16039231419563293, 0.0057891118340194225, 0.09407611936330795, -0.2473020851612091, 0.18896743655204773, 0.023449653759598732, 0.1134328618645668, -0.09105262905359268, 0.04841331019997597, 0.011843842454254627, 0.09794804453849792, 0.07964363694190979, -0.0055975522845983505, -0.020991938188672066, -0.15894483029842377, -0.07492637634277344, 0.016907749697566032, 0.11902490258216858, -0.02180933766067028, 0.07331483066082001, -0.047338422387838364, 0.025527430698275566, 0.036915697157382965, -0.12046503275632858, -0.1624874621629715, -0.09390916675329208, 0.05272384732961655, 0.01800084114074707, -0.00892698671668768, -0.0853017121553421, -0.1008848026394844, -0.04380911961197853, 0.18609216809272766, -0.009079852141439915, -0.09300452470779419, -0.15013866126537323, 0.09028484672307968, 0.11780062317848206, -0.06354431062936783, 0.030917279422283173, -0.0009233102900907397, 0.13622085750102997, 0.027759691700339317, -0.09878776967525482, 0.059115756303071976, -0.07979830354452133, -0.20013315975666046, -0.023469891399145126, 0.1321849673986435, 0.03156936541199684, 0.04256920889019966, -0.017516806721687317, 0.021345745772123337, -0.01525464840233326, -0.11013095825910568, 0.04409889504313469, 0.06306921690702438, 0.0988459661602974, 0.05890767648816109, -0.03144418075680733, 0.06665212661027908, -0.01675114780664444, -0.0218418650329113, 0.14562886953353882, 0.19903160631656647, -0.07456424832344055, 0.07278188318014145, 0.05891812965273857, -0.07007093727588654, -0.1824810802936554, 0.057446178048849106, 0.10328491777181625, 0.00666420254856348, -0.01661405898630619, -0.22366495430469513, 0.08141542971134186, 0.10188455134630203, -0.03288797289133072, 0.10230743139982224, -0.3256417214870453, -0.121519535779953, 0.0770055279135704, 0.10886307805776596, 0.070397287607193, -0.15052711963653564, -0.04610797390341759, -0.031626664102077484, -0.12921178340911865, 0.10364644974470139, -0.0967973917722702, 0.12293843179941177, -0.04800591990351677, 0.11418897658586502, 0.021620137616991997, -0.05637173354625702, 0.14664123952388763, 0.008819102309644222, 0.06088145449757576, -0.064769446849823, 0.03817703574895859, 0.1358441561460495, -0.049242157489061356, 0.07781194150447845, -0.008238011039793491, 0.08201039582490921, -0.14171990752220154, -0.031097931787371635, -0.09173180162906647, 0.07264290750026703, -0.03558911755681038, -0.0766160860657692, -0.04379522427916527, 0.032618846744298935, 0.011265543289482594, -0.04387889429926872, 0.08956394344568253, 0.05486435815691948, 0.10279273986816406, 0.07419534772634506, 0.08373094350099564, -0.03694687411189079, -0.10978885740041733, -0.010462277568876743, -0.021966388449072838, 0.07978864759206772, -0.1559978574514389, 0.008518794551491737, 0.11242262274026871, 0.02999117039144039, 0.11330956220626831, 0.06906444579362869, -0.07806465029716492, 0.01797710359096527, 0.043068286031484604, -0.15405981242656708, -0.11418302357196808, -0.05519236996769905, -0.023385506123304367, -0.08082432299852371, 0.08147097378969193, 0.1096758246421814, -0.07239266484975815, -0.02466718479990959, -0.016836319118738174, 0.002457221271470189, -0.02360456995666027, 0.1854589432477951, 0.05213760584592819, 0.057929087430238724, -0.10513630509376526, 0.1325315684080124, 0.05980668589472771, -0.061766304075717926, 0.0644475519657135, 0.08593089133501053, -0.09329124540090561, -0.03005082905292511, 0.04334596171975136, 0.11873089522123337, -0.07283110171556473, -0.04766807332634926, -0.1033206507563591, -0.09611372649669647, 0.07915076613426208, 0.041024137288331985, 0.05478503182530403, -0.005230814218521118, -0.056908752769231796, 0.049814704805612564, -0.1747293919324875, 0.0720025897026062, 0.022968461737036705, 0.07730323076248169, -0.17674453556537628, 0.1539926826953888, 0.016897378489375114, 0.07097823172807693, -0.030346127226948738, -0.0024127771612256765, -0.09218800812959671, -0.0009724416304379702, -0.15296564996242523, -0.038954973220825195, -0.026094969362020493, -0.0045993877574801445, -0.026329580694437027, -0.03691373020410538, -0.029548807069659233, 0.06896379590034485, -0.07317454367876053, -0.059780072420835495, 0.00642798375338316, 0.026550721377134323, -0.11026069521903992, 0.008048689924180508, 0.0008219040464609861, -0.09037970006465912, 0.10109749436378479, 0.07963890582323074, 0.029130656272172928, 0.04326366260647774, -0.046711914241313934, -0.013339859433472157, 0.04245494306087494, 0.008929590694606304, 0.05669697746634483, -0.06833982467651367, 0.01577766425907612, -0.025015464052557945, 0.0364147312939167, 0.0233305636793375, 0.05020127817988396, -0.13488155603408813, -0.024606890976428986, -0.02381345070898533, -0.019446687772870064, -0.08880294859409332, 0.05836031585931778, 0.09302583336830139, 0.034678809344768524, 0.1258208006620407, -0.07061324268579483, 0.032971158623695374, -0.1792069673538208, -0.0199568048119545, -0.02098342962563038, -0.07190575450658798, -0.054798468947410583, -0.010136924684047699, 0.08717287331819534, -0.048940934240818024, 0.1520620435476303, 0.006730204448103905, 0.09757667034864426, 0.03121134452521801, 0.008817163296043873, -0.006213412154465914, -0.0006594614824280143, 0.1726102977991104, 0.07409612834453583, -0.017626402899622917, 0.07566215842962265, 0.0340210385620594, 0.09149796515703201, 0.04572117328643799, 0.1644638031721115, 0.08042505383491516, -0.024424239993095398, 0.07869653403759003, 0.05151807889342308, -0.09405029565095901, -0.21398766338825226, 0.013301350176334381, -0.01636723428964615, 0.1171288937330246, -0.04688435420393944, 0.10871154814958572, 0.10310300439596176, -0.13703365623950958, 0.04617294669151306, -0.047300852835178375, -0.09381797909736633, -0.11527000367641449, -0.057587966322898865, -0.05575291067361832, -0.13940423727035522, 0.0298498272895813, -0.1270035356283188, 0.022762944921851158, 0.05983971804380417, 0.016783520579338074, -0.021470163017511368, 0.15716855227947235, 0.013616486452519894, -0.014347899705171585, 0.07332690060138702, -0.0012766250874847174, -0.006764135789126158, -0.05497428774833679, -0.06311926245689392, 0.024464648216962814, 0.013443328440189362, 0.09437407553195953, -0.03651810437440872, -0.05104826018214226, 0.03532322868704796, -0.01757938787341118, -0.0686647966504097, 0.015504193492233753, 0.038014210760593414, 0.03081708960235119, 0.04891503229737282, 0.010088415816426277, -0.012179696932435036, -0.043005574494600296, 0.26882168650627136, -0.07918774336576462, -0.0735970139503479, -0.11713453382253647, 0.2392052412033081, 0.04362645000219345, -0.03736133500933647, 0.0749567300081253, -0.1066281646490097, -0.04499750956892967, 0.1858161836862564, 0.17674866318702698, -0.03039935976266861, -0.03445291519165039, 0.005385329015552998, -0.02652573771774769, -0.03996475413441658, 0.14812971651554108, 0.12132016569375992, 0.07739956676959991, -0.06155546382069588, -0.004560163244605064, -0.019519032910466194, -0.02053377591073513, -0.07486405968666077, 0.09429533034563065, 0.020086903125047684, 0.002947992878034711, -0.030298147350549698, 0.03953562304377556, -0.03989291191101074, -0.11124446243047714, 0.004511664621531963, -0.13331477344036102, -0.16718988120555878, -0.007013489026576281, 0.04269063100218773, -0.01579403318464756, 0.07169925421476364, 0.010107075795531273, -0.00010013616702053696, 0.1077994778752327, -0.01028897799551487, -0.08647549897432327, -0.07700691372156143, 0.10015381872653961, -0.0372232086956501, 0.21662160754203796, -0.012426680885255337, 0.09107642620801926, 0.11287083476781845, 0.015006963163614273, -0.16134832799434662, 0.02898653782904148, 0.04500270634889603, -0.05461054667830467, 0.023089392110705376, 0.17412054538726807, -0.02494603954255581, 0.014807967469096184, 0.010530211962759495, -0.1320381462574005, -0.04125247150659561, -0.02198587730526924, -0.011624375358223915, -0.07222957164049149, -0.017430055886507034, -0.06791235506534576, 0.14012779295444489, 0.19682984054088593, -0.035552650690078735, -0.014918497763574123, -0.09804369509220123, 0.0460672602057457, 0.05376909673213959, 0.06107949838042259, -0.02320932410657406, -0.21819627285003662, -0.002659719903022051, 0.07669095695018768, 0.011333881877362728, -0.24412479996681213, -0.07203985005617142, 0.03430383279919624, -0.05579990893602371, -0.08932530134916306, 0.09609952569007874, 0.040418170392513275, 0.03500905632972717, -0.03924477845430374, -0.04720274358987808, -0.07884521782398224, 0.13668152689933777, -0.17088907957077026, -0.05216681584715843 ]
null
null
transformers
TrainingArgs: lr=2e-5, train-batch-size=16, eval-batch-size=16, num-train-epochs=5, weight-decay=0.01,
{"license": "mit"}
text-classification
aypan17/roberta-base-imdb
[ "transformers", "pytorch", "roberta", "text-classification", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #roberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
TrainingArgs: lr=2e-5, train-batch-size=16, eval-batch-size=16, num-train-epochs=5, weight-decay=0.01,
[]
[ "TAGS\n#transformers #pytorch #roberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 42 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.015270475298166275, 0.07784062623977661, -0.006913014221936464, 0.03031000867486, 0.17507308721542358, 0.03787829726934433, 0.13349303603172302, 0.11901473999023438, 0.03586392477154732, -0.056423649191856384, 0.12209897488355637, 0.2674638628959656, -0.021239958703517914, 0.08964911848306656, -0.10582119226455688, -0.28342393040657043, 0.06279760599136353, 0.05949212238192558, 0.029677292332053185, 0.11079192906618118, 0.12207958102226257, -0.06350982189178467, 0.06103965640068054, -0.0075846510007977486, -0.12125470489263535, 0.0224524587392807, 0.051496244966983795, -0.13292483985424042, 0.10982797294855118, 0.057890959084033966, 0.11578663438558578, 0.06923933327198029, -0.02956966683268547, -0.16099999845027924, 0.02904612571001053, -0.02903473936021328, -0.10480178147554398, 0.04837451130151749, 0.05592615529894829, -0.0880395919084549, 0.08493572473526001, 0.0586806945502758, 0.021680809557437897, 0.06915151327848434, -0.1315837800502777, -0.12729276716709137, -0.06244611740112305, 0.07053236663341522, 0.05955991894006729, 0.0594349130988121, 0.015373914502561092, 0.15122906863689423, -0.12147130072116852, 0.08778580278158188, 0.0778718963265419, -0.33632248640060425, 0.0064614033326506615, 0.11653807759284973, 0.05098944157361984, 0.009749379940330982, -0.044952694326639175, 0.05613875389099121, 0.04166017845273018, 0.002612762153148651, -0.006552654784172773, -0.07282614707946777, -0.06960464268922806, 0.0324726365506649, -0.05974531173706055, -0.058297861367464066, 0.1973942518234253, -0.05117504671216011, 0.025441374629735947, -0.04526829347014427, -0.05149608105421066, -0.01900477707386017, -0.017570530995726585, 0.03600542992353439, -0.016283631324768066, 0.0850866436958313, 0.025114666670560837, -0.008670778013765812, -0.12982097268104553, 0.02045367658138275, -0.2270353138446808, 0.15843048691749573, 0.0231480710208416, 0.0514945350587368, -0.14581288397312164, 0.06100359186530113, 0.0028405007906258106, -0.09457235783338547, -0.000022629032173426822, -0.09547387063503265, 0.07135052978992462, -0.03012135811150074, -0.059840310364961624, 0.004347297362983227, 0.09348815679550171, 0.21816371381282806, 0.0443388931453228, 0.01524397823959589, -0.03507334366440773, 0.11535859107971191, 0.012910556979477406, 0.10663365572690964, 0.049676600843667984, -0.007867014035582542, 0.06446689367294312, -0.14735578000545502, 0.03369944170117378, -0.04079921171069145, -0.18291178345680237, -0.04178571701049805, 0.014934933744370937, 0.10382059216499329, 0.017724808305501938, 0.08124621212482452, -0.05354369059205055, 0.0003068434016313404, 0.10956144332885742, -0.054405368864536285, 0.006861052941530943, -0.0055258809588849545, 0.02841925621032715, 0.05529003590345383, 0.001399414730258286, 0.01291539054363966, -0.04495726898312569, 0.14278638362884521, -0.05306093767285347, -0.010857263579964638, -0.02675461582839489, -0.05550506338477135, 0.063965804874897, -0.13550405204296112, 0.04723324999213219, -0.16556844115257263, -0.1306067556142807, 0.01240731030702591, 0.034010883420705795, -0.00809480156749487, -0.054907042533159256, 0.002982628997415304, 0.006921630352735519, 0.03136621043086052, -0.07198341190814972, -0.06856188923120499, -0.07519263029098511, 0.10866942256689072, -0.04961789771914482, 0.04371850937604904, -0.16168536245822906, 0.06343785673379898, -0.1165781319141388, -0.026021864265203476, -0.0764901265501976, 0.008297312073409557, -0.06159660220146179, 0.1758906990289688, -0.0016613139305263758, -0.0499991849064827, -0.01788114570081234, 0.052205998450517654, -0.07580771297216415, 0.13490018248558044, -0.06457376480102539, -0.09110291302204132, 0.19505517184734344, -0.1147320568561554, -0.16391760110855103, 0.06900615990161896, -0.023903682827949524, 0.05970091372728348, 0.0971871167421341, 0.19706526398658752, 0.09366099536418915, -0.039323512464761734, 0.08136890083551407, 0.10433478653430939, -0.11174235492944717, -0.17095845937728882, 0.026194633916020393, -0.02895878255367279, -0.095587357878685, 0.05216395854949951, 0.02606705017387867, 0.07253661751747131, -0.03143586963415146, -0.055816203355789185, -0.021290462464094162, -0.010919405147433281, 0.06601686030626297, 0.04695510491728783, 0.09242820739746094, -0.09555437415838242, -0.0053232512436807156, 0.009102428331971169, -0.00839668232947588, 0.04668902978301048, 0.029317669570446014, -0.07625672966241837, 0.09293614327907562, 0.0494563952088356, 0.013909727334976196, -0.1704079806804657, -0.03549288213253021, -0.023880979046225548, 0.09412987530231476, 0.03620443493127823, 0.11305099725723267, 0.020733535289764404, -0.05411464348435402, -0.020075665786862373, 0.0021124649792909622, 0.1668119877576828, 0.035079728811979294, -0.027349665760993958, -0.08981102705001831, 0.05978449434041977, -0.0496072880923748, 0.006959720980376005, -0.05649307370185852, 0.02386421151459217, 0.07855677604675293, 0.07280320674180984, -0.025105740875005722, 0.09389583766460419, -0.03761984035372734, 0.04909336194396019, -0.07304345071315765, 0.02069101855158806, 0.12558050453662872, 0.019845236092805862, -0.06709694117307663, 0.19800424575805664, -0.11315086483955383, 0.2786068320274353, 0.21330387890338898, -0.2204921543598175, -0.012253672815859318, -0.053387951105833054, -0.01724988967180252, 0.02073272131383419, 0.028262760490179062, 0.033962346613407135, 0.04619245603680611, -0.005125383846461773, 0.18483199179172516, -0.04157929867506027, -0.03394966945052147, -0.007212119176983833, -0.0637417882680893, -0.031927574425935745, 0.07520434260368347, 0.13584652543067932, -0.24044634401798248, 0.19553662836551666, 0.2324822098016739, 0.03956875950098038, 0.15502867102622986, -0.03908134624361992, 0.05053986608982086, 0.04845346137881279, -0.046062830835580826, -0.009078843519091606, -0.032668422907590866, -0.13057707250118256, -0.019331978633999825, 0.08402013033628464, 0.009111189283430576, 0.050766415894031525, -0.13957303762435913, -0.07296991348266602, 0.0018473673844709992, 0.004090183414518833, -0.05544304847717285, 0.10410027205944061, 0.04097779467701912, 0.09662675857543945, -0.016733869910240173, -0.08736216276884079, 0.13199540972709656, 0.00934749748557806, -0.07697417587041855, 0.16552728414535522, -0.12926150858402252, -0.26877132058143616, -0.17131268978118896, -0.19835111498832703, -0.001197736244648695, 0.04739713668823242, 0.13941949605941772, -0.04854447394609451, -0.04738545045256615, 0.03327532112598419, -0.04846900328993797, -0.051398519426584244, 0.0002684048959054053, -0.09390732645988464, 0.0833142101764679, -0.06339555978775024, -0.09736098349094391, -0.08219321817159653, -0.017654718831181526, -0.031187741085886955, 0.13919170200824738, -0.10779121518135071, 0.07774496078491211, 0.13188351690769196, -0.018924381583929062, 0.04298737272620201, -0.06929879635572433, 0.15436851978302002, -0.07636065781116486, -0.01919984631240368, 0.20733731985092163, -0.027719710022211075, 0.07598876953125, 0.18826213479042053, 0.047519393265247345, -0.05463387817144394, 0.0137303052470088, -0.06568129360675812, -0.09735734760761261, -0.25262048840522766, -0.12996253371238708, -0.11226397007703781, 0.05002486705780029, 0.0601210817694664, 0.07743754237890244, 0.1538868546485901, 0.08952434360980988, -0.0030563578475266695, -0.005884705111384392, 0.019249992445111275, 0.10371819883584976, 0.31695225834846497, 0.0054613021202385426, 0.14492563903331757, -0.10269840806722641, -0.10169844329357147, 0.10138961672782898, 0.02683618851006031, 0.13173425197601318, 0.1270015984773636, 0.038184475153684616, 0.06791332364082336, 0.10098575800657272, 0.15022225677967072, 0.09922037273645401, 0.05206388980150223, -0.006741958670318127, -0.041383516043424606, -0.0060977027751505375, -0.039472728967666626, 0.031164024025201797, 0.026805898174643517, -0.16824765503406525, -0.056291595101356506, -0.1636076122522354, 0.060390230268239975, 0.09997217357158661, 0.03048548474907875, -0.1729135811328888, 0.012998228892683983, 0.07789775729179382, -0.013635927811264992, -0.07155255973339081, 0.07992423325777054, -0.10004531592130661, -0.15026292204856873, 0.11903895437717438, -0.03595196083188057, 0.12902599573135376, -0.03640344738960266, 0.06701364368200302, -0.030396554619073868, -0.13778935372829437, 0.036197829991579056, 0.1336929202079773, -0.297087699174881, 0.24621820449829102, 0.0028801700100302696, -0.036294467747211456, -0.06860916316509247, -0.02695000171661377, 0.040473926812410355, 0.22316795587539673, 0.08514604717493057, -0.005774241406470537, -0.11481992900371552, -0.12862500548362732, 0.012787027284502983, 0.004851765464991331, 0.0824078768491745, -0.01669466122984886, -0.024995585903525352, -0.06252328306436539, -0.008624742738902569, -0.02322501689195633, -0.006707038730382919, 0.019691307097673416, -0.1737595647573471, 0.04783300310373306, 0.047035228461027145, 0.06449396908283234, -0.009612901136279106, -0.03980349376797676, -0.11866278201341629, 0.18364669382572174, -0.11264199763536453, -0.08266022056341171, -0.11174386739730835, -0.11401553452014923, 0.01027686707675457, -0.07806487381458282, 0.07217659056186676, -0.08958840370178223, -0.012904161587357521, -0.08586584031581879, -0.19054993987083435, 0.1145515888929367, -0.0903288796544075, -0.0478232204914093, -0.05519554018974304, 0.136574849486351, -0.10933735966682434, 0.019947033375501633, 0.04939620569348335, 0.031960614025592804, -0.09788037091493607, -0.10498933494091034, -0.012736644595861435, -0.027464143931865692, 0.054405082017183304, -0.005568138789385557, -0.10241598635911942, -0.08123550564050674, -0.02200615406036377, -0.04413958266377449, 0.2541355788707733, 0.21845771372318268, -0.0755271166563034, 0.1833929866552353, 0.15257172286510468, -0.09755069017410278, -0.31997695565223694, -0.12474818527698517, -0.1468094289302826, -0.06784229725599289, 0.001698772539384663, -0.14466093480587006, 0.08183603733778, 0.03201159089803696, -0.05853882059454918, 0.08518130332231522, -0.16338516771793365, -0.10443266481161118, 0.18369856476783752, -0.026523740962147713, 0.363707035779953, -0.12112439423799515, -0.10024087131023407, -0.06649547815322876, -0.16418156027793884, 0.15201453864574432, 0.01191655732691288, 0.09744561463594437, -0.025186466053128242, 0.04800332710146904, 0.011516163125634193, -0.043817322701215744, 0.0963980033993721, -0.016982033848762512, 0.041190993040800095, -0.1284676194190979, -0.09320864826440811, 0.08384976536035538, 0.0014603876043111086, 0.017602967098355293, -0.07630874961614609, 0.014466874301433563, -0.12091780453920364, -0.04699680209159851, -0.06836505979299545, 0.07777136564254761, 0.012498159892857075, -0.0590004026889801, -0.04401300847530365, 0.002755352295935154, -0.02179114706814289, -0.0292428620159626, 0.26064690947532654, -0.04229084402322769, 0.14728142321109772, 0.09909746050834656, 0.11174019426107407, -0.13144609332084656, 0.02099216729402542, -0.09894736111164093, -0.08778215199708939, 0.043843187391757965, -0.0925530418753624, 0.029660064727067947, 0.12278140336275101, -0.048463333398103714, 0.07383280992507935, 0.09448559582233429, 0.03149595111608505, -0.010675218887627125, 0.16516442596912384, -0.17339476943016052, -0.008298659697175026, -0.025579513981938362, 0.011636839248239994, 0.09724070876836777, 0.05041550099849701, 0.11229787021875381, 0.02200443111360073, -0.04651240259408951, 0.01783752255141735, 0.0002603822504170239, -0.02590739168226719, 0.033787280321121216, 0.060501523315906525, 0.01867355778813362, -0.13181856274604797, 0.055331822484731674, 0.05336341634392738, -0.11851124465465546, -0.022729607298970222, 0.11263979226350784, -0.14935390651226044, -0.12500666081905365, -0.03449619188904762, 0.08382796496152878, -0.20107947289943695, -0.0627276822924614, -0.06141974776983261, -0.14385314285755157, 0.05530200153589249, 0.17932336032390594, 0.11182599514722824, 0.09113465994596481, -0.031184641644358635, -0.06587179005146027, 0.00849085208028555, -0.00620881374925375, -0.05049935728311539, 0.03716537356376648, -0.09921111911535263, 0.01544321607798338, 0.00028337084222584963, 0.11487166583538055, -0.07467514276504517, -0.05330612510442734, -0.154265895485878, 0.039193131029605865, -0.0885206088423729, -0.01501183956861496, -0.09268225729465485, -0.01915832795202732, 0.00839703343808651, -0.05318891629576683, -0.044614437967538834, -0.04450055956840515, -0.123932383954525, 0.024953139945864677, -0.008391705341637135, 0.07974749803543091, -0.0936661809682846, -0.04432099685072899, 0.09043833613395691, -0.0058213830925524235, 0.08874953538179398, 0.06038675084710121, -0.07630153000354767, 0.0808194950222969, -0.1722293645143509, -0.08981925249099731, 0.10626255720853806, 0.01822861284017563, 0.04785257950425148, 0.03518310561776161, 0.039443861693143845, 0.10445536673069, -0.010431250557303429, 0.07540077716112137, 0.010070890188217163, -0.14828680455684662, 0.026441678404808044, -0.013704473152756691, -0.1616998314857483, -0.026094067841768265, -0.05116421729326248, 0.1053810715675354, -0.00774990813806653, 0.19868077337741852, -0.05671078339219093, 0.06625524908304214, -0.04184311255812645, 0.006671397015452385, -0.028573712334036827, -0.1893147975206375, -0.11162614822387695, -0.10436969995498657, -0.02859046682715416, 0.007691019680351019, 0.2865038812160492, 0.07780726999044418, -0.0450611412525177, 0.07391577214002609, 0.10370909422636032, 0.003091786289587617, -0.00010806451609823853, 0.23723892867565155, 0.07564480602741241, -0.020800065249204636, -0.09167128801345825, 0.05714334174990654, -0.02069506049156189, -0.07697904855012894, 0.11720095574855804, 0.10316652059555054, -0.0002765992539934814, 0.04601084813475609, 0.04112879931926727, 0.049115363508462906, -0.09670879691839218, -0.15948835015296936, 0.028923645615577698, 0.06226532533764839, 0.00148869794793427, 0.0648285374045372, 0.1389636993408203, -0.04829956218600273, 0.04710688814520836, -0.051290687173604965, -0.023547137156128883, -0.18692182004451752, -0.1290079802274704, -0.08661231398582458, -0.090308777987957, 0.03638651221990585, -0.04452250152826309, 0.01754397712647915, 0.08303368836641312, 0.04893208667635918, -0.07871778309345245, -0.006304668262600899, -0.035229988396167755, -0.05718562752008438, 0.046240683645009995, -0.037448685616254807, 0.020526904612779617, -0.04951948672533035, -0.02386397309601307, -0.12109237909317017, -0.04646764695644379, -0.03888213634490967, 0.04731560871005058, -0.027351465076208115, 0.0010358095169067383, -0.14763909578323364, -0.09346037358045578, -0.024507954716682434, 0.07281776517629623, 0.00029216770781204104, 0.15874627232551575, -0.005167014431208372, 0.026776039972901344, 0.07080712914466858, 0.1828041970729828, -0.03054393082857132, -0.12350297719240189, -0.013865289278328419, 0.24565623700618744, 0.0922081246972084, 0.08599456399679184, 0.00009756956569617614, -0.001107318908907473, -0.0319879986345768, 0.29813119769096375, 0.3259623050689697, -0.018910080194473267, 0.042327504605054855, -0.0009393712971359491, 0.02983325906097889, 0.14872710406780243, 0.1491290032863617, 0.06971724331378937, 0.23898208141326904, -0.06451400369405746, -0.011734003201127052, -0.04143473505973816, 0.004139584023505449, -0.12279152125120163, 0.08885862678289413, 0.02614206075668335, -0.07095813751220703, -0.03854086622595787, 0.12840433418750763, -0.16902241110801697, 0.12485120445489883, 0.0431097187101841, -0.1550978720188141, -0.04252185672521591, -0.011266125366091728, 0.14257916808128357, 0.0043437909334897995, 0.05507756397128105, -0.01884179189801216, -0.0787227600812912, 0.021923622116446495, 0.01255471259355545, -0.23746921122074127, -0.01628023572266102, 0.07739441841840744, -0.012119514867663383, 0.05764603987336159, -0.02557375468313694, 0.05586700886487961, 0.08470562100410461, 0.07087631523609161, -0.039128292351961136, 0.09702734649181366, 0.01442712265998125, -0.04549090936779976, 0.03184929117560387, -0.0695037990808487, 0.007013747934252024, -0.09403301030397415, 0.0546768419444561, -0.11834867298603058, 0.07570833712816238, -0.05788019299507141, -0.07456602156162262, -0.04536823555827141, 0.0738741010427475, -0.07075820863246918, 0.06266383826732635, 0.06193368509411812, -0.0030944172758609056, -0.043973278254270554, -0.06268013268709183, -0.026040056720376015, 0.046088673174381256, -0.17924536764621735, -0.10759919881820679, -0.03798272833228111, -0.059929776936769485, 0.10294733941555023, 0.019251933321356773, -0.18812671303749084, -0.007768756244331598, -0.11977460980415344, 0.0383276529610157, -0.18218262493610382, 0.08152299374341965, 0.07156456261873245, 0.011516316793859005, -0.005514270626008511, -0.08244802057743073, 0.028891025111079216, 0.03567333146929741, -0.11413950473070145, -0.07847294211387634 ]
null
null
transformers
# RudeRick discord bot
{"tags": ["conversational"]}
text-generation
ayush19/rick-sanchez
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# RudeRick discord bot
[ "# RudeRick discord bot" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# RudeRick discord bot" ]
[ 51, 7 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# RudeRick discord bot" ]
[ -0.004051297437399626, -0.019657569006085396, -0.007262144237756729, 0.033118169754743576, 0.15250718593597412, -0.005543639417737722, 0.13142916560173035, 0.12788236141204834, 0.0871502235531807, 0.008951302617788315, 0.12517599761486053, 0.20869563519954681, -0.010614423081278801, 0.112012580037117, -0.08498121798038483, -0.21173754334449768, 0.07946047931909561, 0.022198397666215897, 0.07553859800100327, 0.11284596472978592, 0.09400453418493271, -0.059532780200242996, 0.09996241331100464, -0.010886628180742264, -0.18067748844623566, 0.00874783843755722, 0.06896424293518066, -0.09325657784938812, 0.1055544912815094, 0.03680276498198509, 0.08627905696630478, 0.02646850049495697, -0.09068945050239563, -0.056919973343610764, 0.05447133630514145, 0.014098191633820534, -0.04503091424703598, 0.06629421561956406, 0.017392171546816826, -0.10419265925884247, 0.16183140873908997, 0.07801566272974014, 0.012379385530948639, 0.08681429177522659, -0.16621752083301544, -0.01230164710432291, -0.07067950814962387, 0.05050495266914368, 0.09692991524934769, 0.1008320227265358, -0.04253174364566803, 0.14817269146442413, -0.09460388869047165, 0.12323788553476334, 0.1707848310470581, -0.3203018605709076, -0.03472569212317467, 0.05672808736562729, 0.03481750190258026, 0.07513585686683655, -0.04533309489488602, 0.04236325994133949, 0.008251480758190155, 0.025378189980983734, -0.043461814522743225, -0.0612306073307991, -0.18191812932491302, -0.00031061904155649245, -0.074305959045887, -0.03805476054549217, 0.20995482802391052, -0.01234474964439869, 0.04642922803759575, -0.05745590478181839, -0.11436521261930466, -0.07685752958059311, -0.04169401153922081, -0.004874608479440212, -0.10353381186723709, 0.05645041912794113, -0.01631922647356987, -0.09045449644327164, -0.11580603569746017, -0.013582836836576462, -0.1500324308872223, 0.17180174589157104, 0.026794498786330223, 0.05199043080210686, -0.20434732735157013, 0.07279285788536072, 0.05772344768047333, -0.09595074504613876, 0.0068405382335186005, -0.0941714197397232, -0.0037909881211817265, -0.001356271910481155, -0.043739646673202515, -0.11034975945949554, 0.10295910388231277, 0.20914490520954132, -0.05629834905266762, 0.03946994990110397, -0.061004240065813065, 0.07838626950979233, 0.03533027321100235, 0.06255855411291122, 0.028446398675441742, 0.016162080690264702, 0.07491371780633926, -0.0930405780673027, 0.05906709283590317, -0.07975813746452332, -0.16053152084350586, -0.02715228870511055, 0.04763944074511528, 0.08134829252958298, 0.06442561000585556, 0.13125161826610565, -0.06021683290600777, -0.03702501207590103, 0.10224398970603943, -0.02617962844669819, -0.0028890909161418676, 0.026717739179730415, -0.009354396723210812, 0.057875532656908035, 0.014737559482455254, 0.02196008339524269, -0.10864590108394623, 0.0402413047850132, -0.06497277319431305, 0.0030839478131383657, -0.023378053680062294, -0.045248597860336304, 0.034976039081811905, -0.03553057834506035, -0.019495878368616104, -0.164489284157753, -0.1254182606935501, 0.020387159660458565, -0.044285405427217484, -0.040098126977682114, -0.09246353805065155, -0.06328707188367844, -0.017229095101356506, 0.04846777394413948, -0.0577671155333519, -0.043097544461488724, -0.04105126112699509, 0.09878603368997574, -0.050706855952739716, 0.10892191529273987, -0.13372297585010529, 0.043557360768318176, -0.10168597102165222, -0.029441341757774353, -0.1414409577846527, 0.08117860555648804, -0.03305210545659065, 0.11304254829883575, -0.008176302537322044, 0.006995058618485928, -0.11412016302347183, 0.046554919332265854, -0.021257562562823296, 0.23934844136238098, -0.09218235313892365, -0.10267744958400726, 0.28260913491249084, -0.06961699575185776, -0.1333465725183487, 0.16240596771240234, 0.007704293821007013, -0.008421165868639946, 0.12367378175258636, 0.17423731088638306, -0.10225261747837067, -0.029576463624835014, 0.02210959978401661, 0.05683998018503189, -0.16089865565299988, 0.018011927604675293, -0.009936884976923466, -0.031235435977578163, -0.04568549990653992, 0.024782555177807808, 0.186803936958313, 0.10285104811191559, -0.03633951395750046, -0.026061665266752243, -0.034112367779016495, 0.000006800697519793175, 0.08990590274333954, -0.011254423297941685, 0.0788644403219223, -0.07711150497198105, -0.06404370814561844, -0.1408022940158844, 0.02161429636180401, 0.010168429464101791, 0.03379184380173683, -0.10120680183172226, 0.1189238429069519, 0.0682699903845787, 0.06926432251930237, -0.17916814982891083, -0.13127505779266357, -0.009423132985830307, 0.1623409390449524, 0.0795457661151886, 0.08338839560747147, 0.05802498012781143, -0.0650431364774704, -0.009675479494035244, -0.0038224835880100727, 0.13484635949134827, -0.0300598107278347, -0.08571558445692062, -0.07641372084617615, 0.07071790099143982, -0.05695025250315666, 0.08076367527246475, -0.048688896000385284, 0.032450299710035324, 0.12494257837533951, 0.1306777000427246, 0.013673516921699047, 0.017153017222881317, 0.029956849291920662, -0.014162194915115833, -0.061319150030612946, -0.017106516286730766, 0.0806996151804924, -0.01157974824309349, -0.0719507560133934, 0.16548104584217072, -0.13872605562210083, 0.08430685847997665, 0.1909583956003189, -0.2111743986606598, -0.021693633869290352, -0.032601747661828995, -0.03634278476238251, 0.02063208818435669, 0.05794040486216545, -0.055769167840480804, 0.14330756664276123, 0.007984546944499016, 0.14947864413261414, -0.027433166280388832, -0.018454957753419876, -0.02642888016998768, -0.046355437487363815, -0.000045687920646741986, 0.07393763959407806, 0.09379240870475769, -0.17900298535823822, 0.15022435784339905, 0.15921476483345032, 0.020177964121103287, 0.1753976196050644, 0.02924812026321888, 0.03634043410420418, 0.05574822425842285, 0.00625483226031065, -0.03509845584630966, -0.07734912633895874, -0.24638409912586212, -0.03279801830649376, 0.06586014479398727, 0.012610887177288532, 0.08885902166366577, -0.07409875839948654, -0.0281024482101202, -0.03887729346752167, -0.0020505813881754875, 0.056490253657102585, 0.14693404734134674, 0.07759766280651093, 0.159124955534935, -0.0007034508744254708, -0.04822347313165665, 0.08514297753572464, 0.020438630133867264, -0.08440033346414566, 0.14502541720867157, -0.14589856564998627, -0.32756149768829346, -0.07921526581048965, -0.14432401955127716, -0.0830225870013237, 0.05446287989616394, 0.1226489469408989, -0.15726029872894287, -0.010663729161024094, 0.01725996658205986, 0.11971447616815567, -0.0022217442747205496, 0.0214186180382967, 0.0009283678373321891, -0.002924374071881175, -0.09144241362810135, -0.1096002534031868, -0.07228237390518188, -0.042602065950632095, -0.09266085177659988, 0.1543123871088028, -0.0963369831442833, 0.06642821431159973, 0.1665918081998825, 0.05130283534526825, 0.0511837936937809, -0.032466135919094086, 0.2483791559934616, -0.10782789438962936, 0.022485222667455673, 0.20744098722934723, -0.023906610906124115, 0.06321345269680023, 0.10205050557851791, -0.004347749520093203, -0.0953555628657341, 0.04154190048575401, -0.016244757920503616, -0.07288249582052231, -0.19904226064682007, -0.13459186255931854, -0.1070149689912796, 0.11828211694955826, 0.022482000291347504, 0.07583434879779816, 0.1827859729528427, 0.0476146899163723, -0.07065626233816147, -0.028861045837402344, 0.08567199856042862, 0.09590783715248108, 0.19275157153606415, -0.05858411639928818, 0.14167800545692444, -0.03307264298200607, -0.11083383113145828, 0.09682846069335938, -0.009390050545334816, 0.08365470916032791, 0.05984801799058914, 0.10840502381324768, 0.018835702911019325, 0.08055760711431503, 0.131032794713974, 0.04461801052093506, 0.027237042784690857, -0.04882653057575226, -0.02089436538517475, -0.01669626496732235, -0.10784732550382614, 0.025143884122371674, 0.06755475699901581, -0.14078183472156525, -0.019780384376645088, -0.05166186764836311, 0.11846735328435898, 0.1417011022567749, 0.037801697850227356, -0.16032880544662476, -0.06810987740755081, 0.06702491641044617, -0.027496566995978355, -0.10085796564817429, 0.07410687953233719, 0.012052211910486221, -0.1484258621931076, 0.04606972634792328, -0.04261937737464905, 0.12070387601852417, -0.060841768980026245, 0.09279312938451767, -0.12325524538755417, -0.02631361223757267, 0.017868690192699432, 0.08483792841434479, -0.31215938925743103, 0.13615302741527557, -0.002074917545542121, -0.040834806859493256, -0.1216224953532219, -0.01630510948598385, 0.013645077124238014, 0.07531376928091049, 0.08049327880144119, -0.0000026908612653642194, -0.03439663350582123, -0.09684213250875473, -0.016091318801045418, 0.012828544713556767, 0.0918772742152214, -0.02115478552877903, -0.02545945905148983, -0.02146442048251629, -0.012920202687382698, -0.023347314447164536, -0.06860028207302094, -0.0188311617821455, -0.17632125318050385, 0.07516885548830032, 0.10293499380350113, 0.07469146698713303, -0.00026518767117522657, 0.0013074910966679454, -0.020345864817500114, 0.2433048039674759, -0.04888593405485153, -0.09627526998519897, -0.09902114421129227, 0.002402437385171652, 0.0028070760890841484, -0.08093087375164032, -0.00016026198863983154, -0.08296157419681549, 0.05346263200044632, -0.07887186110019684, -0.1731288582086563, 0.11972032487392426, -0.08881057798862457, -0.08628346771001816, -0.019536344334483147, 0.18137651681900024, 0.015487182885408401, -0.0015342400874942541, 0.03184230998158455, 0.010635457001626492, -0.09267191588878632, -0.09858614951372147, 0.04208497703075409, 0.006818912457674742, 0.018366796895861626, 0.06449694186449051, -0.03735930845141411, -0.08289336413145065, -0.0857548862695694, 0.001195126329548657, 0.32607531547546387, 0.18460814654827118, -0.03465323895215988, 0.1765696108341217, 0.08426442742347717, -0.01354992389678955, -0.2867382764816284, -0.11522028595209122, -0.08454004675149918, -0.0535363145172596, -0.04196605831384659, -0.18682467937469482, 0.04534748196601868, 0.0020402371883392334, -0.005059211980551481, 0.10509736835956573, -0.23870886862277985, -0.09650015830993652, 0.15516629815101624, -0.035653915256261826, 0.3725944459438324, -0.09365910291671753, -0.07454187422990799, -0.02928813174366951, -0.136939138174057, 0.19000999629497528, -0.023695392534136772, 0.09344306588172913, -0.0104478495195508, 0.17096762359142303, 0.052827201783657074, -0.00023938847880344838, 0.06808638572692871, 0.010685344226658344, -0.00824067648500204, -0.1156863123178482, -0.09497223794460297, 0.04062409698963165, 0.015432486310601234, 0.021438583731651306, -0.06209973618388176, 0.0014917263761162758, -0.1348809450864792, -0.027125954627990723, -0.12509305775165558, 0.0797140970826149, 0.021439649164676666, -0.022730628028512, -0.02970554679632187, -0.056862927973270416, -0.039571259170770645, 0.037561316043138504, 0.23432017862796783, -0.04896041750907898, 0.1838422566652298, 0.07931889593601227, 0.11043897271156311, -0.17888812720775604, 0.011677575297653675, -0.03287019953131676, -0.053682610392570496, 0.05940951406955719, -0.04383280128240585, 0.04449264705181122, 0.07190138846635818, -0.054617319256067276, 0.09166956692934036, 0.08177383244037628, -0.021025488153100014, 0.04658159241080284, 0.0985029861330986, -0.2703382968902588, -0.0657627210021019, -0.030854512006044388, 0.05781712755560875, 0.08046699315309525, 0.06810393929481506, 0.19748513400554657, 0.04626622423529625, -0.07477296143770218, 0.01327232364565134, 0.048608750104904175, -0.03189757466316223, 0.02467403933405876, -0.011517536826431751, 0.019343184307217598, -0.1532917022705078, 0.05181441083550453, 0.03696589916944504, -0.14932048320770264, 0.05523395538330078, 0.19755232334136963, -0.13422295451164246, -0.132910817861557, -0.12144945561885834, 0.10449910908937454, -0.0686165913939476, 0.014200721867382526, -0.046760644763708115, -0.113393135368824, 0.05292917788028717, 0.0961272120475769, 0.042450565844774246, 0.09076735377311707, -0.03529809042811394, -0.027966739609837532, -0.0003990547847934067, -0.014749879948794842, 0.008919527754187584, -0.003066648030653596, -0.038380686193704605, 0.04550251364707947, -0.026024846360087395, 0.1252385675907135, -0.08263692259788513, -0.08811428397893906, -0.17275482416152954, 0.0009987696539610624, -0.127787783741951, -0.10067319869995117, -0.10357688367366791, -0.04942990094423294, -0.014417366124689579, -0.05923264101147652, -0.032387297600507736, -0.048703476786613464, -0.11992163956165314, 0.018372133374214172, -0.027582472190260887, 0.03478296846151352, -0.11024816334247589, 0.015767768025398254, 0.0774446576833725, -0.032603032886981964, 0.18452145159244537, 0.18730169534683228, -0.10677547007799149, 0.08906043320894241, -0.1441984921693802, -0.09339692443609238, 0.09842756390571594, 0.016786636784672737, 0.054315805435180664, 0.10400868952274323, -0.0012006615288555622, 0.048869866877794266, 0.05910594388842583, 0.05985870957374573, 0.08587117493152618, -0.09463752806186676, 0.05953231081366539, 0.016318028792738914, -0.13280726969242096, -0.0659424290060997, -0.019993027672171593, 0.07805422693490982, 0.03519459813833237, 0.14901071786880493, -0.05837226286530495, 0.08249519765377045, -0.06733103841543198, 0.04185079410672188, 0.02234342321753502, -0.1532207876443863, -0.050744056701660156, -0.07991328090429306, 0.05758707597851753, -0.008699063211679459, 0.117976114153862, 0.003559028496965766, -0.057989511638879776, 0.06439366936683655, 0.10017991811037064, -0.03516251593828201, -0.0009690507431514561, 0.02369009517133236, 0.07005898654460907, -0.06954904645681381, -0.07307285815477371, 0.01928657665848732, 0.04031972587108612, 0.010001272894442081, 0.2069065272808075, 0.009243779815733433, 0.017052050679922104, 0.05216255784034729, -0.006322366651147604, 0.003050084924325347, -0.0950966477394104, -0.11364231258630753, -0.11468597501516342, 0.039149560034275055, -0.0641418844461441, 0.13795630633831024, 0.11378670483827591, 0.0056852721609175205, -0.0008054337231442332, -0.030612416565418243, -0.05069145932793617, -0.1320904940366745, -0.15395872294902802, -0.07022935152053833, -0.1153189092874527, 0.01412342581897974, -0.09632264077663422, 0.06566950678825378, -0.018575675785541534, 0.057638589292764664, -0.08427782356739044, 0.12167072296142578, 0.01932121440768242, -0.07547447830438614, 0.05717523768544197, -0.03973325714468956, 0.03411007300019264, 0.016208674758672714, -0.0202607661485672, -0.07488101720809937, 0.031054802238941193, 0.01341310702264309, 0.07830573618412018, -0.04456506296992302, 0.010645879432559013, -0.15214715898036957, -0.10933972895145416, -0.034121640026569366, 0.07629593461751938, -0.04414378106594086, 0.14747729897499084, 0.0398230217397213, 0.012438365258276463, 0.029912566766142845, 0.22372016310691833, -0.03610505163669586, -0.03688298910856247, -0.08155881613492966, 0.17809636890888214, -0.00940372422337532, 0.05787484720349312, -0.04237418994307518, -0.0014296327717602253, -0.09367341548204422, 0.2893730103969574, 0.2559322714805603, -0.09098641574382782, 0.038901571184396744, -0.04884760454297066, 0.04904242604970932, 0.06788794696331024, 0.11332841962575912, 0.1083584576845169, 0.23593758046627045, -0.05223436281085014, 0.025820259004831314, -0.009733397513628006, -0.031953468918800354, -0.12422209233045578, 0.037356238812208176, 0.009926198050379753, -0.03670632466673851, -0.03551424667239189, 0.07973384112119675, -0.22493185102939606, 0.04887348785996437, -0.13154569268226624, -0.2085157334804535, -0.07319334149360657, 0.012606464326381683, 0.09916914999485016, 0.025932032614946365, 0.13388699293136597, -0.0034587057307362556, -0.09593294560909271, 0.07117877900600433, -0.006777314469218254, -0.18059498071670532, -0.027874749153852463, 0.10202175378799438, -0.13531562685966492, -0.012642327696084976, -0.03182796388864517, 0.07092048227787018, 0.0996398776769638, 0.024843111634254456, -0.050300344824790955, 0.026903949677944183, 0.001999322324991226, -0.029622377827763557, -0.011174309067428112, 0.0931021049618721, 0.0182312224060297, -0.053840409964323044, 0.08185680955648422, -0.16167336702346802, 0.029324328526854515, -0.03239167481660843, -0.007167826872318983, -0.017861319705843925, 0.07204554229974747, -0.0507335364818573, 0.06293915957212448, 0.09225155413150787, -0.029422936961054802, -0.013937256298959255, -0.026870980858802795, -0.04703833535313606, -0.02055569924414158, -0.07596271485090256, -0.14321690797805786, -0.19424547255039215, -0.07321759313344955, 0.08639638870954514, 0.006085546687245369, -0.1324247270822525, 0.009858396835625172, -0.15618564188480377, 0.0708700492978096, -0.14747849106788635, 0.11030254513025284, 0.04889186844229698, -0.004106552805751562, 0.01699703373014927, 0.03454083949327469, 0.05270007252693176, 0.09557764232158661, -0.10117481648921967, -0.06443745642900467 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mbert-finetuned-azerbaijani-ner This model is a fine-tuned version of [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased) on the wikiann dataset. It achieves the following results on the evaluation set: - Loss: 0.1385 - Precision: 0.8899 - Recall: 0.9154 - F1: 0.9025 - Accuracy: 0.9669 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.2928 | 1.0 | 625 | 0.1415 | 0.8584 | 0.8918 | 0.8748 | 0.9595 | | 0.1254 | 2.0 | 1250 | 0.1335 | 0.8875 | 0.9119 | 0.8996 | 0.9637 | | 0.077 | 3.0 | 1875 | 0.1385 | 0.8899 | 0.9154 | 0.9025 | 0.9669 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.6
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["wikiann"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "mbert-finetuned-azerbaijani-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "wikiann", "type": "wikiann", "args": "az"}, "metrics": [{"type": "precision", "value": 0.8898541731306236, "name": "Precision"}, {"type": "recall", "value": 0.915416533673795, "name": "Recall"}, {"type": "f1", "value": 0.9024543738200126, "name": "F1"}, {"type": "accuracy", "value": 0.966948310139165, "name": "Accuracy"}]}]}]}
token-classification
azizbarank/mbert-finetuned-azerbaijani-ner
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:wikiann", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-wikiann #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
mbert-finetuned-azerbaijani-ner =============================== This model is a fine-tuned version of bert-base-multilingual-cased on the wikiann dataset. It achieves the following results on the evaluation set: * Loss: 0.1385 * Precision: 0.8899 * Recall: 0.9154 * F1: 0.9025 * Accuracy: 0.9669 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.6
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.6" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-wikiann #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.6" ]
[ 66, 98, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-wikiann #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.6" ]
[ -0.10958761721849442, 0.0958002582192421, -0.0018517913995310664, 0.12605558335781097, 0.15620048344135284, 0.03152143210172653, 0.11130799353122711, 0.12501031160354614, -0.08756159245967865, 0.018895640969276428, 0.1338922679424286, 0.1655699610710144, 0.014896656386554241, 0.10021331906318665, -0.043103691190481186, -0.2634996473789215, 0.0001310136285610497, 0.046963080763816833, -0.06226341798901558, 0.12766137719154358, 0.08973588049411774, -0.13641871511936188, 0.09600426256656647, 0.02618812769651413, -0.1963932365179062, 0.004984634462743998, 0.009496445767581463, -0.061431560665369034, 0.1476271003484726, 0.027026575058698654, 0.12403721362352371, -0.007606336381286383, 0.0930604413151741, -0.18194565176963806, 0.010810405015945435, 0.05402581766247749, 0.010015562176704407, 0.09298313409090042, 0.044387880712747574, 0.015252281911671162, 0.13882753252983093, -0.07209942489862442, 0.0545584112405777, 0.014716418460011482, -0.11361300200223923, -0.22405287623405457, -0.08794049173593521, 0.047057729214429855, 0.08380433917045593, 0.10304297506809235, 0.0029367958195507526, 0.14055949449539185, -0.09825866669416428, 0.08549246191978455, 0.22535115480422974, -0.3033199906349182, -0.06805708259344101, 0.020215313881635666, -0.0025899012107402086, 0.024472033604979515, -0.10961952805519104, -0.03250085189938545, 0.05288899689912796, 0.04774133488535881, 0.1287582814693451, -0.03367232903838158, -0.10717187076807022, 0.019043706357479095, -0.13675542175769806, -0.03573748841881752, 0.1578582376241684, 0.05399251729249954, -0.036190759390592575, -0.04253276810050011, -0.052659500390291214, -0.1429373174905777, -0.028315840288996696, -0.011595931835472584, 0.04262052848935127, -0.03230062127113342, -0.05864059552550316, 0.0072117033414542675, -0.11076385527849197, -0.06186363473534584, -0.08099312335252762, 0.1351952999830246, 0.04219238460063934, 0.025094758719205856, -0.03798333927989006, 0.11004552245140076, 0.011187722906470299, -0.1260959953069687, 0.008542277850210667, 0.025139153003692627, 0.0012843306176364422, -0.04996492713689804, -0.0435732863843441, -0.04574880376458168, 0.013741160742938519, 0.12162915617227554, -0.049613043665885925, 0.03710218891501427, 0.04879226163029671, 0.04047086462378502, -0.08456820994615555, 0.182933047413826, -0.05806358531117439, -0.012073526158928871, 0.013511859811842442, 0.03210557997226715, 0.013995070941746235, -0.0012455829419195652, -0.11742576956748962, 0.0030900088604539633, 0.08826413005590439, 0.006904860958456993, -0.07143015414476395, 0.07495231926441193, -0.046625733375549316, -0.02848111279308796, 0.022915547713637352, -0.08241324871778488, 0.03398052230477333, -0.0013174588093534112, -0.08288544416427612, -0.010962538421154022, 0.022232796996831894, 0.023434078320860863, -0.012224379926919937, 0.12262524664402008, -0.09164536744356155, 0.022126130759716034, -0.09252399951219559, -0.10126431286334991, 0.02941211685538292, -0.09692751616239548, 0.04817909002304077, -0.09576054662466049, -0.17381323873996735, -0.004008473828434944, 0.06743038445711136, -0.018518617376685143, -0.06755658239126205, -0.04391643777489662, -0.06856579333543777, 0.015170915052294731, -0.014514697715640068, 0.13435113430023193, -0.06654480844736099, 0.08382724970579147, 0.014849434606730938, 0.06360279768705368, -0.04289638623595238, 0.04818841442465782, -0.10387354344129562, 0.021202508360147476, -0.15908470749855042, 0.015207553282380104, -0.04524841159582138, 0.07398597151041031, -0.09528649598360062, -0.09675804525613785, 0.01470832247287035, -0.006474542431533337, 0.06723429262638092, 0.08734080940485, -0.15733855962753296, -0.06852599233388901, 0.1496555656194687, -0.06747204065322876, -0.13347145915031433, 0.11900060623884201, -0.06522528827190399, 0.04384332150220871, 0.06123409792780876, 0.1455872803926468, 0.06956622004508972, -0.09079058468341827, 0.0011691423133015633, 0.01110027264803648, 0.04629766568541527, -0.07538487762212753, 0.07345661520957947, 0.01323297992348671, 0.01380879431962967, 0.029026968404650688, -0.027203546836972237, 0.06202084943652153, -0.09002231806516647, -0.09859024733304977, -0.04123881831765175, -0.09665112942457199, 0.026332946494221687, 0.0732664242386818, 0.07699009776115417, -0.09502697736024857, -0.08227867633104324, 0.05705247446894646, 0.09320659935474396, -0.04981129989027977, 0.02481592632830143, -0.07129161804914474, 0.06337753683328629, -0.06020396575331688, -0.028124190866947174, -0.16589684784412384, -0.02517622523009777, 0.003136832034215331, -0.0010483681689947844, 0.0020877073984593153, 0.02707556262612343, 0.06854245811700821, 0.05940594896674156, -0.049021460115909576, -0.014147416688501835, -0.02441168576478958, 0.0021259544882923365, -0.13283050060272217, -0.2084018588066101, -0.04219810292124748, -0.02201271802186966, 0.11262901872396469, -0.2138851135969162, 0.024876557290554047, -0.013407430611550808, 0.07956621050834656, 0.01562650129199028, -0.004144037142395973, -0.04324570670723915, 0.06727109849452972, -0.04946143925189972, -0.049702856689691544, 0.06558958441019058, 0.006277765147387981, -0.08722064644098282, -0.038815829902887344, -0.06885340809822083, 0.17416737973690033, 0.1312985122203827, -0.11025267839431763, -0.06878241896629333, -0.00037919668829999864, -0.062110938131809235, -0.039871737360954285, -0.05404550954699516, 0.04561083763837814, 0.1838746815919876, -0.010784171521663666, 0.13782334327697754, -0.06995625793933868, -0.04647640511393547, 0.03175222501158714, -0.031828682869672775, 0.018923595547676086, 0.12191256135702133, 0.14848022162914276, -0.10129430890083313, 0.14585942029953003, 0.14973455667495728, -0.09212110936641693, 0.12465020269155502, -0.042758677154779434, -0.07750259339809418, -0.01809605583548546, -0.03138771280646324, -0.005152126774191856, 0.11969289928674698, -0.15955758094787598, -0.0031132649164646864, 0.029880788177251816, 0.02566184289753437, 0.013888920657336712, -0.21964164078235626, -0.04695729911327362, 0.03808568790555, -0.03144281730055809, -0.016734512522816658, -0.013908415101468563, 0.0018741725943982601, 0.10171209275722504, 0.009569639340043068, -0.09443642944097519, 0.03962300345301628, 0.009413491934537888, -0.08040270954370499, 0.2121707797050476, -0.07010069489479065, -0.13556969165802002, -0.11242256313562393, -0.08360930532217026, -0.04803154990077019, -0.0013495365856215358, 0.062349770218133926, -0.0903528556227684, -0.03625527024269104, -0.06551778316497803, 0.0077688926830887794, -0.00648609409108758, 0.039297908544540405, 0.012814756482839584, -0.0014709378592669964, 0.061620503664016724, -0.10146649181842804, -0.012562897987663746, -0.05396951735019684, -0.05247657001018524, 0.03594595938920975, 0.03677096217870712, 0.11521665006875992, 0.1404682993888855, -0.012858479283750057, 0.011368470266461372, -0.025171345099806786, 0.23669570684432983, -0.06605953723192215, -0.01791241392493248, 0.13099826872348785, -0.015420905314385891, 0.041023045778274536, 0.1112326830625534, 0.07339291274547577, -0.0827561542391777, -0.0012331773759797215, 0.0401405394077301, -0.0364905409514904, -0.22498536109924316, -0.036549825221300125, -0.054074663668870926, -0.015243188478052616, 0.1038239449262619, 0.030055731534957886, 0.03163512796163559, 0.08255165815353394, 0.0355834886431694, 0.08788827806711197, -0.05812469497323036, 0.057019591331481934, 0.0952666699886322, 0.047981519252061844, 0.11844716966152191, -0.0509147122502327, -0.05933060497045517, 0.04140467196702957, 0.007031518500298262, 0.23209597170352936, 0.010256070643663406, 0.13436342775821686, 0.06375083327293396, 0.1867857575416565, -0.005973643623292446, 0.08437740057706833, -0.015017503872513771, -0.04964961111545563, -0.01248865481466055, -0.03542684018611908, -0.025113623589277267, 0.03223311901092529, -0.07208890467882156, 0.07338391989469528, -0.11644057929515839, 0.022372078150510788, 0.049876414239406586, 0.28343167901039124, 0.045097965747117996, -0.32125183939933777, -0.10031809657812119, 0.0019995728507637978, -0.04263104870915413, -0.022767964750528336, 0.03487231954932213, 0.09278889745473862, -0.09279337525367737, 0.02545984275639057, -0.06630145013332367, 0.09566432982683182, -0.05545753985643387, 0.03910095989704132, 0.08222442120313644, 0.08231395483016968, 0.005635988432914019, 0.0916265919804573, -0.2846216857433319, 0.28239068388938904, 0.0032196047250181437, 0.07690320909023285, -0.09026820212602615, 0.0006727706640958786, 0.0382712259888649, 0.0736212208867073, 0.08629223704338074, -0.009242960251867771, -0.05537813529372215, -0.1837763488292694, -0.05992956832051277, 0.03353868052363396, 0.049954138696193695, -0.029813773930072784, 0.09365174919366837, -0.023774102330207825, 0.006764404941350222, 0.0731472373008728, 0.021464398130774498, -0.04504748061299324, -0.0963577851653099, -0.022525813430547714, 0.032175347208976746, -0.06507863849401474, -0.0617746040225029, -0.11161281168460846, -0.12649156153202057, 0.15528643131256104, -0.019588354974985123, -0.02972579188644886, -0.11079620569944382, 0.0803317129611969, 0.06708823889493942, -0.09257043153047562, 0.036354124546051025, 0.004469707608222961, 0.0702224150300026, 0.03353588655591011, -0.0723877027630806, 0.10971641540527344, -0.07723395526409149, -0.1574288308620453, -0.06801959127187729, 0.08730727434158325, 0.04157638177275658, 0.06098831444978714, -0.00218007480725646, 0.01064167357981205, -0.04205532371997833, -0.08808611333370209, 0.026918962597846985, -0.00040487851947546005, 0.08142272382974625, -0.01324852742254734, -0.04603198543190956, 0.04307066276669502, -0.05592317506670952, -0.036818087100982666, 0.19018739461898804, 0.21609260141849518, -0.1026768758893013, 0.01715746894478798, 0.02481461875140667, -0.07289334386587143, -0.19109952449798584, 0.038210321217775345, 0.05074393004179001, 0.011023870669305325, 0.028558019548654556, -0.18172858655452728, 0.1641886681318283, 0.11429160833358765, -0.017880264669656754, 0.11105671525001526, -0.3035794198513031, -0.12108085304498672, 0.13653801381587982, 0.1419295370578766, 0.12694154679775238, -0.12777802348136902, -0.007107523735612631, -0.017897093668580055, -0.13339795172214508, 0.1077156737446785, -0.09656116366386414, 0.1062578335404396, -0.035167451947927475, 0.09086983650922775, 0.0027249690610915422, -0.06266588717699051, 0.11299264430999756, 0.02041347697377205, 0.10140646994113922, -0.05786030367016792, -0.03522659093141556, 0.028812013566493988, -0.04393459111452103, 0.03080398589372635, -0.08340568840503693, 0.024242384359240532, -0.11295371502637863, -0.027507470920681953, -0.07522739470005035, 0.04683716595172882, -0.03678802028298378, -0.06714598834514618, -0.03710714355111122, 0.026461346074938774, 0.05497542768716812, -0.008879969827830791, 0.16179513931274414, 0.03838213533163071, 0.13131630420684814, 0.08457902818918228, 0.0801510289311409, -0.07342622429132462, -0.0693487673997879, -0.02356107532978058, -0.016268474981188774, 0.06812115758657455, -0.16219189763069153, 0.023766115307807922, 0.13838157057762146, 0.02440461330115795, 0.12714526057243347, 0.07803363353013992, -0.031716201454401016, 0.009282023645937443, 0.05586545914411545, -0.15610289573669434, -0.07713290303945541, -0.004131979309022427, -0.052021414041519165, -0.11201754212379456, 0.058185476809740067, 0.0875353291630745, -0.07972071319818497, -0.009900163859128952, -0.009764480404555798, 0.008701551705598831, -0.06237638369202614, 0.18229451775550842, 0.060920193791389465, 0.045599449425935745, -0.10174538195133209, 0.07340383529663086, 0.04789476469159126, -0.07494326680898666, 0.011265731416642666, 0.05669436603784561, -0.08135435730218887, -0.05201702192425728, 0.0700196698307991, 0.17686901986598969, -0.051887206733226776, -0.05851620435714722, -0.12699002027511597, -0.12122185528278351, 0.08282507956027985, 0.14811797440052032, 0.11332039535045624, 0.01393983419984579, -0.05551851913332939, 0.002806261647492647, -0.12314142286777496, 0.08821329474449158, 0.028557712212204933, 0.06614193320274353, -0.158407524228096, 0.12269162386655807, 0.00685510877519846, 0.05265906825661659, -0.015286439098417759, 0.030124682933092117, -0.10634183883666992, 0.014406777918338776, -0.1285777986049652, -0.01848839595913887, -0.05183494836091995, 0.013174296356737614, 0.0008219468290917575, -0.061842773109674454, -0.06278521567583084, 0.02002761699259281, -0.11289758235216141, -0.012642789632081985, 0.04161641374230385, 0.06310227513313293, -0.10506492108106613, -0.035643305629491806, 0.02397957257926464, -0.061036478728055954, 0.06789329648017883, 0.04707229137420654, 0.026572847738862038, 0.05731719359755516, -0.1272619068622589, 0.022958233952522278, 0.06920655071735382, 0.017951451241970062, 0.08110576868057251, -0.09864151477813721, -0.009644039906561375, 0.002128937281668186, 0.046040166169404984, 0.012956181541085243, 0.060228779911994934, -0.136691614985466, -0.014735862612724304, -0.001693789497949183, -0.08655042946338654, -0.060774534940719604, 0.02440808154642582, 0.10710575431585312, -0.0032192852813750505, 0.20515282452106476, -0.0641036182641983, 0.041000932455062866, -0.2098947912454605, 0.006183238700032234, -0.012758342549204826, -0.0986744835972786, -0.11466677486896515, -0.05476886034011841, 0.055531833320856094, -0.06379182636737823, 0.14200322329998016, 0.02256152778863907, 0.02944597229361534, 0.025515394285321236, -0.03242615982890129, 0.007186587434262037, 0.022251978516578674, 0.2075914740562439, 0.029662244021892548, -0.036906011402606964, 0.05689357593655586, 0.042344845831394196, 0.09912135452032089, 0.13871809840202332, 0.18433812260627747, 0.14505358040332794, 0.004330390132963657, 0.09853049367666245, 0.03893144056200981, -0.05747396498918533, -0.1647120714187622, 0.03072671964764595, -0.048108480870723724, 0.10721815377473831, -0.009794848039746284, 0.1958099752664566, 0.08676163107156754, -0.16496075689792633, 0.03258766606450081, -0.06015928462147713, -0.07774004340171814, -0.11633853614330292, -0.08260798454284668, -0.08497586101293564, -0.12555843591690063, 0.011793959885835648, -0.11604738980531693, 0.0035916182678192854, 0.11698733270168304, 0.01337617076933384, -0.028796769678592682, 0.15605878829956055, 0.039089854806661606, 0.039426665753126144, 0.0513903982937336, 0.014061380177736282, -0.042874936014413834, -0.10350725054740906, -0.0519987978041172, -0.021158691495656967, -0.033849943429231644, 0.026956556364893913, -0.07339663058519363, -0.05230464041233063, 0.04243633896112442, -0.007213371340185404, -0.09363409876823425, 0.009480916894972324, 0.006624981295317411, 0.0612935870885849, 0.03340050205588341, 0.011862779036164284, 0.030575193464756012, -0.012299008667469025, 0.1977318972349167, -0.07576285302639008, -0.05265802517533302, -0.1193670853972435, 0.24739181995391846, 0.03369578719139099, -0.020000673830509186, 0.035058293491601944, -0.05919549986720085, 0.008999735116958618, 0.25588878989219666, 0.20093491673469543, -0.07310285419225693, -0.010358650237321854, 0.0017890377203002572, -0.01269195880740881, -0.03849621117115021, 0.10462015867233276, 0.13613756000995636, 0.04359206557273865, -0.09633524715900421, -0.052021536976099014, -0.06149905547499657, -0.013981563039124012, -0.035110075026750565, 0.06985684484243393, 0.052551306784152985, 0.003433028468862176, -0.04339585080742836, 0.04611322283744812, -0.06655710935592651, -0.09213785082101822, 0.07613363862037659, -0.1916002631187439, -0.16343870759010315, -0.009855414740741253, 0.0917009711265564, 0.012965495698153973, 0.06418749690055847, -0.03062623180449009, -0.004265947733074427, 0.07961124181747437, -0.015325542539358139, -0.10415074229240417, -0.07749532163143158, 0.10289470106363297, -0.08620823919773102, 0.22684551775455475, -0.04197849705815315, 0.057771362364292145, 0.12714120745658875, 0.0692199245095253, -0.09198703616857529, 0.04582955315709114, 0.04459278658032417, -0.06726815551519394, 0.019602494314312935, 0.06961847841739655, -0.028734635561704636, 0.07142730057239532, 0.04197410121560097, -0.1326012760400772, 0.016140267252922058, -0.06122696027159691, -0.04535089060664177, -0.04518931359052658, -0.022984830662608147, -0.05580376461148262, 0.1388321816921234, 0.2168717384338379, -0.03323856741189957, -0.011137392371892929, -0.07433009892702103, 0.02395566552877426, 0.049475766718387604, 0.005361078307032585, -0.07094139605760574, -0.2067987620830536, 0.020042048767209053, 0.03778420761227608, -0.018887866288423538, -0.22498704493045807, -0.09162367135286331, 0.012024355120956898, -0.08657156676054001, -0.08956211060285568, 0.057352591305971146, 0.09284313768148422, 0.05813885107636452, -0.06354185193777084, -0.039627932012081146, -0.07225310057401657, 0.13633403182029724, -0.13497765362262726, -0.09960457682609558 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-base-gn-demo This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.7426 - Wer: 0.7256 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine_with_restarts - lr_scheduler_warmup_steps: 50 - num_epochs: 60 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 4.0 | 100 | 0.7045 | 0.7409 | | No log | 8.0 | 200 | 0.7200 | 0.75 | | No log | 12.0 | 300 | 0.7400 | 0.7439 | | No log | 16.0 | 400 | 0.7677 | 0.7515 | | 0.0846 | 20.0 | 500 | 0.7765 | 0.7271 | | 0.0846 | 24.0 | 600 | 0.7821 | 0.7287 | | 0.0846 | 28.0 | 700 | 0.7671 | 0.7180 | | 0.0846 | 32.0 | 800 | 0.7594 | 0.7180 | | 0.0846 | 36.0 | 900 | 0.7500 | 0.7165 | | 0.0713 | 40.0 | 1000 | 0.7351 | 0.7287 | | 0.0713 | 44.0 | 1100 | 0.7361 | 0.7241 | | 0.0713 | 48.0 | 1200 | 0.7389 | 0.7378 | | 0.0713 | 52.0 | 1300 | 0.7424 | 0.7210 | | 0.0713 | 56.0 | 1400 | 0.7425 | 0.7256 | | 0.0669 | 60.0 | 1500 | 0.7426 | 0.7256 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.2+cu102 - Datasets 1.18.3 - Tokenizers 0.10.3
{"language": ["gn"], "license": "apache-2.0", "tags": ["generated_from_trainer", "mozilla-foundation/common_voice_8_0", "robust-speech-event", "hf-asr-leaderboard"], "datasets": ["common_voice", "mozilla-foundation/common_voice_8_0"], "model-index": [{"name": "wav2vec2-base-gn-demo", "results": []}]}
automatic-speech-recognition
azuur/wav2vec2-base-gn-demo
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "mozilla-foundation/common_voice_8_0", "robust-speech-event", "hf-asr-leaderboard", "gn", "dataset:common_voice", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "gn" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #mozilla-foundation/common_voice_8_0 #robust-speech-event #hf-asr-leaderboard #gn #dataset-common_voice #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-base-gn-demo ===================== This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common\_voice dataset. It achieves the following results on the evaluation set: * Loss: 0.7426 * Wer: 0.7256 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: cosine\_with\_restarts * lr\_scheduler\_warmup\_steps: 50 * num\_epochs: 60 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.2+cu102 * Datasets 1.18.3 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 60\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #mozilla-foundation/common_voice_8_0 #robust-speech-event #hf-asr-leaderboard #gn #dataset-common_voice #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 60\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.10.3" ]
[ 121, 140, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #mozilla-foundation/common_voice_8_0 #robust-speech-event #hf-asr-leaderboard #gn #dataset-common_voice #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\\_with\\_restarts\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 60\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.3\n* Tokenizers 0.10.3" ]
[ -0.10688646882772446, 0.09505119919776917, -0.00515039311721921, 0.0431925430893898, 0.11434032022953033, 0.018510842695832253, 0.1408039927482605, 0.1513698697090149, -0.0821479931473732, 0.08104146271944046, 0.04119870066642761, 0.0706731453537941, 0.08011022955179214, 0.09863688796758652, -0.026617703959345818, -0.2775384783744812, 0.02913520485162735, -0.0200668778270483, -0.11188948154449463, 0.10311473160982132, 0.08301793783903122, -0.09621796011924744, 0.03699545934796333, 0.016232989728450775, -0.10856426507234573, 0.012577686458826065, -0.021295646205544472, -0.047551099210977554, 0.11430887877941132, 0.03668620064854622, 0.057457104325294495, 0.03963867202401161, 0.10990285873413086, -0.25250566005706787, 0.014764181338250637, 0.06509692221879959, 0.026960985735058784, 0.05432897061109543, 0.07332029938697815, -0.004131274297833443, 0.10599926859140396, -0.038009725511074066, 0.04454110935330391, 0.06455566734075546, -0.09981480985879898, -0.22859784960746765, -0.07050349563360214, 0.018004661425948143, 0.10250306129455566, 0.09226951748132706, -0.032148368656635284, 0.04463157802820206, -0.07759775966405869, 0.10460438579320908, 0.23702411353588104, -0.2280595749616623, -0.05946031212806702, 0.0015250496799126267, 0.04937219247221947, 0.022672336548566818, -0.11124703288078308, -0.0036534126847982407, 0.00944803562015295, 0.01688217557966709, 0.05252731963992119, -0.009075701236724854, -0.029767945408821106, -0.00236178794875741, -0.12060919404029846, -0.03607143461704254, 0.14864392578601837, 0.0506584532558918, 0.0010426163207739592, -0.10263379663228989, -0.04967986047267914, -0.16969439387321472, -0.043980423361063004, 0.010456874966621399, 0.027086166664958, -0.031200386583805084, -0.06524146348237991, 0.003050912171602249, -0.05677017942070961, -0.0897819772362709, 0.041797440499067307, 0.09817299246788025, 0.03663993999361992, -0.02112368680536747, -0.008262102492153645, 0.1023772805929184, 0.0771392285823822, -0.16320045292377472, -0.016623253002762794, 0.03349458798766136, -0.11092669516801834, -0.001595628447830677, -0.018680226057767868, 0.013036482967436314, 0.06286665797233582, 0.15739652514457703, -0.023118937388062477, 0.08749482780694962, -0.02586670033633709, 0.006548426579684019, -0.07651768624782562, 0.14587201178073883, -0.06239346042275429, -0.0784185454249382, -0.028702445328235626, 0.11530657857656479, -0.006388009060174227, -0.015111387707293034, -0.08217457681894302, 0.03216056153178215, 0.1005907952785492, 0.05859975144267082, -0.023407619446516037, 0.05505519360303879, -0.04310295358300209, -0.0255811158567667, -0.02536439709365368, -0.13526467978954315, 0.02993193082511425, 0.06696128100156784, -0.0707777887582779, -0.0096671087667346, 0.007015181705355644, 0.02102591283619404, -0.033739905804395676, 0.06584964692592621, -0.046101849526166916, -0.006079626269638538, -0.07215186953544617, -0.09986559301614761, 0.03541743382811546, -0.06478040665388107, 0.001924516400322318, -0.052399374544620514, -0.1157665029168129, -0.0745677575469017, 0.045088689774274826, -0.05601280927658081, -0.06330525130033493, -0.09322775900363922, -0.11215762048959732, 0.042702291160821915, -0.021812697872519493, 0.13045330345630646, -0.05649908259510994, 0.08358820527791977, 0.02419969253242016, 0.03716607764363289, 0.08982934802770615, 0.06927221268415451, -0.03451094776391983, 0.036520157009363174, -0.13342620432376862, 0.09946881979703903, -0.10878784954547882, 0.029521765187382698, -0.11119057983160019, -0.0923144519329071, -0.009243455715477467, 0.018156347796320915, 0.07392435520887375, 0.11492745578289032, -0.19186818599700928, -0.10235751420259476, 0.1735275685787201, -0.07843607664108276, -0.07580327242612839, 0.1177501454949379, -0.0181382168084383, -0.008906377479434013, 0.0323169007897377, 0.19578565657138824, 0.11785398423671722, -0.10890708118677139, -0.009208234027028084, -0.06009592115879059, 0.1101156547665596, 0.05689993128180504, 0.0487276166677475, -0.027699606493115425, 0.06510273367166519, 0.016478169709444046, -0.03378843143582344, 0.04846059903502464, -0.06659089773893356, -0.07912127673625946, -0.006592862773686647, -0.0768832340836525, -0.00010384244524175301, 0.06632731109857559, 0.006912219803780317, -0.06722395122051239, -0.11728379875421524, 0.03394210711121559, 0.09835284948348999, -0.12619607150554657, 0.03195156902074814, -0.09708987176418304, 0.033814944326877594, 0.02227741852402687, 0.0010503121884539723, -0.1530178338289261, -0.03141908720135689, 0.031561508774757385, -0.03370288759469986, 0.02208702266216278, 0.021722499281167984, 0.07965672016143799, 0.0371708907186985, -0.04331275075674057, -0.05741899833083153, -0.03855940327048302, -0.0019441463518887758, -0.03013586439192295, -0.24432969093322754, -0.061520613729953766, -0.026420660316944122, 0.15367156267166138, -0.1716475635766983, 0.023415764793753624, 0.058992575854063034, 0.13434167206287384, 0.025582194328308105, -0.04804610460996628, 0.0066337427124381065, 0.056758128106594086, -0.01871960610151291, -0.06669563800096512, 0.018399886786937714, 0.018481817096471786, -0.07939782738685608, 0.02549612522125244, -0.18467475473880768, 0.08310402184724808, 0.08853611350059509, 0.015865670517086983, -0.06281774491071701, -0.00802240613847971, -0.061267390847206116, -0.059793442487716675, -0.024788280948996544, -0.024064386263489723, 0.17034973204135895, 0.01321603637188673, 0.10652271658182144, -0.06575921922922134, -0.05073459446430206, 0.025437137112021446, 0.0199241004884243, 0.016568902879953384, 0.1331976056098938, 0.06220285966992378, -0.03707214444875717, 0.10435885936021805, -0.014513473957777023, -0.05553976818919182, 0.17701052129268646, -0.08351624757051468, -0.06706571578979492, -0.04248103126883507, -0.0020343796350061893, 0.01337443944066763, 0.1198047325015068, -0.19562457501888275, -0.02334628812968731, 0.01315425056964159, 0.019304074347019196, 0.02705548144876957, -0.18149632215499878, 0.003205179003998637, 0.02481914311647415, -0.10828028619289398, -0.0021320462692528963, 0.020862117409706116, 0.007701693568378687, 0.07774706929922104, -0.024182550609111786, -0.10059840232133865, -0.030680662021040916, -0.05028456076979637, -0.0925433337688446, 0.15751013159751892, -0.09755110740661621, -0.15369337797164917, -0.11314784735441208, -0.031604740768671036, -0.024629874154925346, -0.005801282823085785, 0.04243941232562065, -0.08737511187791824, -0.044444385915994644, -0.06061353161931038, 0.031138647347688675, -0.011248520575463772, 0.012911413796246052, 0.04224285110831261, 0.017161840572953224, 0.052270203828811646, -0.11381151527166367, 0.012573630549013615, -0.02521592006087303, -0.06269653141498566, -0.014206977561116219, 0.027630038559436798, 0.08784711360931396, 0.1637447625398636, 0.05903758108615875, 0.04529288783669472, -0.0276007279753685, 0.19380487501621246, -0.13272680342197418, 0.01440929900854826, 0.11172576248645782, -0.01501564309000969, 0.029520435258746147, 0.18276293575763702, 0.04984775185585022, -0.0659484937787056, 0.003427238669246435, 0.05301272124052048, -0.022527026012539864, -0.2386864721775055, -0.023343414068222046, -0.07064012438058853, -0.0070305801928043365, 0.06145340949296951, 0.036629728972911835, 0.04541076347231865, 0.0030350496526807547, -0.03251197561621666, -0.01622776873409748, 0.08476772904396057, 0.030393004417419434, 0.1151152104139328, 0.029579907655715942, 0.11718302965164185, -0.006795752327889204, -0.048157352954149246, 0.005754116922616959, 0.007111858110874891, 0.21933476626873016, 0.005771976429969072, 0.17709881067276, 0.049346908926963806, 0.11506334692239761, 0.00037051583058200777, 0.060691043734550476, 0.008316164836287498, 0.010361448861658573, 0.034233950078487396, -0.04783612862229347, -0.01321503147482872, 0.02217222936451435, 0.08672750741243362, 0.0030528728384524584, -0.07446414977312088, 0.04288480803370476, 0.033723678439855576, 0.3433581590652466, 0.05583817884325981, -0.24570409953594208, -0.07500480860471725, 0.004819868598133326, -0.06691453605890274, -0.028919614851474762, 0.041072092950344086, 0.1358053833246231, -0.0808011144399643, 0.07366973161697388, -0.06359745562076569, 0.07494327425956726, -0.07702658325433731, 0.029812708497047424, 0.09790754318237305, 0.08749248087406158, 0.015214175917208195, 0.048430558294057846, -0.26836860179901123, 0.2570020258426666, -0.004305493086576462, 0.07474049180746078, -0.044710081070661545, 0.04867951199412346, 0.02998574636876583, -0.028324464336037636, 0.09005764126777649, -0.024849770590662956, -0.15762048959732056, -0.15758047997951508, -0.08501410484313965, 0.01829957216978073, 0.13118793070316315, -0.04879482835531235, 0.1226138323545456, -0.054353903979063034, -0.041867680847644806, 0.05584355443716049, -0.07295911759138107, -0.12354038655757904, -0.1182120218873024, 0.049953073263168335, 0.027690773829817772, 0.06524647772312164, -0.07474736869335175, -0.09440138190984726, -0.07088987529277802, 0.13880537450313568, -0.13012418150901794, -0.03619858995079994, -0.12547840178012848, 0.04699765518307686, 0.16723912954330444, -0.0651983767747879, 0.029599493369460106, 0.03948044776916504, 0.12239154428243637, 0.0231308676302433, -0.03883665055036545, 0.09635665267705917, -0.08087089657783508, -0.21552471816539764, -0.0320734977722168, 0.16930174827575684, 0.040468331426382065, 0.0745687335729599, -0.008721676655113697, 0.028593575581908226, 0.00912979431450367, -0.059474289417266846, 0.06742680072784424, 0.05830831825733185, -0.022755002602934837, 0.04062419384717941, -0.056244660168886185, -0.012138775549829006, -0.07359714061021805, -0.0314330980181694, 0.11696408689022064, 0.21963518857955933, -0.0751354843378067, 0.08015910536050797, 0.027901172637939453, -0.08862446248531342, -0.17362350225448608, -0.011526232585310936, 0.10336507111787796, 0.04131372645497322, -0.021234558895230293, -0.223286435008049, 0.02404206432402134, 0.05201341584324837, -0.025022994726896286, 0.06728464365005493, -0.32195672392845154, -0.13390089571475983, 0.09865286201238632, 0.02928874082863331, -0.014495382085442543, -0.142862468957901, -0.05530459061264992, -0.012302515096962452, -0.07090596854686737, 0.04606322571635246, -0.03587062656879425, 0.13282211124897003, 0.00853657629340887, 0.045419320464134216, 0.024050196632742882, -0.04440324380993843, 0.10576570779085159, 0.03277014568448067, 0.048895906656980515, -0.020438358187675476, -0.0032515237107872963, 0.028628593310713768, -0.05363442376255989, 0.01149647869169712, -0.06423139572143555, 0.01859804056584835, -0.08714495599269867, -0.02550075203180313, -0.09752901643514633, 0.01883980818092823, -0.04419441893696785, -0.0191926509141922, -0.007266633678227663, 0.04478258639574051, 0.09082842618227005, 0.011920896358788013, 0.1165953129529953, -0.06487445533275604, 0.15012045204639435, 0.1447170078754425, 0.1326053887605667, -0.00011694826389430091, -0.07851200550794601, -0.005113481543958187, 0.0075133549980819225, 0.030667973682284355, -0.06944870948791504, 0.058981068432331085, 0.13972444832324982, 0.039353661239147186, 0.1397557258605957, 0.06368543952703476, -0.09809979051351547, 0.0036528639029711485, 0.05568811297416687, -0.10222136974334717, -0.15397779643535614, -0.030044497922062874, -0.03467719256877899, -0.12398205697536469, 0.017843490466475487, 0.14035069942474365, -0.03116011805832386, 0.004241549409925938, 0.024153107777237892, 0.05610106140375137, -0.021790681406855583, 0.2307928502559662, 0.024269770830869675, 0.08982986211776733, -0.1146012619137764, 0.06705395877361298, 0.03204246982932091, -0.07522004097700119, 0.05645986273884773, 0.09379781037569046, -0.05464671924710274, -0.021891681477427483, 0.008425094187259674, 0.10658735036849976, 0.03598806634545326, -0.05010605975985527, -0.1527453362941742, -0.1577688753604889, 0.06982747465372086, 0.13612458109855652, 0.011088358238339424, 0.04840053617954254, -0.030740756541490555, 0.04446713998913765, -0.09614390134811401, 0.12594182789325714, 0.09208577871322632, 0.05312706157565117, -0.13093404471874237, 0.11073515564203262, 0.0021270927973091602, 0.005685604177415371, -0.00737187871709466, 0.007079578470438719, -0.11398842185735703, 0.03054167330265045, -0.12183697521686554, -0.016529196873307228, -0.037822116166353226, -0.002956945914775133, 0.016783319413661957, -0.053028907626867294, -0.059951964765787125, 0.03270597383379936, -0.11225449293851852, -0.02811625227332115, -0.02701343223452568, 0.04763074219226837, -0.1200910285115242, -0.0044227298349142075, 0.03710828721523285, -0.12833330035209656, 0.0971977487206459, 0.05895278602838516, -0.009506184607744217, 0.037823233753442764, -0.100161612033844, -0.0377650111913681, 0.04416501894593239, 0.014744315296411514, 0.030453355982899666, -0.15466971695423126, -0.007635118905454874, -0.01692337729036808, 0.027540024369955063, -0.009037301875650883, 0.0131254056468606, -0.10382147878408432, 0.03030705638229847, -0.0395834818482399, -0.03657035902142525, -0.04874889552593231, 0.050591617822647095, 0.08086486160755157, 0.03000379353761673, 0.15530255436897278, -0.08910820633172989, 0.07218765467405319, -0.22483542561531067, 0.01466320175677538, 0.010387529619038105, -0.07844842970371246, -0.0487213097512722, -0.036833349615335464, 0.1060444563627243, -0.0531214214861393, 0.09230685979127884, 0.00023387007240671664, 0.054225143045186996, 0.021792855113744736, -0.09392647445201874, 0.00278747477568686, 0.053480155766010284, 0.14380645751953125, 0.049130432307720184, -0.0281178280711174, 0.10050585120916367, -0.034722816199064255, 0.049006398767232895, 0.08657989650964737, 0.16576403379440308, 0.11643624305725098, 0.061791837215423584, 0.0678502768278122, 0.13379913568496704, -0.13960103690624237, -0.10687454044818878, 0.12754520773887634, -0.07599887996912003, 0.13193540275096893, -0.04236208647489548, 0.19628870487213135, 0.09687238931655884, -0.17668357491493225, 0.08395691961050034, -0.04591239616274834, -0.08918950706720352, -0.14213553071022034, -0.08873207122087479, -0.08497361838817596, -0.1691886931657791, 0.027103714644908905, -0.1248808354139328, 0.060252845287323, 0.05620971694588661, 0.03927876427769661, 0.040958430618047714, 0.097907654941082, 0.051308564841747284, 0.009369784034788609, 0.09408650547266006, 0.011067233979701996, -0.027437930926680565, -0.06379523128271103, -0.07953143119812012, 0.05798361822962761, 0.011254411190748215, 0.05625690519809723, -0.035340651869773865, -0.08384902775287628, 0.05404981970787048, -0.01724303886294365, -0.09353043884038925, 0.02223294973373413, -0.013658401556313038, 0.0473346970975399, 0.06895841658115387, 0.038824569433927536, -0.02073695883154869, -0.001897657522931695, 0.2123231440782547, -0.0974249467253685, -0.0941854938864708, -0.13991589844226837, 0.1530712991952896, -0.0015070053050294518, 0.0049628750421106815, 0.024246951565146446, -0.06869525462388992, -0.016239585354924202, 0.1962280571460724, 0.196092426776886, -0.011954179033637047, -0.008393443189561367, 0.018541518598794937, -0.003345449920743704, -0.028464393690228462, 0.05464057996869087, 0.12446612119674683, 0.06974421441555023, -0.02663780003786087, -0.023074299097061157, -0.0012647020630538464, -0.0772433653473854, -0.016138767823576927, 0.08519089967012405, 0.008554281666874886, -0.007739742286503315, -0.013772484846413136, 0.10065833479166031, -0.11548903584480286, -0.16850152611732483, 0.01088785845786333, -0.1828320175409317, -0.16050076484680176, -0.03677467256784439, 0.09492479264736176, 0.0506102591753006, 0.04199584573507309, -0.008313165046274662, -0.038260236382484436, 0.14428597688674927, 0.0018796593649312854, -0.06434062868356705, -0.08742073178291321, 0.06394036114215851, -0.15346230566501617, 0.15505938231945038, -0.04797845333814621, 0.025182055309414864, 0.10016128420829773, 0.047155581414699554, -0.0750790685415268, 0.03029349073767662, 0.0726214125752449, -0.1656031310558319, 0.02191844768822193, 0.20710963010787964, -0.04342810437083244, 0.15068891644477844, 0.0345812663435936, -0.11287461966276169, 0.025966661050915718, -0.05287927761673927, -0.05129807069897652, -0.055786386132240295, -0.008848975412547588, -0.03615279495716095, 0.12963111698627472, 0.20169951021671295, -0.059439778327941895, 0.0036473218351602554, -0.04968373477458954, 0.0038864517118781805, 0.02952422760426998, 0.1117418184876442, -0.044685546308755875, -0.2836780548095703, 0.01044027041643858, -0.006461372599005699, 0.014319985173642635, -0.1822502762079239, -0.08795465528964996, 0.04230278357863426, -0.05871974304318428, -0.08235762268304825, 0.11249763518571854, 0.081666499376297, 0.05285228416323662, -0.04597948119044304, -0.07050973176956177, -0.004528617020696402, 0.18101495504379272, -0.19472892582416534, -0.06235068291425705 ]
null
null
transformers
#Ragnar Lothbrok DialoGPT Model
{"tags": ["conversational"]}
text-generation
b0shakk/DialoGPT-small-Ragnar
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#Ragnar Lothbrok DialoGPT Model
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
# shirt_identifier Autogenerated by HuggingPics๐Ÿค—๐Ÿ–ผ๏ธ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb). Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics). ## Example Images #### Big Check shirt ![Big Check shirt](images/Big_Check_shirt.jpg) #### Formal Shirt ![Formal Shirt](images/Formal_Shirt.jpg) #### casual shirt ![casual shirt](images/casual_shirt.jpg) #### denim shirt ![denim shirt](images/denim_shirt.jpg)
{"tags": ["image-classification", "pytorch", "huggingpics"], "metrics": ["accuracy"]}
image-classification
b25mayank3/shirt_identifier
[ "transformers", "pytorch", "tensorboard", "vit", "image-classification", "huggingpics", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #region-us
# shirt_identifier Autogenerated by HuggingPics๏ธ Create your own image classifier for anything by running the demo on Google Colab. Report any issues with the demo at the github repo. ## Example Images #### Big Check shirt !Big Check shirt #### Formal Shirt !Formal Shirt #### casual shirt !casual shirt #### denim shirt !denim shirt
[ "# shirt_identifier\n\n\nAutogenerated by HuggingPics๏ธ\n\nCreate your own image classifier for anything by running the demo on Google Colab.\n\nReport any issues with the demo at the github repo.", "## Example Images", "#### Big Check shirt\n\n!Big Check shirt", "#### Formal Shirt\n\n!Formal Shirt", "#### casual shirt\n\n!casual shirt", "#### denim shirt\n\n!denim shirt" ]
[ "TAGS\n#transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "# shirt_identifier\n\n\nAutogenerated by HuggingPics๏ธ\n\nCreate your own image classifier for anything by running the demo on Google Colab.\n\nReport any issues with the demo at the github repo.", "## Example Images", "#### Big Check shirt\n\n!Big Check shirt", "#### Formal Shirt\n\n!Formal Shirt", "#### casual shirt\n\n!casual shirt", "#### denim shirt\n\n!denim shirt" ]
[ 49, 44, 4, 11, 11, 10, 11 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #vit #image-classification #huggingpics #model-index #autotrain_compatible #endpoints_compatible #region-us \n# shirt_identifier\n\n\nAutogenerated by HuggingPics๏ธ\n\nCreate your own image classifier for anything by running the demo on Google Colab.\n\nReport any issues with the demo at the github repo.## Example Images#### Big Check shirt\n\n!Big Check shirt#### Formal Shirt\n\n!Formal Shirt#### casual shirt\n\n!casual shirt#### denim shirt\n\n!denim shirt" ]
[ -0.03138252720236778, 0.05613218992948532, -0.0012805734295397997, 0.0988122746348381, 0.11850672960281372, 0.03656696528196335, 0.11040925979614258, 0.1275201439857483, 0.08940932899713516, -0.04565230756998062, 0.14869830012321472, 0.16167841851711273, 0.05534561350941658, 0.20792777836322784, 0.03007795661687851, -0.1895153969526291, -0.09575142711400986, 0.0790141299366951, 0.11375850439071655, 0.10928841680288315, 0.07723957300186157, -0.09716899693012238, 0.17426443099975586, -0.007133326027542353, -0.25088444352149963, -0.04673813655972481, -0.020886609330773354, -0.058048103004693985, 0.11784316599369049, 0.03057112917304039, 0.07757572829723358, 0.04209150746464729, 0.004570163320749998, -0.03627926483750343, 0.054979197680950165, 0.015718959271907806, 0.00338978273794055, 0.05241960287094116, 0.08771141618490219, 0.12696970999240875, 0.06864359974861145, 0.11055053025484085, -0.05844299867749214, 0.056286659091711044, -0.07174825668334961, -0.11156293749809265, -0.07510465383529663, 0.10965417325496674, 0.05512649565935135, 0.010398798622190952, -0.006588766351342201, 0.21458321809768677, -0.11449100822210312, 0.05453547090291977, 0.18503788113594055, -0.10696723312139511, -0.045509763062000275, 0.09590549767017365, 0.07876439392566681, -0.10814501345157623, -0.06745083630084991, 0.04246358945965767, 0.04326564446091652, 0.057698335498571396, 0.03775352984666824, -0.07927573472261429, -0.07929074019193649, -0.05545556917786598, -0.037605270743370056, 0.08580058813095093, 0.146022230386734, 0.03738648071885109, -0.052881114184856415, -0.102090023458004, 0.03127581253647804, 0.04232342168688774, -0.027703896164894104, -0.09248684346675873, 0.010542171075940132, -0.0032959028612822294, -0.07576452195644379, -0.12961271405220032, -0.16896986961364746, 0.012377297505736351, -0.16077028214931488, -0.037601497024297714, -0.016987264156341553, 0.0028696029912680387, -0.12865059077739716, 0.03043697215616703, -0.08659068495035172, -0.04730388522148132, -0.021887633949518204, -0.08210697025060654, 0.015629740431904793, -0.12965331971645355, 0.050082605332136154, -0.0030909685883671045, 0.13104881346225739, 0.0725325345993042, 0.0012753031915053725, 0.001784709864296019, -0.04832093045115471, 0.05461075156927109, 0.06599365174770355, 0.18568819761276245, -0.0818135067820549, 0.030574318021535873, 0.0682232528924942, -0.0320783331990242, 0.044660355895757675, 0.04239465668797493, -0.15540874004364014, -0.04899114370346069, 0.05948233604431152, -0.026775600388646126, -0.008620655164122581, 0.06753907352685928, -0.030965182930231094, -0.0517255924642086, 0.2425779104232788, 0.010376554913818836, 0.03140201419591904, -0.03948837146162987, -0.02407151646912098, 0.012449781410396099, 0.049326349049806595, 0.07293302565813065, -0.001948135090060532, 0.040115032345056534, -0.05675356462597847, -0.013075369410216808, 0.028179852291941643, 0.04077010601758957, 0.004742234945297241, -0.23122963309288025, -0.0078401705250144, -0.11744566261768341, -0.044981710612773895, 0.001368425670079887, 0.023550309240818024, -0.030416196212172508, -0.07037315517663956, 0.012215277180075645, -0.036261849105358124, -0.07616832852363586, 0.06262267380952835, -0.013834387995302677, -0.016591424122452736, 0.05309423804283142, 0.048827268183231354, 0.14006821811199188, -0.07732753455638885, 0.011688116006553173, -0.17858394980430603, 0.014397836290299892, -0.14350880682468414, -0.001597253023646772, 0.03652643784880638, 0.1573612540960312, -0.10357359796762466, -0.05954902246594429, -0.108385369181633, 0.0491521954536438, -0.026793094351887703, 0.13833443820476532, -0.21123062074184418, -0.1466846615076065, 0.11177530139684677, -0.06441815197467804, -0.20092353224754333, 0.15222080051898956, 0.04589902237057686, 0.11717642098665237, 0.05865254998207092, 0.08112199604511261, 0.03604556620121002, -0.1308838129043579, -0.032185327261686325, 0.029013177379965782, -0.18482550978660583, 0.006495898123830557, -0.03681141510605812, 0.01460289116948843, -0.0005171566153876483, 0.02997615374624729, 0.05100857838988304, 0.04823312908411026, -0.053121909499168396, -0.05374020338058472, -0.02510049007833004, -0.0722975805401802, 0.0823960155248642, 0.09879984706640244, 0.015871545299887657, -0.06354932487010956, -0.03706597536802292, -0.12062160670757294, 0.0076925549656152725, 0.07636094838380814, -0.052933741360902786, -0.10943090170621872, 0.04078405722975731, 0.006652748677879572, -0.012608927674591541, 0.0007437232416123152, -0.07559345662593842, 0.037453360855579376, 0.08198096603155136, 0.1265920251607895, 0.04470695182681084, 0.09717131406068802, -0.012156709097325802, -0.046574462205171585, -0.021394209936261177, 0.11339625716209412, -0.0701623186469078, -0.06228881701827049, -0.0520385280251503, 0.04231682047247887, -0.020101385191082954, 0.11215867847204208, -0.16628344357013702, -0.039861250668764114, 0.024797026067972183, 0.11249691992998123, 0.11462252587080002, -0.012364855036139488, 0.06194460764527321, -0.048779211938381195, -0.05631977692246437, -0.06847472488880157, 0.12251546233892441, -0.030998151749372482, -0.0718708336353302, 0.12729474902153015, -0.012689572758972645, -0.03180941939353943, 0.1817696988582611, -0.14374695718288422, -0.16909965872764587, -0.030368447303771973, -0.050018567591905594, 0.012785437516868114, -0.0257758516818285, -0.014010202139616013, 0.03336936607956886, -0.003942824900150299, 0.04102920740842819, -0.02524062804877758, -0.048904068768024445, 0.07174450159072876, -0.057800643146038055, -0.10906749218702316, 0.011621125042438507, 0.19359229505062103, -0.15532167255878448, 0.07850997895002365, 0.16871456801891327, 0.005972142331302166, 0.13524824380874634, 0.03946409374475479, 0.05646240711212158, 0.00666848523542285, -0.07009946554899216, 0.038022082298994064, 0.06734325736761093, -0.08144988864660263, -0.009889300912618637, 0.02657652460038662, -0.1162547767162323, -0.011108584702014923, -0.11876236647367477, -0.035203736275434494, -0.006107859313488007, -0.01107816956937313, 0.17349065840244293, 0.1004972904920578, 0.015963638201355934, 0.05186983570456505, 0.05871700495481491, -0.016106300055980682, 0.0551326610147953, 0.05514994263648987, -0.05810622498393059, 0.07818233966827393, -0.013701318763196468, -0.32937362790107727, -0.09136757254600525, 0.038156136870384216, -0.031929392367601395, 0.06911949813365936, 0.052286166697740555, -0.1883382648229599, -0.023798299953341484, -0.07192254066467285, 0.05744152143597603, 0.11366751044988632, 0.0644516721367836, -0.13639074563980103, 0.023758798837661743, -0.07625214755535126, -0.04452358931303024, -0.052475739270448685, -0.024931831285357475, -0.1179635301232338, 0.12649138271808624, -0.1232398971915245, 0.036999139934778214, 0.08224305510520935, -0.02227143384516239, 0.05194483697414398, -0.0010015295119956136, 0.17708061635494232, -0.12432919442653656, 0.1537400782108307, 0.2119077742099762, 0.07611941546201706, 0.04447806254029274, 0.13519765436649323, -0.005950593855232, -0.09353144466876984, 0.06430818140506744, 0.054086606949567795, -0.10088372230529785, -0.055759575217962265, -0.015497192740440369, -0.012846909463405609, 0.09924110770225525, 0.08430619537830353, 0.0395362414419651, 0.16768208146095276, 0.10431180149316788, 0.010089155286550522, 0.06938742101192474, -0.03202959522604942, 0.011248398572206497, -0.07405157387256622, 0.035722438246011734, 0.045464616268873215, -0.026623234152793884, -0.1402577906847, 0.13729813694953918, -0.029728129506111145, 0.045455895364284515, 0.007962482050061226, 0.09214078634977341, -0.007849005982279778, 0.037987615913152695, 0.03451739624142647, -0.009386830031871796, -0.08833286911249161, -0.048428092151880264, 0.01371767558157444, -0.04288005083799362, 0.021823346614837646, -0.008955081924796104, 0.11595696955919266, -0.0729205533862114, 0.04982488602399826, -0.0878380760550499, 0.06232447177171707, 0.1126239001750946, 0.119560606777668, -0.299096941947937, -0.051372379064559937, -0.024654965847730637, 0.04170946404337883, -0.05112799257040024, -0.07050219923257828, 0.09039053320884705, -0.09212371706962585, 0.11524055898189545, 0.0037769563496112823, 0.057840414345264435, -0.09719191491603851, 0.08454213291406631, 0.07866054773330688, 0.10319539159536362, -0.002492237137630582, -0.0023816819302737713, -0.11100052297115326, 0.02503242716193199, -0.012089159339666367, 0.012474186718463898, -0.10330752283334732, -0.020524902269244194, 0.09807899594306946, 0.14124487340450287, 0.06912560760974884, 0.01964903622865677, 0.08553581684827805, -0.09246345609426498, 0.008619576692581177, 0.02278284542262554, 0.028618335723876953, -0.07852537930011749, -0.0060091097839176655, -0.035847827792167664, -0.07578493654727936, -0.0046595316380262375, 0.09261278808116913, -0.13929344713687897, 0.008780776523053646, 0.0406496524810791, -0.001276226481422782, -0.019403398036956787, 0.05687740072607994, -0.03433500602841377, -0.19391794502735138, 0.1470021903514862, 0.09887845069169998, -0.013242010027170181, -0.06715087592601776, -0.0398826003074646, 0.009029206819832325, -0.05302079766988754, -0.031280357390642166, -0.0588388592004776, 0.13521839678287506, 0.0024700963404029608, -0.10531768202781677, 0.08026210218667984, -0.04768991097807884, -0.07180999219417572, -0.05312369018793106, 0.05747000500559807, -0.04065924510359764, -0.008807064034044743, 0.09565204381942749, 0.009644911624491215, -0.030525457113981247, -0.055594608187675476, 0.1691090315580368, -0.016876745969057083, -0.008050855249166489, -0.04110044240951538, -0.00170657760463655, -0.04504922032356262, 0.0019942240323871374, -0.012151816859841347, 0.14393621683120728, 0.15212783217430115, -0.07200344651937485, 0.05477451533079147, -0.07158462703227997, -0.0024745159316807985, -0.32361793518066406, 0.015256204642355442, -0.017757682129740715, -0.016704019159078598, 0.06205258145928383, -0.09002909064292908, 0.15163715183734894, 0.09317798912525177, -0.016324147582054138, 0.24897883832454681, -0.0698811486363411, -0.1258501261472702, 0.0013643119018524885, 0.06795233488082886, 0.02973426692187786, -0.15056036412715912, -0.07421073317527771, -0.06950027495622635, -0.12110034376382828, 0.2512061893939972, 0.045326072722673416, 0.08776958286762238, -0.029484881088137627, 0.011741049587726593, 0.05027487501502037, 0.004373773001134396, 0.06277847290039062, -0.07705771178007126, -0.013047325424849987, -0.13579733669757843, -0.16110479831695557, -0.08560236543416977, -0.007539796642959118, 0.10588423907756805, -0.0032464636024087667, -0.022443030029535294, -0.10333170741796494, 0.006544208619743586, -0.10870499163866043, 0.1667362004518509, -0.00501254154369235, -0.05239409580826759, -0.04535357654094696, 0.08728496730327606, -0.11323866993188858, 0.001757577178068459, 0.07404010742902756, -0.0776715949177742, 0.1642725020647049, 0.15196983516216278, 0.15633779764175415, -0.0557592399418354, 0.08246266841888428, -0.01592704840004444, -0.05915440618991852, 0.04254840314388275, -0.06460849195718765, 0.09743008017539978, 0.04776196926832199, 0.010471578687429428, 0.0635562315583229, 0.029025644063949585, -0.002902377163991332, 0.04534708708524704, 0.13914212584495544, -0.05978057533502579, -0.11222808808088303, -0.08481334149837494, -0.06519667059183121, -0.056792549788951874, 0.09541837871074677, 0.07671483606100082, 0.026432838290929794, -0.07800545543432236, 0.016856065019965172, 0.05048351734876633, -0.03638947382569313, 0.023054486140608788, 0.13622848689556122, -0.040533121675252914, -0.060515061020851135, 0.04086092486977577, 0.03550039604306221, 0.011452416889369488, -0.06254605203866959, 0.15097196400165558, -0.06977253407239914, -0.13461804389953613, 0.2303866147994995, 0.13948120176792145, -0.03915117308497429, 0.0052409786731004715, 0.014063918963074684, -0.09017739444971085, 0.0019879965111613274, 0.031010214239358902, 0.08237291127443314, 0.05281693860888481, 0.11359991133213043, 0.012210139073431492, -0.0851544514298439, 0.046895213425159454, -0.0030879618134349585, 0.12053129076957703, -0.18075716495513916, 0.010720682330429554, -0.0313037745654583, 0.1570529192686081, -0.09106139093637466, -0.0001901793002616614, -0.14990073442459106, -0.07340540736913681, 0.0000034881306874012807, 0.03272590413689613, -0.04734882339835167, -0.013345700688660145, 0.02633226104080677, 0.00029865617398172617, -0.001899517374113202, -0.05362578481435776, -0.09988543391227722, -0.030397990718483925, -0.007606952916830778, 0.0676756277680397, -0.03951558843255043, -0.014688807539641857, 0.10724879801273346, -0.008800187148153782, 0.1771910935640335, -0.0360415019094944, -0.041369907557964325, -0.0026484401896595955, -0.1596333235502243, -0.05685507133603096, 0.057741764932870865, -0.010608394630253315, 0.02933640405535698, 0.021295398473739624, -0.0005134172970429063, -0.03179680556058884, -0.0015707486309111118, 0.003663033479824662, 0.2119070440530777, -0.05821554362773895, 0.040639374405145645, -0.04552178084850311, -0.1402081400156021, -0.02356783300638199, 0.0667559802532196, 0.1309075802564621, -0.02244560979306698, 0.09829575568437576, -0.07083309441804886, -0.008636881597340107, -0.12398391216993332, 0.007849348708987236, -0.04803195223212242, -0.12198945134878159, -0.04026776924729347, -0.03659067302942276, 0.008683498948812485, 0.044789887964725494, -0.05838401988148689, 0.10742393136024475, 0.08387578278779984, -0.008808004669845104, 0.0728156715631485, -0.05922018364071846, 0.026327351108193398, 0.06597369909286499, -0.06537825614213943, -0.05555505305528641, -0.060067374259233475, 0.06535140424966812, 0.0010616828221827745, -0.10115458816289902, 0.08207888901233673, -0.02006623148918152, -0.14977812767028809, 0.019797077402472496, 0.11931328475475311, 0.02538483403623104, -0.07229987531900406, 0.09962330758571625, -0.02966429851949215, 0.06163523346185684, -0.04729516804218292, 0.06116780266165733, 0.1080479621887207, -0.0960085466504097, 0.054181620478630066, -0.05601707473397255, -0.0160951130092144, -0.06980656832456589, -0.21757389605045319, -0.13233689963817596, -0.127160906791687, 0.02460833266377449, -0.011967639438807964, -0.044897787272930145, -0.019476765766739845, 0.033969320356845856, -0.1147337406873703, 0.09576664119958878, -0.04533833637833595, -0.0703582689166069, 0.06845976412296295, -0.04354226961731911, -0.16597148776054382, -0.05085122585296631, 0.016484268009662628, -0.057044658809900284, 0.11719340831041336, 0.010619306936860085, -0.015905285254120827, -0.0678391084074974, -0.005820297170430422, -0.07602006942033768, -0.11361633986234665, -0.013222461566329002, -0.028856471180915833, -0.03722185641527176, -0.19223237037658691, 0.002975190756842494, 0.033359594643116, 0.02030486986041069, 0.1395193487405777, -0.019042976200580597, 0.04648330807685852, -0.09303723275661469, 0.15679392218589783, -0.12827424705028534, 0.012120689265429974, -0.02827296033501625, -0.054476093500852585, -0.008174888789653778, 0.3174155056476593, 0.17024336755275726, -0.07670419663190842, 0.0030390857718884945, 0.024038545787334442, 0.010730437003076077, -0.022819073870778084, 0.10246697813272476, 0.06325102597475052, 0.012740172445774078, -0.06009947508573532, 0.04647421836853027, 0.013591966591775417, -0.021731015294790268, 0.01282535120844841, -0.036200959235429764, 0.07744075357913971, 0.04837013781070709, -0.07124314457178116, 0.0830589085817337, -0.13360607624053955, -0.10800882428884506, 0.173357754945755, -0.17740529775619507, -0.0758528858423233, 0.01245159562677145, 0.0440874882042408, -0.015465091913938522, 0.11564730852842331, -0.038215287029743195, 0.08792073279619217, -0.061003901064395905, -0.019443649798631668, -0.15114997327327728, -0.0012082681059837341, -0.0443529412150383, -0.1254996955394745, 0.14738835394382477, -0.04871672764420509, -0.0364689864218235, 0.06530887633562088, -0.015989525243639946, -0.09097082167863846, 0.010099184699356556, -0.02808724343776703, 0.004116851836442947, 0.0056869881227612495, 0.11747153848409653, -0.0013541373191401362, -0.09601014107465744, 0.03315739333629608, -0.14727887511253357, 0.0030233445577323437, 0.02734518051147461, 0.011129211634397507, -0.05114548280835152, 0.17339737713336945, -0.13739106059074402, 0.10421328991651535, 0.06182974949479103, -0.0036654816940426826, -0.03297249972820282, -0.0525880791246891, 0.026257574558258057, 0.06652012467384338, -0.06662624329328537, -0.0996522605419159, -0.06495995819568634, -0.0018962806789204478, -0.009984895586967468, -0.050486717373132706, -0.08985564112663269, 0.018009694293141365, -0.136211097240448, 0.024792606011033058, -0.08047597855329514, 0.08046259731054306, 0.15777313709259033, 0.012441269122064114, 0.018868733197450638, 0.027950365096330643, 0.009467163123190403, 0.07447315752506256, -0.116397425532341, -0.06169818341732025 ]
null
null
transformers
# GPT-Neo 125M finetuned with beer recipes ## Model Description GPT-Neo 125M is a transformer model based on EleutherAI's replication of the GPT-3 architecture https://huggingface.co/EleutherAI/gpt-neo-125M. It generates recipes for brewing beer in a YAML-like format which can be easily used for different purposes. ## Training data This model was trained on a custom dataset of ~ 76,800 beer recipes from the internet. It includes recipes for the following styles of beer: * Strong American Ale * Pale American Ale * India Pale Ale (IPA) * Standard American Beer * Stout * English Pale Ale * IPA * American Porter and Stout * Sour Ale * Irish Beer * Strong British Ale * Belgian and French Ale * German Wheat and Rye Beer * Czech Lager * Spice/Herb/Vegetable Beer * Specialty Beer * American Ale * Pilsner * Belgian Ale * Strong Belgian Ale * Bock * Brown British Beer * German Wheat Beer * Fruit Beer * Amber Malty European Lager * Pale Malty European Lager * British Bitter * Amber and Brown American Beer * Light Hybrid Beer * Pale Commonwealth Beer * American Wild Ale * European Amber Lager * Belgian Strong Ale * International Lager * Amber Bitter European Lager * Light Lager * Scottish and Irish Ale * European Sour Ale * Trappist Ale * Strong European Beer * Porter * Historical Beer * Pale Bitter European Beer * Amber Hybrid Beer * Smoke Flavored/Wood-Aged Beer * Spiced Beer * Dark European Lager * Alternative Fermentables Beer * Mead * Strong Ale * Dark British Beer * Scottish Ale * Smoked Beer * English Brown Ale * Dark Lager * Cider or Perry * Wood Beer ### How to use You can use this model directly with a pipeline for text generation. This example generates a different recipe each time it's run: ```py >>> from transformers import pipeline >>> generator = pipeline('text-generation', model='b3ck1/gpt-neo-125M-finetuned-beer-recipes') >>> generator("style: Pilsner\nbatch_size: 20\nefficiency: 75\nboil_size:", do_sample=True, min_length=50, max_length=500) >>> print(output[0]['generated_text']) style: Pilsner batch_size: 20 efficiency: 70 boil_size: 24 boil_time: 60 fermentables: - name: Pale Ale type: Grain amount: 6.5 hops: - name: Saaz alpha: 3.5 use: Boil time: 60 amount: 0.06 - name: Saaz alpha: 3.5 use: Boil time: 30 amount: 0.06 - name: Saaz alpha: 3.5 use: Boil time: 10 amount: 0.06 - name: Saaz alpha: 3.5 use: Boil time: 0 amount: 0.06 yeasts: - name: Safale - American Ale Yeast US-05 amount: 0.11 min_temperature: 12 max_temperature: 25 primary_temp: null mash_steps: - step_temp: 65 step_time: 60 miscs: [] ``` ### See this model in action This model was used to build https://beerai.net.
{"language": ["en"], "license": "apache-2.0", "tags": ["text generation", "pytorch", "causal-lm"], "datasets": ["custom"], "widget": [{"text": "style: Pilsner\nbatch_size: 20\nefficiency: 75\nboil_size:", "example_title": "Pilsener"}, {"text": "style: IPA\nbatch_size: 20\nefficiency: 75\nboil_size:", "example_title": "IPA"}, {"text": "style: Scottish Ale\nbatch_size: 20\nefficiency: 75\nboil_size:", "example_title": "Scottish Ale"}], "inference": {"parameters": {"do_sample": true, "top_k": 10, "top_p": 0.99, "max_length": 500}}}
text-generation
b3ck1/gpt-neo-125M-finetuned-beer-recipes
[ "transformers", "pytorch", "gpt_neo", "text-generation", "text generation", "causal-lm", "en", "dataset:custom", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt_neo #text-generation #text generation #causal-lm #en #dataset-custom #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# GPT-Neo 125M finetuned with beer recipes ## Model Description GPT-Neo 125M is a transformer model based on EleutherAI's replication of the GPT-3 architecture URL It generates recipes for brewing beer in a YAML-like format which can be easily used for different purposes. ## Training data This model was trained on a custom dataset of ~ 76,800 beer recipes from the internet. It includes recipes for the following styles of beer: * Strong American Ale * Pale American Ale * India Pale Ale (IPA) * Standard American Beer * Stout * English Pale Ale * IPA * American Porter and Stout * Sour Ale * Irish Beer * Strong British Ale * Belgian and French Ale * German Wheat and Rye Beer * Czech Lager * Spice/Herb/Vegetable Beer * Specialty Beer * American Ale * Pilsner * Belgian Ale * Strong Belgian Ale * Bock * Brown British Beer * German Wheat Beer * Fruit Beer * Amber Malty European Lager * Pale Malty European Lager * British Bitter * Amber and Brown American Beer * Light Hybrid Beer * Pale Commonwealth Beer * American Wild Ale * European Amber Lager * Belgian Strong Ale * International Lager * Amber Bitter European Lager * Light Lager * Scottish and Irish Ale * European Sour Ale * Trappist Ale * Strong European Beer * Porter * Historical Beer * Pale Bitter European Beer * Amber Hybrid Beer * Smoke Flavored/Wood-Aged Beer * Spiced Beer * Dark European Lager * Alternative Fermentables Beer * Mead * Strong Ale * Dark British Beer * Scottish Ale * Smoked Beer * English Brown Ale * Dark Lager * Cider or Perry * Wood Beer ### How to use You can use this model directly with a pipeline for text generation. This example generates a different recipe each time it's run: ### See this model in action This model was used to build URL.
[ "# GPT-Neo 125M finetuned with beer recipes", "## Model Description\n\nGPT-Neo 125M is a transformer model based on EleutherAI's replication of the GPT-3 architecture URL\nIt generates recipes for brewing beer in a YAML-like format which can be easily used for different purposes.", "## Training data\n\nThis model was trained on a custom dataset of ~ 76,800 beer recipes from the internet. It includes recipes for the following \nstyles of beer:\n\n* Strong American Ale \n* Pale American Ale\n* India Pale Ale (IPA)\n* Standard American Beer\n* Stout\n* English Pale Ale\n* IPA\n* American Porter and Stout\n* Sour Ale\n* Irish Beer\n* Strong British Ale\n* Belgian and French Ale\n* German Wheat and Rye Beer\n* Czech Lager\n* Spice/Herb/Vegetable Beer\n* Specialty Beer\n* American Ale\n* Pilsner\n* Belgian Ale\n* Strong Belgian Ale\n* Bock\n* Brown British Beer\n* German Wheat Beer\n* Fruit Beer\n* Amber Malty European Lager\n* Pale Malty European Lager\n* British Bitter\n* Amber and Brown American Beer\n* Light Hybrid Beer\n* Pale Commonwealth Beer\n* American Wild Ale\n* European Amber Lager\n* Belgian Strong Ale\n* International Lager\n* Amber Bitter European Lager\n* Light Lager\n* Scottish and Irish Ale\n* European Sour Ale\n* Trappist Ale\n* Strong European Beer\n* Porter\n* Historical Beer\n* Pale Bitter European Beer\n* Amber Hybrid Beer\n* Smoke Flavored/Wood-Aged Beer\n* Spiced Beer\n* Dark European Lager\n* Alternative Fermentables Beer\n* Mead\n* Strong Ale\n* Dark British Beer\n* Scottish Ale\n* Smoked Beer\n* English Brown Ale\n* Dark Lager\n* Cider or Perry\n* Wood Beer", "### How to use\n\nYou can use this model directly with a pipeline for text generation. This example generates a different recipe each time it's run:", "### See this model in action\n\nThis model was used to build URL." ]
[ "TAGS\n#transformers #pytorch #gpt_neo #text-generation #text generation #causal-lm #en #dataset-custom #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# GPT-Neo 125M finetuned with beer recipes", "## Model Description\n\nGPT-Neo 125M is a transformer model based on EleutherAI's replication of the GPT-3 architecture URL\nIt generates recipes for brewing beer in a YAML-like format which can be easily used for different purposes.", "## Training data\n\nThis model was trained on a custom dataset of ~ 76,800 beer recipes from the internet. It includes recipes for the following \nstyles of beer:\n\n* Strong American Ale \n* Pale American Ale\n* India Pale Ale (IPA)\n* Standard American Beer\n* Stout\n* English Pale Ale\n* IPA\n* American Porter and Stout\n* Sour Ale\n* Irish Beer\n* Strong British Ale\n* Belgian and French Ale\n* German Wheat and Rye Beer\n* Czech Lager\n* Spice/Herb/Vegetable Beer\n* Specialty Beer\n* American Ale\n* Pilsner\n* Belgian Ale\n* Strong Belgian Ale\n* Bock\n* Brown British Beer\n* German Wheat Beer\n* Fruit Beer\n* Amber Malty European Lager\n* Pale Malty European Lager\n* British Bitter\n* Amber and Brown American Beer\n* Light Hybrid Beer\n* Pale Commonwealth Beer\n* American Wild Ale\n* European Amber Lager\n* Belgian Strong Ale\n* International Lager\n* Amber Bitter European Lager\n* Light Lager\n* Scottish and Irish Ale\n* European Sour Ale\n* Trappist Ale\n* Strong European Beer\n* Porter\n* Historical Beer\n* Pale Bitter European Beer\n* Amber Hybrid Beer\n* Smoke Flavored/Wood-Aged Beer\n* Spiced Beer\n* Dark European Lager\n* Alternative Fermentables Beer\n* Mead\n* Strong Ale\n* Dark British Beer\n* Scottish Ale\n* Smoked Beer\n* English Brown Ale\n* Dark Lager\n* Cider or Perry\n* Wood Beer", "### How to use\n\nYou can use this model directly with a pipeline for text generation. This example generates a different recipe each time it's run:", "### See this model in action\n\nThis model was used to build URL." ]
[ 64, 15, 59, 307, 33, 15 ]
[ "passage: TAGS\n#transformers #pytorch #gpt_neo #text-generation #text generation #causal-lm #en #dataset-custom #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# GPT-Neo 125M finetuned with beer recipes## Model Description\n\nGPT-Neo 125M is a transformer model based on EleutherAI's replication of the GPT-3 architecture URL\nIt generates recipes for brewing beer in a YAML-like format which can be easily used for different purposes.## Training data\n\nThis model was trained on a custom dataset of ~ 76,800 beer recipes from the internet. It includes recipes for the following \nstyles of beer:\n\n* Strong American Ale \n* Pale American Ale\n* India Pale Ale (IPA)\n* Standard American Beer\n* Stout\n* English Pale Ale\n* IPA\n* American Porter and Stout\n* Sour Ale\n* Irish Beer\n* Strong British Ale\n* Belgian and French Ale\n* German Wheat and Rye Beer\n* Czech Lager\n* Spice/Herb/Vegetable Beer\n* Specialty Beer\n* American Ale\n* Pilsner\n* Belgian Ale\n* Strong Belgian Ale\n* Bock\n* Brown British Beer\n* German Wheat Beer\n* Fruit Beer\n* Amber Malty European Lager\n* Pale Malty European Lager\n* British Bitter\n* Amber and Brown American Beer\n* Light Hybrid Beer\n* Pale Commonwealth Beer\n* American Wild Ale\n* European Amber Lager\n* Belgian Strong Ale\n* International Lager\n* Amber Bitter European Lager\n* Light Lager\n* Scottish and Irish Ale\n* European Sour Ale\n* Trappist Ale\n* Strong European Beer\n* Porter\n* Historical Beer\n* Pale Bitter European Beer\n* Amber Hybrid Beer\n* Smoke Flavored/Wood-Aged Beer\n* Spiced Beer\n* Dark European Lager\n* Alternative Fermentables Beer\n* Mead\n* Strong Ale\n* Dark British Beer\n* Scottish Ale\n* Smoked Beer\n* English Brown Ale\n* Dark Lager\n* Cider or Perry\n* Wood Beer### How to use\n\nYou can use this model directly with a pipeline for text generation. This example generates a different recipe each time it's run:### See this model in action\n\nThis model was used to build URL." ]
[ -0.05138006806373596, 0.09788857400417328, -0.0036381310783326626, 0.035594966262578964, 0.042012784630060196, 0.05534058436751366, 0.09942968934774399, 0.04458291083574295, 0.13875900208950043, 0.11461103707551956, -0.029648706316947937, -0.016917986795306206, 0.0849643275141716, 0.12265969812870026, 0.10826718807220459, -0.1674015372991562, 0.036027371883392334, -0.10553622990846634, 0.02785572223365307, 0.06676670163869858, 0.08552197366952896, -0.043842945247888565, 0.04774772748351097, -0.06090201064944267, -0.0243307463824749, 0.049756694585084915, -0.06490012258291245, -0.024885213002562523, 0.005882767029106617, -0.0002262345515191555, 0.06463160365819931, -0.0376150943338871, -0.008549576625227928, -0.2509196996688843, 0.019711634144186974, 0.011664032004773617, -0.011394962668418884, -0.0006757593364454806, 0.0736897811293602, -0.07252904772758484, 0.08356236666440964, -0.08614369481801987, 0.0639435276389122, 0.09411603957414627, -0.11305549740791321, -0.16775144636631012, -0.09606927633285522, 0.20338819921016693, 0.09847867488861084, 0.03341842442750931, -0.0796855241060257, -0.15902329981327057, -0.09402289986610413, -0.043940991163253784, 0.227031871676445, -0.06359462440013885, -0.028667941689491272, -0.05996517091989517, -0.0014093326171860099, 0.0774998664855957, -0.10278090089559555, 0.03149522840976715, -0.02809304930269718, 0.06727518141269684, -0.025784708559513092, -0.008854107931256294, 0.11701207607984543, 0.022465277463197708, -0.04959278553724289, -0.06299860030412674, -0.014765576459467411, 0.0033951103687286377, -0.04970713332295418, -0.21597270667552948, 0.06705094128847122, -0.026299232617020607, -0.13250039517879486, -0.07378732413053513, 0.01864822767674923, 0.002376174321398139, 0.0691491961479187, -0.08427313715219498, -0.05642997473478317, -0.017859825864434242, 0.030961142852902412, 0.21053430438041687, 0.03641002997756004, -0.060918696224689484, 0.10335540026426315, 0.07078447937965393, -0.0035995645448565483, -0.053759075701236725, -0.030433766543865204, -0.003611884079873562, -0.16314731538295746, 0.0019457307644188404, -0.022669579833745956, 0.046495623886585236, 0.03782537579536438, 0.23572245240211487, -0.09734924137592316, 0.02567141130566597, 0.009136020205914974, -0.03340485319495201, -0.027948690578341484, 0.19727718830108643, -0.046384334564208984, -0.15507562458515167, -0.09487703442573547, -0.08229691535234451, -0.06381339579820633, 0.016183484345674515, -0.05198473855853081, 0.0859932079911232, 0.03296620398759842, 0.039992090314626694, 0.019268447533249855, 0.000704197445884347, -0.16249890625476837, -0.07003305852413177, 0.2208920568227768, -0.0757199227809906, 0.0901809111237526, 0.1014523133635521, -0.01546503510326147, 0.009235674515366554, -0.04457928240299225, 0.09898511320352554, -0.056344881653785706, 0.10863129794597626, -0.005679043475538492, -0.012234868481755257, -0.07693006098270416, -0.017761990427970886, 0.024806303903460503, 0.06448622792959213, -0.06308245658874512, -0.008608766831457615, -0.14749231934547424, -0.10019733011722565, 0.06930845975875854, -0.0718231126666069, -0.012594734318554401, -0.08681093156337738, -0.019573450088500977, 0.029676401987671852, 0.03086327575147152, -0.06481251865625381, -0.012068379670381546, -0.006604836788028479, -0.161518856883049, -0.005738761276006699, 0.06104394793510437, 0.07529579848051071, -0.11363855004310608, -0.015642844140529633, -0.36937087774276733, 0.17413602769374847, -0.15039964020252228, 0.08234789967536926, -0.17035041749477386, -0.009467700496315956, 0.006397726945579052, 0.055538300424814224, -0.05528721213340759, 0.1126277968287468, -0.10432954132556915, -0.10425762832164764, 0.05242181941866875, -0.08527769148349762, -0.034854475408792496, 0.15522974729537964, 0.011772828176617622, 0.00001017344584397506, 0.14317776262760162, 0.17569270730018616, 0.05847695469856262, -0.05090079456567764, -0.013527977280318737, -0.10505926609039307, -0.02893351949751377, 0.2349548190832138, 0.09831782430410385, -0.16331303119659424, 0.08372944593429565, -0.00554842920973897, 0.0010196418734267354, 0.009216276928782463, 0.005347601603716612, -0.04697649925947189, 0.04642422869801521, -0.04399934783577919, 0.09249728918075562, 0.010526597499847412, -0.04314933344721794, -0.023380188271403313, -0.13072456419467926, 0.1399248093366623, 0.04252620041370392, -0.004245496354997158, 0.020752668380737305, -0.060105811804533005, 0.05812641605734825, 0.06491706520318985, -0.016870317980647087, -0.03475954383611679, -0.1101183369755745, 0.059962198138237, -0.09880203753709793, 0.16251079738140106, -0.06550037860870361, 0.05718261003494263, 0.03029056079685688, -0.03822312131524086, -0.10755057632923126, -0.01194838061928749, -0.0038560200482606888, -0.024031467735767365, -0.1972305327653885, 0.03469514846801758, 0.043253522366285324, 0.17364507913589478, -0.014323428273200989, -0.04731877148151398, 0.1403411626815796, 0.17270465195178986, 0.07277598977088928, -0.06422499567270279, -0.006684722378849983, 0.033028360456228256, 0.03863365575671196, -0.035480331629514694, 0.02270674891769886, -0.09749509394168854, -0.017280789092183113, 0.1154627799987793, 0.033609941601753235, -0.09362128376960754, 0.0511074997484684, 0.11114125698804855, -0.09051358699798584, -0.05935852974653244, -0.026139602065086365, -0.037426188588142395, 0.04219978675246239, -0.10359174758195877, 0.21806824207305908, 0.03206931799650192, -0.002612067386507988, -0.06992404907941818, -0.10659533739089966, -0.04663765802979469, 0.008145908825099468, -0.08385007083415985, 0.15505585074424744, -0.09128610789775848, -0.14215797185897827, 0.08840295672416687, 0.05639166012406349, 0.169586181640625, 0.2148926854133606, 0.025634972378611565, -0.10542148351669312, -0.025020234286785126, 0.027235472574830055, 0.04593224450945854, 0.09231202304363251, -0.006130638066679239, 0.026615124195814133, -0.007933747954666615, -0.01366871315985918, 0.007449110969901085, -0.006690515670925379, 0.050193995237350464, -0.024877455085515976, -0.05759350210428238, 0.1380121260881424, 0.0013002032646909356, 0.09210386127233505, 0.05633252114057541, 0.08422067016363144, 0.02650102786719799, -0.03300433233380318, -0.028740527108311653, -0.09224561601877213, 0.10433118790388107, -0.19295728206634521, -0.13487966358661652, -0.1295837163925171, 0.0017777620814740658, -0.010832260362803936, 0.0018788460874930024, 0.05758640170097351, -0.07105223834514618, -0.0761641263961792, -0.0688798800110817, 0.2300335168838501, 0.05727289989590645, -0.13098183274269104, -0.12579652667045593, -0.05145043507218361, 0.017034849151968956, -0.09063862264156342, 0.018363608047366142, 0.03602251783013344, -0.08087107539176941, -0.047494109719991684, -0.023942045867443085, 0.10829237848520279, 0.07087850570678711, 0.05150827020406723, -0.014108527451753616, -0.03194068372249603, 0.18974338471889496, -0.1926441788673401, 0.11499695479869843, -0.07268817722797394, -0.03702373802661896, 0.06378701329231262, 0.11742107570171356, 0.05496608838438988, -0.03264901414513588, -0.002962705912068486, 0.0645872950553894, 0.01951921172440052, -0.16711243987083435, -0.1306367665529251, -0.012671565636992455, 0.007581827696412802, 0.04011015594005585, 0.09518556296825409, 0.07434321194887161, -0.020125944167375565, -0.16424818336963654, -0.036153797060251236, 0.028336266055703163, 0.06297862529754639, 0.06245061010122299, 0.023494157940149307, -0.04026509076356888, -0.011718299239873886, 0.004562995862215757, 0.12218181043863297, -0.026572968810796738, 0.09592509269714355, 0.0019497564062476158, 0.27831172943115234, 0.08659875392913818, 0.05235115438699722, -0.01676928997039795, -0.022490642964839935, -0.04104509949684143, -0.010550595819950104, 0.0020853665191680193, -0.06918808072805405, 0.10833229124546051, 0.12546776235103607, 0.07970842719078064, -0.05301094427704811, 0.03236781060695648, -0.11184030771255493, 0.14787480235099792, 0.18203610181808472, -0.02556145004928112, -0.01984458975493908, -0.021685203537344933, 0.05689668655395508, -0.10132183879613876, -0.03209690377116203, -0.025390714406967163, -0.013213577680289745, -0.11308003216981888, 0.04653364419937134, -0.017362533137202263, 0.06916005909442902, -0.004095283802598715, -0.026111401617527008, 0.10969146341085434, 0.10360388457775116, 0.0026518043596297503, 0.0981520414352417, -0.2613433599472046, 0.027509447187185287, 0.009373515844345093, 0.04562588408589363, -0.05657898262143135, 0.038014769554138184, -0.005189551971852779, -0.0843866616487503, 0.13886308670043945, 0.06847716867923737, -0.06727585196495056, -0.12062256783246994, -0.10275907069444656, -0.07433024048805237, 0.0947173610329628, -0.2522171437740326, 0.048421747982501984, 0.0263786930590868, -0.012974194251000881, -0.08926162868738174, -0.05942189320921898, -0.18611456453800201, -0.13148272037506104, 0.013697476126253605, -0.15847523510456085, 0.13189616799354553, -0.056698139756917953, -0.005519772879779339, -0.17802174389362335, 0.14507411420345306, -0.1829942911863327, -0.0861884355545044, -0.08412685990333557, -0.11837491393089294, 0.1365048885345459, -0.12120189517736435, 0.04214729368686676, 0.03917963430285454, 0.11284494400024414, -0.07001417130231857, -0.01769919879734516, 0.07756365090608597, -0.061133865267038345, -0.22871901094913483, -0.03743734210729599, 0.1033586785197258, 0.16612732410430908, 0.04885946214199066, 0.032575611025094986, 0.08119377493858337, 0.033028438687324524, -0.0975155308842659, -0.012688839808106422, 0.0030379313975572586, 0.0018203116487711668, 0.13636614382266998, -0.014117070473730564, -0.0881151482462883, -0.11035289615392685, -0.03210987150669098, 0.010215295478701591, 0.10020692646503448, -0.029525723308324814, 0.21410220861434937, 0.07391881197690964, -0.1052178218960762, -0.194255068898201, -0.07317643612623215, 0.030551670119166374, -0.06388381123542786, 0.02151079848408699, -0.2183842957019806, 0.12218544632196426, 0.03822698816657066, -0.04263898730278015, -0.0087185213342309, -0.22732441127300262, -0.18440759181976318, 0.08407571166753769, 0.0475882925093174, -0.27695369720458984, -0.11150874197483063, -0.08633091300725937, -0.025948671624064445, 0.015797629952430725, 0.024629639461636543, 0.12683221697807312, 0.009343303740024567, -0.002309986390173435, 0.023274589329957962, 0.04654689133167267, 0.004796623717993498, 0.1019316092133522, 0.049058932811021805, 0.038486547768116, -0.10318329930305481, -0.011453800834715366, 0.07649208605289459, 0.006040405947715044, 0.08937165886163712, -0.11285331100225449, -0.04790160804986954, -0.13555017113685608, 0.0030210153199732304, -0.06859123706817627, 0.07264534384012222, -0.12519951164722443, 0.008852554485201836, -0.10105250030755997, 0.11457408964633942, 0.014208126813173294, 0.011111644096672535, -0.020695151761174202, -0.06153029203414917, 0.038356903940439224, -0.04703693836927414, 0.10515255481004715, 0.09313026815652847, -0.1393871009349823, -0.018572237342596054, -0.01609625108540058, -0.0009770154720172286, 0.007570186164230108, 0.0633152648806572, 0.061385199427604675, -0.03813629224896431, 0.1513599306344986, -0.099500872194767, -0.18335369229316711, -0.021122673526406288, 0.110544353723526, -0.04584207385778427, -0.2009223848581314, -0.05477036163210869, -0.004807735327631235, 0.012100768275558949, -0.09834323078393936, 0.0826243907213211, -0.012992707081139088, -0.09753704816102982, 0.0018637881148606539, 0.07001478224992752, -0.014079303480684757, 0.06644772738218307, -0.009290735237300396, 0.04342668876051903, -0.10130993276834488, 0.08997814357280731, 0.08389127254486084, -0.0767979770898819, 0.02844054438173771, 0.2069302499294281, 0.01771482266485691, -0.04592733457684517, 0.12929461896419525, 0.24390481412410736, -0.059869784861803055, -0.010995110496878624, -0.027938904240727425, -0.09803951531648636, 0.09282555431127548, 0.06429985165596008, 0.019815998151898384, 0.0574236661195755, 0.0749199315905571, 0.021573882550001144, 0.06766432523727417, 0.12577544152736664, 0.030478591099381447, -0.0003117050800938159, -0.035882022231817245, 0.03952745720744133, -0.01452045701444149, -0.09016578644514084, 0.021878164261579514, 0.00849526934325695, -0.08850488066673279, -0.040643587708473206, -0.15834154188632965, -0.01841716282069683, -0.043883971869945526, -0.012525836005806923, -0.01898684911429882, -0.03853902593255043, 0.0905451625585556, 0.035125233232975006, -0.07211368530988693, -0.09269918501377106, -0.008918259292840958, 0.1380203813314438, -0.1968160718679428, 0.009267231449484825, 0.10146432369947433, -0.012246890924870968, 0.1258755326271057, 0.039614319801330566, 0.01289881207048893, -0.0340522825717926, -0.0980997085571289, 0.050190269947052, -0.0899316668510437, -0.025222886353731155, 0.00862310454249382, -0.1208031103014946, -0.07419589161872864, -0.030499547719955444, -0.0024600301403552294, 0.04513942077755928, -0.05118975415825844, -0.03594475984573364, 0.07192066311836243, 0.07652387768030167, -0.10195976495742798, -0.06513861566781998, 0.04785334691405296, 0.06334991008043289, -0.06251712143421173, 0.06531583517789841, -0.05858098715543747, 0.01839282177388668, -0.06558738648891449, -0.003239013720303774, 0.03471415117383003, 0.020800834521651268, -0.17851078510284424, 0.006721203215420246, -0.007460448890924454, 0.03558044880628586, 0.10044752061367035, -0.03352166339755058, -0.051678355783224106, 0.07127762585878372, 0.047571055591106415, -0.021002499386668205, 0.08379770815372467, -0.07276566326618195, -0.07191725075244904, 0.021547580137848854, -0.07063799351453781, -0.03529863804578781, -0.07047050446271896, -0.042438313364982605, 0.10748879611492157, 0.16228990256786346, 0.18669337034225464, -0.029240820556879044, 0.07123954594135284, -0.021803908050060272, -0.16372893750667572, 0.14628496766090393, -0.05453816056251526, -0.014933998696506023, -0.07689472287893295, 0.09280971437692642, 0.09993214160203934, -0.29080331325531006, 0.13607154786586761, 0.012591755017638206, -0.03551896661520004, 0.07083237916231155, -0.15258461236953735, -0.07913132756948471, -0.09508902579545975, -0.0025171046145260334, -0.07532601058483124, 0.09710836410522461, 0.02790040709078312, 0.031174376606941223, 0.018982525914907455, 0.16817936301231384, -0.07094953209161758, -0.055414386093616486, 0.12040704488754272, 0.0068510752171278, 0.012333095073699951, 0.05826954543590546, 0.026902616024017334, -0.011782943271100521, 0.043305348604917526, 0.06932590156793594, 0.06827067583799362, -0.051384877413511276, -0.04266788065433502, -0.1039893627166748, -0.09338735044002533, 0.07379839569330215, -0.04213700443506241, -0.0435253344476223, 0.12670478224754333, 0.06864169985055923, -0.01606222242116928, -0.022170528769493103, 0.19454927742481232, -0.011492184363305569, -0.12588346004486084, -0.09504342079162598, 0.082832932472229, -0.002411193447187543, 0.034011367708444595, -0.00987517274916172, -0.08596531301736832, 0.06684978306293488, 0.08002660423517227, 0.08573486655950546, 0.008981854654848576, 0.01907266117632389, -0.017819778993725777, -0.013306774199008942, 0.014718354679644108, 0.0018320108065381646, -0.01574636809527874, 0.24467863142490387, -0.1060577929019928, 0.03458253666758537, 0.007707701064646244, -0.05775982886552811, -0.13660137355327606, 0.12546898424625397, -0.04256879538297653, 0.017450757324695587, -0.06812452524900436, 0.16992901265621185, -0.11719515919685364, -0.22173087298870087, -0.02687882073223591, -0.09056860208511353, -0.10745614022016525, -0.057855792343616486, 0.12923356890678406, 0.0778256356716156, 0.03452926501631737, 0.10687842220067978, -0.02932327426970005, 0.03887943923473358, 0.018469780683517456, -0.03906865790486336, -0.013184169307351112, 0.07311346381902695, -0.20221419632434845, 0.2588109076023102, 0.020646952092647552, 0.013393422588706017, 0.10671738535165787, -0.056229978799819946, -0.09497269243001938, -0.07516475021839142, 0.0018080385634675622, -0.10178812593221664, 0.027177536860108376, 0.24516788125038147, -0.0063541121780872345, 0.07585813105106354, 0.09312613308429718, -0.016717925667762756, 0.10934408754110336, 0.09796208143234253, -0.02062062919139862, -0.016193192452192307, 0.13948293030261993, -0.09966161102056503, 0.11214610934257507, 0.1337304562330246, -0.019522549584507942, 0.05526024475693703, -0.05438946187496185, -0.004134560935199261, -0.0233207605779171, 0.16426676511764526, 0.010631129145622253, -0.1432396024465561, 0.057576123625040054, -0.012386759743094444, 0.048585809767246246, -0.14036038517951965, -0.08613527566194534, -0.01969948224723339, 0.06122497096657753, -0.0612492561340332, 0.1690998524427414, 0.14906145632266998, -0.01610252447426319, 0.01490053627640009, -0.13134106993675232, 0.005272715352475643, 0.12854818999767303, -0.01651088520884514, 0.0421549528837204 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [hf-test/xls-r-dummy](https://huggingface.co/hf-test/xls-r-dummy) on the COMMON_VOICE - AB dataset. It achieves the following results on the evaluation set: - Loss: 133.5167 - Wer: 18.9286 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 2.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.11.0
{"language": ["ab"], "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "", "results": []}]}
automatic-speech-recognition
baaastien/xls-r-ab-test
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "common_voice", "generated_from_trainer", "ab", "dataset:common_voice", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ab" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #ab #dataset-common_voice #endpoints_compatible #region-us
# This model is a fine-tuned version of hf-test/xls-r-dummy on the COMMON_VOICE - AB dataset. It achieves the following results on the evaluation set: - Loss: 133.5167 - Wer: 18.9286 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 2.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.11.0
[ "# \n\nThis model is a fine-tuned version of hf-test/xls-r-dummy on the COMMON_VOICE - AB dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 133.5167\n- Wer: 18.9286", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 2.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1+cu102\n- Datasets 1.17.1.dev0\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #ab #dataset-common_voice #endpoints_compatible #region-us \n", "# \n\nThis model is a fine-tuned version of hf-test/xls-r-dummy on the COMMON_VOICE - AB dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 133.5167\n- Wer: 18.9286", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 2.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1+cu102\n- Datasets 1.17.1.dev0\n- Tokenizers 0.11.0" ]
[ 61, 59, 6, 12, 8, 3, 140, 4, 41 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #ab #dataset-common_voice #endpoints_compatible #region-us \n# \n\nThis model is a fine-tuned version of hf-test/xls-r-dummy on the COMMON_VOICE - AB dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 133.5167\n- Wer: 18.9286## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 2.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1+cu102\n- Datasets 1.17.1.dev0\n- Tokenizers 0.11.0" ]
[ -0.0865170955657959, 0.1464160829782486, -0.0035482191015034914, 0.020073749125003815, 0.1262628734111786, 0.03187305107712746, 0.06765814870595932, 0.14301787316799164, -0.06566157191991806, 0.10688281804323196, 0.05168229341506958, -0.006196088157594204, 0.08671777695417404, 0.06779796630144119, 0.012509878724813461, -0.2680591344833374, 0.006310082972049713, -0.021381428465247154, -0.057771217077970505, 0.07868572324514389, 0.1128728985786438, -0.08253269642591476, 0.014997917227447033, 0.022388679906725883, -0.10981163382530212, 0.029735377058386803, -0.060549765825271606, -0.04700643941760063, 0.08475647121667862, 0.021587632596492767, 0.0461520180106163, 0.01983875036239624, 0.09625318646430969, -0.3003072440624237, 0.0011436209315434098, 0.0752488523721695, 0.054204463958740234, 0.07207559794187546, 0.05753417685627937, 0.012106821872293949, 0.15071320533752441, -0.15327666699886322, 0.08043478429317474, 0.0696341022849083, -0.04274512454867363, -0.18767766654491425, -0.07587018609046936, 0.055885665118694305, 0.09858714789152145, 0.11435750126838684, -0.028266631066799164, 0.1238635703921318, -0.09498210996389389, 0.05515396222472191, 0.17789411544799805, -0.23085102438926697, -0.03226229548454285, -0.010468780063092709, 0.05946333706378937, 0.061590526252985, -0.10913368314504623, -0.0040224166586995125, 0.044322118163108826, 0.016472790390253067, 0.04340915381908417, 0.02267039753496647, -0.007224193308502436, -0.0013959555653855205, -0.11073993891477585, -0.03811667114496231, 0.1651875078678131, 0.09203513711690903, -0.04296953231096268, -0.1471332311630249, 0.006531805265694857, -0.15775328874588013, -0.020378388464450836, -0.020310821011662483, 0.015374848619103432, -0.046392157673835754, -0.06773797422647476, -0.012310684658586979, -0.07224691659212112, -0.06902046501636505, 0.06019808351993561, 0.1049044206738472, 0.044687602669000626, -0.029869256541132927, 0.024003440514206886, 0.0836365595459938, 0.020756322890520096, -0.11963790655136108, -0.03636451065540314, 0.0186153631657362, -0.14223317801952362, -0.03113933466374874, -0.04588833078742027, -0.06659234315156937, 0.02569439634680748, 0.09149022400379181, -0.024940678849816322, 0.0870533362030983, 0.009112470783293247, -0.007615116890519857, 0.024469347670674324, 0.13923287391662598, -0.03828741982579231, -0.06593247503042221, -0.03604000061750412, 0.09950895607471466, 0.007909994572401047, -0.02338656224310398, -0.06410471349954605, -0.016433339565992355, 0.11634708195924759, 0.08268473297357559, -0.035529207438230515, -0.01679185964167118, -0.06110373139381409, -0.02117595449090004, 0.005001616198569536, -0.1314934641122818, 0.05243711173534393, 0.018848631531000137, -0.03773009404540062, 0.02233484573662281, 0.002508101286366582, 0.009220690466463566, -0.06631128489971161, 0.08408951014280319, -0.055397067219018936, -0.027917833998799324, -0.04822726920247078, -0.059010136872529984, 0.03208311274647713, -0.03086562640964985, 0.001008143532089889, -0.05512825399637222, -0.08507051318883896, -0.05826260894536972, 0.030422745272517204, -0.07688718289136887, -0.07795204222202301, -0.03872251510620117, -0.01949296146631241, 0.04021294787526131, -0.017278043553233147, 0.11887712776660919, -0.043095678091049194, 0.07048710435628891, 0.01308386866003275, 0.008988233283162117, 0.0692664384841919, 0.05947981774806976, -0.041745323687791824, 0.03917216509580612, -0.05451730638742447, 0.11069565266370773, -0.11204083263874054, 0.011950121261179447, -0.13717040419578552, -0.11501109600067139, -0.04745849221944809, -0.015654554590582848, 0.07987061887979507, 0.10981444269418716, -0.15886084735393524, -0.06412947177886963, 0.13903819024562836, -0.060321830213069916, -0.06885766237974167, 0.14019888639450073, -0.014898152090609074, -0.02801167033612728, 0.0517907589673996, 0.13863512873649597, 0.11119776219129562, -0.1325884461402893, -0.04572343826293945, -0.005802667699754238, 0.08941663056612015, 0.048237692564725876, 0.08149243891239166, -0.04607449471950531, 0.005586548708379269, 0.002164937788620591, -0.010329500772058964, 0.02390761859714985, -0.06081511825323105, -0.07341484725475311, -0.030624769628047943, -0.07555285841226578, 0.009293505921959877, 0.035086844116449356, 0.0038598808459937572, -0.07969211786985397, -0.16036152839660645, 0.06412258744239807, 0.13975954055786133, -0.07522298395633698, 0.01958637498319149, -0.077760249376297, -0.0015179204056039453, -0.026724783703684807, -0.013552279211580753, -0.16643126308918, -0.03488839790225029, 0.0573163665831089, -0.10344111919403076, 0.058808837085962296, 0.014485102146863937, 0.05048484727740288, 0.036587655544281006, -0.03708188980817795, -0.032157592475414276, -0.09905760735273361, 0.012627092190086842, -0.061880409717559814, -0.17159350216388702, -0.0640345886349678, -0.02171727456152439, 0.23446272313594818, -0.22631791234016418, -0.017859550192952156, 0.01869121752679348, 0.12987835705280304, 0.006157025694847107, -0.072512187063694, 0.015437756665050983, 0.048395927995443344, -0.0033329627476632595, -0.07650842517614365, 0.010055072605609894, 0.007943726144731045, -0.1382085382938385, -0.03102150373160839, -0.15032018721103668, -0.043238092213869095, 0.06911665201187134, 0.10023649036884308, -0.07466430217027664, -0.07167881727218628, -0.06360204517841339, -0.054348424077034, -0.06043020635843277, -0.04327791929244995, 0.19863763451576233, 0.04228082299232483, 0.09455578029155731, -0.04603259265422821, -0.06375814229249954, 0.017499035224318504, 0.03675113618373871, -0.038473110646009445, 0.08014905452728271, 0.04892465099692345, -0.0969267413020134, 0.04373086243867874, 0.028016841039061546, -0.0421534962952137, 0.14717595279216766, -0.05538291856646538, -0.12901239097118378, -0.017656957730650902, 0.011600444093346596, 0.011005697771906853, 0.1086932122707367, -0.15770889818668365, 0.00751466816291213, 0.04048335552215576, 0.009328882209956646, 0.03765904903411865, -0.1367560476064682, 0.02067256160080433, 0.051066506654024124, -0.02538708969950676, -0.051178935915231705, -0.027604106813669205, 0.009861239232122898, 0.06056874990463257, 0.029749689623713493, 0.0062653422355651855, -0.015471618622541428, -0.03886021301150322, -0.0958109125494957, 0.14243818819522858, -0.10610882937908173, -0.17749443650245667, -0.13355161249637604, 0.036597367376089096, -0.015132400207221508, -0.038573604077100754, 0.02903469279408455, -0.11837437748908997, -0.06813369691371918, -0.08759534358978271, -0.01106277946382761, -0.08702415227890015, 0.0063042547553777695, 0.06082996726036072, 0.022498490288853645, 0.09404686093330383, -0.13099318742752075, 0.021948914974927902, -0.001857275259681046, -0.034610144793987274, -0.029122579842805862, 0.036203280091285706, 0.09179038554430008, 0.14658550918102264, 0.012795490212738514, 0.02840939536690712, -0.02516065537929535, 0.19028231501579285, -0.11702343076467514, -0.016470691189169884, 0.08850046992301941, 0.012644624337553978, 0.03762149438261986, 0.10234557092189789, 0.030477270483970642, -0.06762973219156265, 0.031693585216999054, 0.06240207329392433, -0.011906234547495842, -0.26201969385147095, -0.059785205870866776, -0.05796315521001816, -0.09707582741975784, 0.10948732495307922, 0.06368371844291687, 0.006936939898878336, 0.013955590315163136, -0.04038712754845619, 0.02541028894484043, -0.003754932200536132, 0.07451573014259338, 0.08054585754871368, 0.03736473619937897, 0.07391215860843658, -0.02282692678272724, -0.02709372714161873, 0.04484027996659279, 0.004648382309824228, 0.22146075963974, -0.0034079598262906075, 0.17158886790275574, 0.014748062938451767, 0.10596919804811478, -0.010733472183346748, 0.029212001711130142, 0.018300168216228485, -0.005879069212824106, 0.032697468996047974, -0.0704614520072937, -0.04025967791676521, 0.04447562247514725, 0.11289390921592712, 0.0180963221937418, -0.08496823161840439, 0.028141286224126816, -0.0023182653822004795, 0.2868930995464325, 0.08397507667541504, -0.24283689260482788, -0.07904798537492752, 0.01223260723054409, -0.05370238050818443, -0.08088359236717224, 0.011041530407965183, 0.09655436873435974, -0.12182606756687164, 0.07550916075706482, -0.034176602959632874, 0.09485645592212677, -0.060959357768297195, 0.0027553571853786707, 0.026224307715892792, 0.10454749315977097, -0.005929470062255859, 0.10019275546073914, -0.1655469387769699, 0.17618834972381592, 0.010814487002789974, 0.10690224170684814, -0.06876230239868164, 0.05074257031083107, -0.004430577624589205, -0.0023692890536040068, 0.10121245682239532, 0.009019356220960617, -0.0567643940448761, -0.1628275066614151, -0.08929813653230667, 0.03719479590654373, 0.12979373335838318, -0.024239864200353622, 0.08232541382312775, -0.0500512570142746, -0.01163694728165865, 0.044495876878499985, -0.07305573672056198, -0.1823044866323471, -0.18489034473896027, 0.045872464776039124, 0.04447732865810394, 0.05444697290658951, -0.07794392853975296, -0.08730914443731308, -0.03289680927991867, 0.21934552490711212, -0.038932379335165024, -0.03190893679857254, -0.1379234790802002, 0.1046777218580246, 0.16505686938762665, -0.0428818054497242, 0.010112067684531212, 0.03308798000216484, 0.17000843584537506, 0.0160294771194458, -0.0381520576775074, 0.051959697157144547, -0.05598039925098419, -0.1379322111606598, -0.04823313280940056, 0.19145797193050385, 0.04309818148612976, 0.08425111323595047, -0.007208545692265034, 0.005460662767291069, 0.012229685671627522, -0.07352246344089508, 0.0552668496966362, 0.07431231439113617, -0.005922123324126005, 0.08964499086141586, -0.0494723804295063, 0.026766078546643257, -0.06661783903837204, -0.04659663140773773, 0.16452926397323608, 0.23991496860980988, -0.0709640234708786, 0.09441894292831421, 0.0735073983669281, -0.05597551167011261, -0.13427016139030457, 0.013056247495114803, 0.13581404089927673, 0.05503693222999573, 0.04423220083117485, -0.1993996948003769, 0.07787343114614487, 0.10861413180828094, -0.015715699642896652, 0.0306396447122097, -0.28391799330711365, -0.11294584721326828, 0.08991329371929169, 0.04482429474592209, -0.046972569078207016, -0.1217045709490776, -0.061639755964279175, -0.07569384574890137, -0.11661578714847565, 0.05282992124557495, -0.009301828220486641, 0.11567942053079605, 0.0070760250091552734, 0.058930788189172745, 0.05309964716434479, -0.03662601858377457, 0.14445911347866058, 0.01289052702486515, 0.01767639070749283, -0.010938504710793495, 0.06522928923368454, 0.02706676721572876, -0.061061177402734756, 0.06956624984741211, -0.08815037459135056, 0.040529076009988785, -0.14092056453227997, -0.037850361317396164, -0.05119550600647926, 0.04365752264857292, -0.043317683041095734, -0.04633752256631851, -0.03299359977245331, 0.04681701585650444, 0.08440390974283218, -0.01856767013669014, 0.0606500618159771, -0.02609330601990223, 0.0898304432630539, 0.08715508878231049, 0.12125474214553833, -0.00603037653490901, -0.12370113283395767, -0.004905471578240395, -0.016726592555642128, 0.0566522553563118, -0.08734682202339172, 0.03858613595366478, 0.11380214989185333, 0.06316151469945908, 0.15657736361026764, 0.012728429399430752, -0.11287005990743637, 0.00011250803072471172, 0.04495970532298088, -0.025864655151963234, -0.19649502635002136, -0.04415076971054077, 0.05400138348340988, -0.16073206067085266, -0.016955656930804253, 0.08935683220624924, -0.050755053758621216, -0.017695952206850052, 0.004578088875859976, 0.02564341388642788, -0.030571771785616875, 0.17911165952682495, 0.006536177825182676, 0.0961395725607872, -0.07301553338766098, 0.09453712403774261, 0.09445162117481232, -0.1211327612400055, 0.08274047821760178, 0.05494675412774086, -0.044739268720149994, -0.011111358180642128, 0.019795741885900497, 0.036567993462085724, 0.032954223453998566, -0.05180931091308594, -0.048864152282476425, -0.11688805371522903, 0.043370988219976425, -0.0326521098613739, 0.00018880573043134063, -0.0069371359422802925, -0.032917510718107224, 0.034388624131679535, -0.1723920702934265, 0.08577467501163483, 0.03868231177330017, 0.06173430383205414, -0.13494300842285156, 0.06206780672073364, 0.02221676893532276, 0.024623414501547813, 0.00085542316082865, -0.026527930051088333, -0.05569557473063469, -0.0015154939610511065, -0.124672532081604, -0.03175598382949829, -0.048526398837566376, 0.00035998818930238485, 0.0023523815907537937, -0.031021812930703163, -0.03883977606892586, 0.060843806713819504, -0.06620874255895615, -0.09115497022867203, -0.008082560263574123, 0.07420837879180908, -0.10334512591362, 0.018419062718749046, 0.03879450261592865, -0.12577810883522034, 0.08371370285749435, 0.05646321177482605, 0.024078745394945145, 0.029109898954629898, -0.08492822200059891, -0.02887718565762043, 0.045542921870946884, 0.030413219705224037, 0.04354022443294525, -0.13778218626976013, -0.01629750244319439, -0.01218046061694622, 0.01486461516469717, -0.0002899124228861183, -0.0044682323932647705, -0.11097657680511475, -0.056194551289081573, -0.08743473887443542, -0.0388140045106411, -0.06162700802087784, 0.0476190447807312, 0.0757892057299614, 0.03835299611091614, 0.14396899938583374, -0.07306774705648422, 0.06364854425191879, -0.20007571578025818, -0.01915602572262287, -0.024199562147259712, 0.01947626657783985, -0.04718134179711342, -0.01438802108168602, 0.07904193550348282, -0.04902055859565735, 0.11753059923648834, -0.061862293630838394, 0.09349334985017776, 0.04627872258424759, -0.05338970571756363, -0.015653472393751144, -0.007433158345520496, 0.21993467211723328, 0.09223348647356033, -0.023740282282233238, 0.08144372701644897, -0.0643107071518898, 0.05231210216879845, 0.11182574927806854, 0.10640526562929153, 0.1525641530752182, 0.02450110763311386, 0.05477128177881241, 0.07874146103858948, -0.12655635178089142, -0.13650411367416382, 0.1489325761795044, -0.026534080505371094, 0.12696291506290436, -0.020680366083979607, 0.20147453248500824, 0.1204296350479126, -0.1604853868484497, 0.05676645785570145, -0.05545121431350708, -0.11435717344284058, -0.09694512188434601, -0.08157793432474136, -0.07719399780035019, -0.139052614569664, 0.03096211887896061, -0.09457283467054367, 0.03748064860701561, 0.03734496235847473, 0.02548123337328434, 0.019155241549015045, 0.1247360110282898, -0.005301198456436396, -0.010274605825543404, 0.09006187319755554, -0.003827861277386546, -0.016883496195077896, -0.023031508550047874, -0.039898891001939774, 0.0986841693520546, 0.006456989794969559, 0.11026953905820847, -0.028547760099172592, -0.03611245006322861, 0.06212848424911499, 0.0064997123554348946, -0.09467612951993942, 0.028709879145026207, -0.009784284979104996, 0.030910050496459007, 0.08065415173768997, 0.06361491978168488, -0.008782659657299519, -0.05769185721874237, 0.23539946973323822, -0.07047727704048157, -0.056737951934337616, -0.13219931721687317, 0.18429958820343018, 0.048446059226989746, -0.0016631721518933773, 0.052222445607185364, -0.13197152316570282, -0.01493233535438776, 0.093991719186306, 0.12956318259239197, -0.006148006301373243, -0.01771606132388115, -0.01052832417190075, -0.017986761406064034, -0.0826605036854744, 0.08627313375473022, 0.09740083664655685, 0.03901202976703644, -0.01775648258626461, 0.055512573570013046, -0.019697602838277817, -0.07160482555627823, -0.06009146198630333, 0.07730019837617874, 0.034056808799505234, -0.007512183394283056, -0.009425322525203228, 0.10879503190517426, -0.007217972073704004, -0.20212842524051666, -0.00037401009467430413, -0.13120363652706146, -0.18795451521873474, -0.023632287979125977, 0.06324834376573563, 0.011113343760371208, 0.04535925015807152, -0.015404729172587395, -0.013473023660480976, 0.145236536860466, 0.0004572852631099522, -0.04730776697397232, -0.07930770516395569, 0.0757928192615509, -0.024515196681022644, 0.18271097540855408, 0.0146954245865345, 0.06453308463096619, 0.09343447536230087, 0.05054669827222824, -0.12813366949558258, 0.04883148521184921, 0.0790862888097763, -0.11689548939466476, 0.06666269898414612, 0.22999131679534912, -0.041140105575323105, 0.1250610053539276, 0.05565589293837547, -0.1154399961233139, -0.02212504670023918, -0.09469262510538101, -0.0057572233490645885, -0.08354884386062622, 0.010845978744328022, -0.03849015757441521, 0.1560291051864624, 0.18211165070533752, -0.059420108795166016, -0.028609048575162888, -0.07008180767297745, 0.0032857628539204597, 0.047045521438121796, 0.1537897139787674, -0.023381829261779785, -0.20304253697395325, 0.02661617286503315, -0.010427114553749561, 0.04319643974304199, -0.2536868453025818, -0.08544424176216125, 0.04489210993051529, -0.055198658257722855, -0.026927579194307327, 0.1199464350938797, 0.06685864925384521, 0.01688828505575657, -0.04727894812822342, -0.15198694169521332, -0.01691005751490593, 0.14950047433376312, -0.18032227456569672, -0.03333010897040367 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-timit_asr-oogway This model is a fine-tuned version of [OthmaneJ/distil-wav2vec2](https://huggingface.co/OthmaneJ/distil-wav2vec2) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-timit_asr-oogway", "results": []}]}
automatic-speech-recognition
baby-oogway/wav2vec2-timit_asr-oogway
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
# wav2vec2-timit_asr-oogway This model is a fine-tuned version of OthmaneJ/distil-wav2vec2 on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# wav2vec2-timit_asr-oogway\n\nThis model is a fine-tuned version of OthmaneJ/distil-wav2vec2 on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "# wav2vec2-timit_asr-oogway\n\nThis model is a fine-tuned version of OthmaneJ/distil-wav2vec2 on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 56, 44, 6, 12, 8, 3, 117, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec2-timit_asr-oogway\n\nThis model is a fine-tuned version of OthmaneJ/distil-wav2vec2 on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.09546970576047897, 0.11961981654167175, -0.0022739695850759745, 0.06511349976062775, 0.12947486340999603, 0.017313964664936066, 0.11261030286550522, 0.13794180750846863, -0.07867956161499023, 0.06558762490749359, 0.061427515000104904, 0.06523459404706955, 0.0788135975599289, 0.1355576515197754, -0.03514258563518524, -0.24267138540744781, 0.009523695334792137, -0.032237812876701355, -0.08338982611894608, 0.09703126549720764, 0.08276646584272385, -0.08842545002698898, 0.06986650079488754, 0.008412940427660942, -0.12042474746704102, 0.00783108826726675, -0.05607171729207039, -0.08304157108068466, 0.09256221354007721, -0.014545390382409096, 0.058369915932416916, 0.026978325098752975, 0.1457679122686386, -0.2331400066614151, -0.0007760290754958987, 0.08884229511022568, 0.037583108991384506, 0.08735129237174988, 0.06468753516674042, -0.01430189236998558, 0.07815956324338913, -0.1690283864736557, 0.09215854853391647, 0.02864832989871502, -0.06449207663536072, -0.14291074872016907, -0.07299298048019409, 0.05602394416928291, 0.09435462206602097, 0.10294242203235626, 0.012352444231510162, 0.13607539236545563, -0.06651662290096283, 0.07947733998298645, 0.22607707977294922, -0.23953472077846527, -0.048859238624572754, 0.021193664520978928, 0.05763652175664902, 0.06326045095920563, -0.11415995657444, -0.01724398322403431, 0.011687574908137321, 0.028422381728887558, 0.09732156246900558, -0.022862063720822334, -0.08920718729496002, -0.000025980933060054667, -0.12271934747695923, 0.003380472306162119, 0.10197418928146362, 0.03563397750258446, -0.047818198800086975, -0.11284709721803665, -0.0538056418299675, -0.1074037104845047, -0.015497803688049316, -0.030565375462174416, 0.055983953177928925, -0.05460871011018753, -0.06233606114983559, -0.038836769759655, -0.06837515532970428, -0.055416401475667953, -0.013780253008008003, 0.15298962593078613, 0.02589331567287445, 0.016723070293664932, -0.02825755812227726, 0.110588438808918, 0.02825445495545864, -0.11565607786178589, -0.02016051486134529, 0.007024346850812435, -0.11411788314580917, -0.041117653250694275, -0.02972392737865448, -0.007711159531027079, 0.012981126084923744, 0.12866990268230438, -0.02596789412200451, 0.08535832166671753, 0.007207712158560753, -0.010060618631541729, -0.00931733287870884, 0.1362956315279007, -0.051613178104162216, -0.058764487504959106, -0.016303136944770813, 0.08516137301921844, -0.0008550635538995266, -0.021979358047246933, -0.062515988945961, -0.020633887499570847, 0.05468273535370827, 0.06294883787631989, -0.008901599794626236, 0.024288740009069443, -0.037138424813747406, -0.0356205515563488, 0.008846351876854897, -0.14140892028808594, 0.049084894359111786, 0.0076053207740187645, -0.09057622402906418, 0.0194214079529047, 0.02033858560025692, 0.02693377062678337, -0.05784500017762184, 0.10270289331674576, -0.062029920518398285, 0.007299421355128288, -0.0505034402012825, -0.05132352188229561, 0.019161446020007133, -0.1213846281170845, -0.0075637297704815865, -0.05964967980980873, -0.1630561500787735, -0.0477469377219677, 0.06149512156844139, -0.06001598387956619, -0.019993280991911888, -0.060495831072330475, -0.0643618032336235, 0.02928791753947735, -0.019033188000321388, 0.18471944332122803, -0.06149118021130562, 0.061386771500110626, -0.03570994362235069, 0.03713294491171837, 0.025544270873069763, 0.05185868963599205, -0.08863452076911926, 0.00970344990491867, -0.06356318295001984, 0.07374914735555649, -0.07157338410615921, 0.014104649424552917, -0.13625894486904144, -0.09333693236112595, 0.0007865997613407671, -0.019601846113801003, 0.053230170160532, 0.12816786766052246, -0.18733344972133636, -0.02760365419089794, 0.12652812898159027, -0.050091106444597244, -0.03384948894381523, 0.0695110484957695, -0.02205956168472767, 0.020606135949492455, 0.041224632412195206, 0.1680416464805603, 0.055158961564302444, -0.16517780721187592, -0.011710614897310734, -0.005398394539952278, 0.04310208931565285, 0.008382300846278667, 0.0694471076130867, -0.018408309668302536, 0.07335913926362991, 0.00033469891059212387, -0.059200581163167953, -0.026570357382297516, -0.06477602571249008, -0.08607957512140274, -0.047819484025239944, -0.07508629560470581, -0.012619161047041416, 0.009742426685988903, 0.018680527806282043, -0.05738576129078865, -0.11695591360330582, 0.05622629448771477, 0.1333804726600647, -0.07754223048686981, 0.03371942788362503, -0.0804932489991188, -0.027466341853141785, -0.013475662097334862, -0.02392023243010044, -0.18189847469329834, -0.06839750707149506, 0.02535214088857174, -0.07644379138946533, 0.026591496542096138, 0.006098973099142313, 0.07584863156080246, 0.055418919771909714, -0.03277386352419853, -0.03038685768842697, -0.058063387870788574, 0.012138078920543194, -0.074567049741745, -0.2118392139673233, -0.05755177512764931, -0.03773553669452667, 0.143183633685112, -0.1807483434677124, -0.0013833768898621202, -0.004337314050644636, 0.13711673021316528, 0.0370599739253521, -0.06392531841993332, 0.011477903462946415, 0.046868812292814255, 0.011487355455756187, -0.11847124248743057, 0.032648660242557526, -0.0025017866864800453, -0.08282030373811722, -0.018734104931354523, -0.15795569121837616, 0.0164082869887352, 0.0979706197977066, 0.09719304740428925, -0.06749393045902252, -0.017587026581168175, -0.0705576166510582, -0.06472338736057281, -0.07591678947210312, 0.03766524791717529, 0.2047550231218338, 0.029053322970867157, 0.09137807786464691, -0.04827038571238518, -0.07216645777225494, 0.019407475367188454, 0.013400143012404442, -0.014968684874475002, 0.08817557245492935, 0.07449159771203995, -0.08150465786457062, 0.07374506443738937, 0.08476913720369339, -0.03269154578447342, 0.1298319548368454, -0.033525675535202026, -0.09562397003173828, -0.01356128603219986, -0.018252529203891754, -0.013482271693646908, 0.11423072218894958, -0.07200299203395844, 0.01407382171601057, 0.0358477458357811, 0.03127656131982803, 0.024935631081461906, -0.16903585195541382, 0.008013444021344185, 0.013610697351396084, -0.050580039620399475, -0.0566994771361351, -0.016959741711616516, 0.01642800308763981, 0.0721701830625534, 0.009160131216049194, -0.02698422782123089, 0.02764524705708027, -0.013738281093537807, -0.08418303728103638, 0.15860101580619812, -0.120407335460186, -0.15267355740070343, -0.09652596712112427, 0.013362269848585129, -0.034456703811883926, -0.03665965050458908, 0.013810128904879093, -0.08562751859426498, -0.06970825791358948, -0.09035258740186691, -0.0019410691456869245, -0.028197482228279114, 0.016189048066735268, 0.0677480697631836, 0.025464897975325584, 0.0603756420314312, -0.11972203850746155, -0.0007516317418776453, -0.026620939373970032, -0.08336707204580307, 0.010793742723762989, 0.07816461473703384, 0.06977494806051254, 0.09910999238491058, 0.0040206522680819035, 0.018074164167046547, -0.020191220566630363, 0.17188547551631927, -0.07950358837842941, 0.018331551924347878, 0.10497523844242096, -0.005584171041846275, 0.04658736661076546, 0.1078733578324318, 0.02927052415907383, -0.08067205548286438, 0.02245405875146389, 0.060124434530735016, -0.018191037699580193, -0.2607972323894501, -0.03025885485112667, -0.024367917329072952, -0.0626496747136116, 0.11183713376522064, 0.05324631556868553, -0.011853392235934734, 0.031086908653378487, 0.005994458682835102, 0.010706568136811256, -0.001495218719355762, 0.06500156223773956, 0.049003154039382935, 0.0601377934217453, 0.08527180552482605, -0.013199453242123127, -0.011509555391967297, 0.0663631409406662, 0.04437082260847092, 0.25388970971107483, -0.010759585537016392, 0.10313767939805984, 0.011172126978635788, 0.12545891106128693, -0.0288954209536314, 0.06620211154222488, 0.035610880702733994, 0.0028609170112758875, 0.00566626712679863, -0.06942250579595566, -0.01163844857364893, 0.04838293790817261, 0.030094856396317482, 0.007959104143083096, -0.07053007185459137, 0.05851886793971062, 0.014120101928710938, 0.3170631527900696, 0.019074421375989914, -0.28456220030784607, -0.05110907927155495, -0.0020298021845519543, -0.0611809641122818, -0.06071667745709419, 0.009766031987965107, 0.11895042657852173, -0.13164539635181427, 0.07838920503854752, -0.06300890445709229, 0.07966020703315735, -0.048299696296453476, -0.0022154790349304676, 0.07799845933914185, 0.14029990136623383, -0.008761991746723652, 0.06148297339677811, -0.23717953264713287, 0.21117912232875824, 0.012238590978085995, 0.136715829372406, -0.08761961758136749, 0.02745254896581173, 0.015138084068894386, -0.00036410699249245226, 0.12158989906311035, -0.0035476679913699627, -0.05734040215611458, -0.12426534295082092, -0.13389551639556885, 0.0423237606883049, 0.13191144168376923, -0.026669608429074287, 0.07748707383871078, -0.039770569652318954, -0.01566438190639019, 0.04039851948618889, -0.09226454049348831, -0.14836692810058594, -0.11203032732009888, 0.04328475147485733, 0.05151103436946869, -0.006586481351405382, -0.05856088921427727, -0.10579998046159744, -0.03883088007569313, 0.1426243633031845, -0.02962370030581951, -0.03933518007397652, -0.15166476368904114, 0.0328606553375721, 0.17196513712406158, -0.04462319239974022, 0.014211256057024002, 0.036352306604385376, 0.10877444595098495, 0.03609967976808548, -0.10137883573770523, 0.06405651569366455, -0.09514953196048737, -0.178016796708107, -0.04665902256965637, 0.12967334687709808, 0.07005809247493744, 0.05299608036875725, -0.01993737556040287, 0.039019547402858734, -0.010653833858668804, -0.08041753619909286, 0.05700848251581192, 0.09726351499557495, 0.054135460406541824, 0.04079856723546982, -0.052985310554504395, -0.015583519823849201, -0.049225110560655594, -0.03689975664019585, 0.11364930868148804, 0.19732598960399628, -0.05979597568511963, 0.1417429894208908, 0.08219321072101593, -0.07857146859169006, -0.15376786887645721, 0.05339274927973747, 0.10928604006767273, 0.023762496188282967, 0.056403785943984985, -0.20218084752559662, 0.1211852878332138, 0.09598352015018463, -0.01283485721796751, 0.06407112628221512, -0.3065272867679596, -0.12893494963645935, 0.08565901964902878, 0.08179806172847748, 0.01904662698507309, -0.11019247025251389, -0.015589924529194832, -0.02455775812268257, -0.13874533772468567, 0.14552892744541168, -0.07803380489349365, 0.11707902699708939, -0.0063387928530573845, 0.11353559046983719, 0.021685181185603142, -0.048194900155067444, 0.12541091442108154, 0.05641739070415497, 0.051068905740976334, -0.048094864934682846, 0.04273080825805664, 0.027375780045986176, -0.05387892201542854, 0.03813784569501877, -0.05105455592274666, 0.044104453176259995, -0.10597484558820724, -0.02169337496161461, -0.06700017303228378, 0.06642366200685501, -0.03534410893917084, -0.06531363725662231, -0.0025205176789313555, 0.05789913982152939, 0.08307760208845139, -0.03639985993504524, 0.050275374203920364, 0.013478325679898262, 0.04538137465715408, 0.08705419301986694, 0.1115950345993042, -0.019861629232764244, -0.07659461349248886, -0.021466467529535294, -0.018129536882042885, 0.045840125530958176, -0.0812871903181076, 0.022296806797385216, 0.1303149312734604, 0.041131410747766495, 0.11735475808382034, 0.036348868161439896, -0.052363328635692596, -0.015391578897833824, 0.023393144831061363, -0.11069381237030029, -0.0909283384680748, 0.016381649300456047, -0.030760908499360085, -0.10521842539310455, 0.024087214842438698, 0.12283994257450104, -0.056300319731235504, -0.010129059664905071, -0.003212652401998639, 0.05336017161607742, -0.044131115078926086, 0.19979552924633026, 0.01877385377883911, 0.06146790459752083, -0.09848043322563171, 0.14322873950004578, 0.05624917522072792, -0.0838015228509903, 0.07672105729579926, 0.07752098143100739, -0.08183502405881882, -0.008639395236968994, 0.06734900921583176, 0.0829874575138092, -0.02000674232840538, -0.061762887984514236, -0.07071062922477722, -0.08851400762796402, 0.05604938417673111, 0.07708879560232162, 0.017043842002749443, -0.0002409628068562597, -0.04875896871089935, 0.01828789710998535, -0.14882303774356842, 0.05516231432557106, 0.054517731070518494, 0.05698220804333687, -0.12597768008708954, 0.11032828688621521, 0.008699242025613785, 0.030228467658162117, -0.01213080994784832, 0.014740408398211002, -0.08369199186563492, -0.005562345031648874, -0.07954510301351547, -0.01002169493585825, -0.046881888061761856, 0.004048951435834169, -0.008719777688384056, -0.06658370047807693, -0.042561307549476624, 0.052915070205926895, -0.07069078087806702, -0.055817291140556335, -0.005913981236517429, 0.03432806581258774, -0.13435804843902588, -0.00504727428779006, 0.017199721187353134, -0.10006507486104965, 0.07980889081954956, 0.08298946917057037, -0.016068648546934128, 0.023930074647068977, -0.13550090789794922, -0.02589641697704792, 0.02474496141076088, 0.018010668456554413, 0.06158231198787689, -0.11820408701896667, -0.02886802889406681, -0.001633239327929914, 0.03507651761174202, 0.008747118525207043, 0.09861978143453598, -0.11613291501998901, -0.028995594009757042, -0.08848825097084045, -0.0461161844432354, -0.061247896403074265, 0.030621634796261787, 0.11092887818813324, 0.03378331661224365, 0.1603870838880539, -0.05946599692106247, 0.05870961770415306, -0.16786035895347595, -0.024542201310396194, -0.006201205775141716, -0.025933587923645973, -0.04316576570272446, -0.011181238107383251, 0.09863509237766266, -0.0593012236058712, 0.11013689637184143, -0.010645986534655094, 0.07073977589607239, 0.03434068337082863, -0.08034279197454453, -0.08647100627422333, 0.015073582530021667, 0.1858755350112915, 0.07090131938457489, -0.00827026180922985, 0.0977444127202034, -0.02095203287899494, 0.05311087146401405, 0.0734020322561264, 0.20389747619628906, 0.15865521132946014, 0.025302408263087273, 0.09245795756578445, 0.06064619868993759, -0.1428687423467636, -0.15710537135601044, 0.14873343706130981, -0.05127396434545517, 0.1281111091375351, -0.06055382266640663, 0.17977787554264069, 0.06197421997785568, -0.17237012088298798, 0.03123854286968708, -0.043522730469703674, -0.09920839965343475, -0.1250796765089035, -0.05256348475813866, -0.07183492928743362, -0.11087770760059357, 0.02522418461740017, -0.09580272436141968, 0.07302743941545486, 0.07100315392017365, 0.03415204584598541, 0.039851125329732895, 0.13690556585788727, -0.04672256112098694, -0.003606608835980296, 0.0751674696803093, 0.0185953788459301, -0.013882430270314217, -0.04704311117529869, -0.08101386576890945, 0.03822442144155502, 0.02560245990753174, 0.0832434743642807, -0.05049216374754906, -0.017703162506222725, 0.055972374975681305, -0.024530593305826187, -0.06866154074668884, 0.021224284544587135, 0.0019718813709914684, 0.04032386466860771, 0.0511818453669548, 0.07547209411859512, -0.0025763516314327717, -0.027194103226065636, 0.27953338623046875, -0.0844014510512352, -0.1099850982427597, -0.1377863883972168, 0.153983473777771, 0.028469447046518326, -0.0365685373544693, 0.0689830482006073, -0.10413441061973572, -0.03602851554751396, 0.14987622201442719, 0.15395691990852356, -0.05671219527721405, -0.023738719522953033, -0.007799619343131781, -0.009207500144839287, -0.05231308937072754, 0.12125830352306366, 0.10383874922990799, 0.04946834594011307, -0.07481838017702103, -0.020025944337248802, -0.003445529378950596, -0.058289315551519394, -0.0691828727722168, 0.0623750239610672, -0.014378482475876808, -0.01187000423669815, -0.019422244280576706, 0.07197680324316025, -0.0193890780210495, -0.21803805232048035, 0.03653978928923607, -0.15462075173854828, -0.17618155479431152, -0.01986663229763508, 0.09866711497306824, 0.000670717447064817, 0.03815951943397522, 0.0038826290983706713, -0.015589010901749134, 0.143105149269104, -0.0021647412795573473, -0.0393524169921875, -0.07504203170537949, 0.07647959887981415, -0.0860094204545021, 0.17754144966602325, -0.009943021461367607, 0.08700747787952423, 0.07906778901815414, 0.05854751169681549, -0.13740496337413788, 0.026517009362578392, 0.05513058230280876, -0.10784868896007538, 0.03941696137189865, 0.17470356822013855, -0.05866086110472679, 0.08046973496675491, 0.03375301510095596, -0.08896560966968536, -0.018447766080498695, -0.033061884343624115, -0.013289210386574268, -0.05080537125468254, -0.013971362262964249, -0.055797066539525986, 0.15577608346939087, 0.17606548964977264, -0.034791164100170135, 0.01123035792261362, -0.08462785184383392, 0.03587685525417328, 0.02809806354343891, 0.051225483417510986, -0.03851718455553055, -0.22065238654613495, 0.01897255890071392, -0.0021775593049824238, 0.03885221481323242, -0.17628373205661774, -0.08558516949415207, 0.0263631921261549, -0.07316546887159348, -0.07148917019367218, 0.08483265340328217, 0.04973085969686508, 0.023403307422995567, -0.03980764001607895, -0.06086891517043114, -0.012414291501045227, 0.13631412386894226, -0.18523195385932922, -0.03783931955695152 ]
null
null
transformers
"hello"
{}
null
bada/test
[ "transformers", "pytorch", "jax", "bert", "pretraining", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #pretraining #endpoints_compatible #region-us
"hello"
[]
[ "TAGS\n#transformers #pytorch #jax #bert #pretraining #endpoints_compatible #region-us \n" ]
[ 29 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #pretraining #endpoints_compatible #region-us \n" ]
[ -0.04758477956056595, 0.025376562029123306, -0.010041696019470692, -0.026890631765127182, 0.12169862538576126, 0.03283354640007019, 0.04852999001741409, 0.06603716313838959, 0.06981325894594193, -0.02207045815885067, 0.16288776695728302, 0.1839056760072708, -0.04884219914674759, -0.004254240542650223, -0.038344986736774445, -0.2528855800628662, 0.053164176642894745, 0.08647771924734116, -0.05832131579518318, 0.11299949139356613, 0.05624457448720932, -0.12078862637281418, 0.052798882126808167, -0.032205283641815186, -0.13158826529979706, 0.021922975778579712, 0.007353800814598799, -0.06394176930189133, 0.12668900191783905, 0.0036244255024939775, 0.1890132576227188, 0.034819357097148895, -0.07439008355140686, -0.17548999190330505, 0.03879281133413315, 0.0039975885301828384, -0.05654344707727432, 0.025172775611281395, 0.042274296283721924, -0.07451727241277695, 0.01493008341640234, 0.06636687368154526, 0.016456041485071182, 0.0354442298412323, -0.21587271988391876, -0.16798201203346252, -0.036612942814826965, 0.03626810014247894, 0.06971926242113113, 0.07811176031827927, 0.009043033234775066, 0.15474732220172882, -0.15645740926265717, 0.06996539980173111, 0.14205828309059143, -0.31152382493019104, -0.01930762641131878, 0.07578881084918976, 0.10666734725236893, 0.05848696827888489, -0.03887427598237991, 0.03007550723850727, 0.018684636801481247, 0.016603868454694748, -0.03030146099627018, -0.09776552766561508, -0.034215107560157776, 0.07640163600444794, -0.11875150352716446, -0.0789615735411644, 0.2424411177635193, -0.025206539779901505, 0.03936931863427162, 0.04909973591566086, -0.06134871393442154, -0.08005161583423615, -0.005129881203174591, -0.03513208404183388, -0.013564311899244785, 0.07285438477993011, 0.01604410447180271, -0.018982702866196632, -0.12464295327663422, 0.022556308656930923, -0.16647079586982727, 0.23840190470218658, 0.02205256186425686, 0.06097600609064102, -0.21879221498966217, 0.06654202938079834, -0.0002153612585971132, -0.0690160021185875, 0.026950260624289513, -0.08134224265813828, 0.016524923965334892, -0.03067360445857048, -0.06570704281330109, 0.03185085952281952, 0.02740018628537655, 0.1444564014673233, 0.01235243584960699, 0.04652935266494751, 0.039747390896081924, 0.11681018024682999, 0.025414664298295975, 0.08319246023893356, 0.03141963854432106, 0.02190125547349453, 0.011679376475512981, -0.08166307210922241, -0.016955742612481117, -0.02284376509487629, -0.09105930477380753, -0.052212171256542206, 0.04482024535536766, 0.09012977033853531, 0.002317315200343728, 0.014513097703456879, -0.09935379773378372, -0.005556248128414154, -0.006815001368522644, -0.03224998340010643, -0.03014262206852436, -0.011581262573599815, 0.04805031791329384, 0.13356344401836395, -0.02790539152920246, 0.010618642903864384, -0.011646706610918045, 0.10964853316545486, -0.09140199422836304, -0.016210924834012985, -0.05830007791519165, -0.006138738244771957, 0.04258488863706589, -0.10275507718324661, 0.09343232214450836, -0.11879654973745346, -0.09725890308618546, 0.006814936175942421, 0.08113539963960648, 0.031696416437625885, -0.0032427951227873564, -0.012564903125166893, -0.016521934419870377, -0.03629060462117195, -0.0573909729719162, -0.0264847781509161, -0.0645684152841568, 0.10177116096019745, -0.0085616959258914, 0.04125301167368889, -0.06143946945667267, 0.052033957093954086, -0.09721038490533829, 0.039662547409534454, -0.10696694254875183, -0.030260084196925163, -0.012990543618798256, 0.19833293557167053, 0.010571029968559742, -0.050731293857097626, -0.14629696309566498, 0.05442944914102554, -0.04291445389389992, 0.17367485165596008, -0.06156576797366142, -0.10692965239286423, 0.2305416315793991, -0.08521339297294617, -0.17447109520435333, 0.05714026466012001, -0.015460262075066566, 0.053842026740312576, 0.07300862669944763, 0.18539166450500488, 0.00903394352644682, -0.11333677172660828, 0.10339004546403885, 0.13637593388557434, -0.10959412157535553, -0.13198797404766083, 0.03938274830579758, -0.02279824949800968, -0.12618263065814972, 0.027526507154107094, 0.011768346652388573, 0.07868161052465439, -0.09588013589382172, -0.029689455404877663, 0.002884102985262871, -0.01735878922045231, 0.029856299981474876, 0.08388436585664749, 0.08727211505174637, -0.08028150349855423, -0.017339203506708145, 0.006027791649103165, 0.0070586916990578175, 0.05130839720368385, 0.041554201394319534, -0.056027501821517944, 0.052375271916389465, -0.037873852998018265, -0.013774937950074673, -0.1958271861076355, -0.0789342075586319, -0.016674423590302467, 0.10328850895166397, -0.025971999391913414, 0.20274023711681366, 0.0918864756822586, -0.08201758563518524, -0.007370562758296728, -0.04041201248764992, 0.08669956028461456, 0.007555154152214527, -0.042645230889320374, -0.0788133516907692, 0.012404808774590492, -0.06526057422161102, -0.039830368012189865, -0.07701463252305984, 0.012522811070084572, 0.0747489482164383, 0.10472042113542557, 0.011516472324728966, 0.014194635674357414, -0.03217078372836113, 0.0518585629761219, -0.008207371458411217, 0.016298700124025345, 0.11433875560760498, 0.017264211550354958, -0.05044892430305481, 0.10508470982313156, -0.0796843022108078, 0.3379996418952942, 0.1805715411901474, -0.3153628706932068, -0.0011415781918913126, -0.009584764949977398, -0.04273047670722008, 0.01491453219205141, 0.10394691675901413, 0.02079625055193901, 0.11760787665843964, 0.021936796605587006, 0.11275681853294373, -0.032868318259716034, -0.053560201078653336, -0.008895720355212688, -0.04473020136356354, -0.052144598215818405, 0.0900842547416687, 0.0660732239484787, -0.15620069205760956, 0.15276029706001282, 0.2834061086177826, 0.0161992609500885, 0.12887687981128693, -0.05239143222570419, -0.01264589186757803, 0.016060220077633858, 0.01790418103337288, -0.05083385109901428, 0.0640946626663208, -0.28027600049972534, -0.048239417374134064, 0.04627029970288277, 0.008735853247344494, 0.08143450319766998, -0.15827003121376038, -0.09083732962608337, 0.014465195126831532, 0.04552235081791878, -0.04861642047762871, 0.08626738935709, 0.006301594898104668, 0.07003342360258102, 0.027432940900325775, -0.04383918270468712, 0.11072288453578949, -0.010219133459031582, -0.0588102862238884, 0.16722124814987183, -0.10981982946395874, -0.23858632147312164, -0.07923832535743713, -0.08523030579090118, 0.05995393916964531, 0.002401184756308794, 0.05964250862598419, -0.09286674112081528, -0.014566340483725071, 0.10035161674022675, 0.043911367654800415, -0.20327003300189972, 0.021479066461324692, -0.029361054301261902, 0.06213165074586868, -0.09410492330789566, -0.056792739778757095, -0.06334526836872101, -0.071891188621521, -0.07211031019687653, 0.0799049437046051, -0.11083874106407166, 0.04225043207406998, 0.11146104335784912, 0.05096304789185524, 0.07234664261341095, -0.015427562408149242, 0.210290789604187, -0.08039557188749313, -0.05970446765422821, 0.15208926796913147, -0.05649347975850105, 0.0653720423579216, 0.10023556649684906, 0.03706760331988335, -0.08244657516479492, -0.02175998128950596, -0.05141335725784302, -0.08270202577114105, -0.2059893012046814, -0.05854338780045509, -0.1326533854007721, 0.0002580733271315694, 0.02137463167309761, 0.0393485464155674, 0.07921664416790009, 0.06292399019002914, 0.039095740765333176, -0.07608423382043839, -0.03379439190030098, 0.04518570378422737, 0.16047303378582, -0.04561249911785126, 0.07162540405988693, -0.03734436258673668, -0.07699783146381378, 0.04507330805063248, 0.05224573612213135, 0.14619846642017365, 0.12699271738529205, 0.009816296398639679, 0.07922627776861191, 0.17724576592445374, 0.14369873702526093, 0.10741399973630905, -0.02959275431931019, -0.0658242255449295, -0.025382867082953453, 0.006822396535426378, -0.093505859375, 0.0228281207382679, 0.09420405328273773, -0.08440548181533813, -0.07247290760278702, -0.21407358348369598, 0.03508314490318298, 0.09763187915086746, 0.02959945797920227, -0.16308599710464478, -0.003423915710300207, 0.08140986412763596, -0.006873996928334236, -0.061884358525276184, 0.1150386705994606, -0.0117102125659585, -0.12265268713235855, 0.036302048712968826, -0.038278207182884216, 0.10297144204378128, -0.02637683041393757, 0.09303030371665955, -0.025390638038516045, -0.11551665514707565, 0.050832588225603104, 0.07477909326553345, -0.27916887402534485, 0.27387961745262146, -0.003368615172803402, -0.03478718921542168, -0.027206802740693092, -0.04565313458442688, -0.016567230224609375, 0.17528851330280304, 0.14892707765102386, 0.035631123930215836, -0.06364226341247559, -0.0763731524348259, 0.05687133967876434, 0.027116822078824043, 0.12139637023210526, -0.050879817456007004, -0.03788547217845917, -0.006101175677031279, 0.008845535106956959, -0.0047314949333667755, 0.005148213356733322, 0.0005474514327943325, -0.12274827063083649, 0.05273427069187164, -0.025029325857758522, 0.05638139694929123, -0.030214587226510048, -0.014232651330530643, -0.03820224106311798, 0.1524651050567627, -0.06369709223508835, -0.03562917560338974, -0.07683729380369186, -0.1372866928577423, 0.0927252322435379, -0.0775918960571289, 0.032606091350317, -0.07675188034772873, -0.05944591760635376, -0.08561501652002335, -0.14271599054336548, 0.14569464325904846, -0.11842286586761475, 0.015895411372184753, -0.06937769055366516, 0.20424962043762207, -0.016673456877470016, 0.021974798291921616, -0.01809825375676155, 0.022552266716957092, -0.10178772360086441, -0.06430039554834366, 0.038193680346012115, -0.05352756381034851, 0.05384057015180588, -0.008993450552225113, -0.0629202350974083, 0.10328070819377899, 0.017419669777154922, 0.0029704980552196503, 0.2251044660806656, 0.2365642786026001, -0.023217396810650826, 0.11565325409173965, 0.17714114487171173, -0.05118212476372719, -0.27821266651153564, -0.02284403145313263, -0.12365742772817612, -0.036728765815496445, -0.04181229695677757, -0.1797652691602707, 0.10324335843324661, 0.04005245119333267, -0.026892634108662605, 0.10708922147750854, -0.2751569449901581, -0.047579795122146606, 0.15807297825813293, 0.020378267392516136, 0.5222597122192383, -0.1348620504140854, -0.06071731448173523, 0.04110337793827057, -0.1898440569639206, 0.1265621781349182, 0.03256361186504364, 0.05081644281744957, -0.005798292811959982, 0.02115809917449951, 0.0230173971503973, -0.08147402107715607, 0.12349952012300491, 0.028232716023921967, 0.03788936138153076, -0.06487257033586502, -0.13532249629497528, 0.05660117045044899, -0.008724461309611797, -0.06592285633087158, 0.07960466295480728, 0.005432679783552885, -0.18660488724708557, -0.019073808565735817, -0.12388879805803299, 0.05996127426624298, 0.007277187891304493, -0.05523562431335449, 0.018301833420991898, 0.009977814741432667, -0.027626417577266693, 0.015138378366827965, 0.29683589935302734, -0.029347561299800873, 0.1637185961008072, -0.0018685496179386973, 0.13350482285022736, -0.18770281970500946, -0.07327540218830109, -0.04503950849175453, -0.04549866542220116, 0.08921278268098831, -0.03790426254272461, 0.022482210770249367, 0.14069199562072754, -0.0021410533227026463, 0.019702013581991196, 0.12155880779027939, 0.006553066894412041, -0.04228489473462105, 0.11387378722429276, -0.24115334451198578, -0.0718904584646225, -0.05027800798416138, 0.013970880769193172, 0.08601261675357819, 0.0834246575832367, 0.08244827389717102, 0.04818698763847351, -0.025859303772449493, -0.01456725224852562, -0.033266037702560425, -0.08887087553739548, -0.001902748248539865, 0.07184335589408875, 0.03294219821691513, -0.09269578009843826, 0.041420046240091324, -0.012934214435517788, -0.2646346688270569, -0.038209155201911926, 0.1358453929424286, -0.12191460281610489, -0.10285573452711105, -0.08504001796245575, 0.12675932049751282, -0.0940762311220169, -0.022578801959753036, -0.016404133290052414, -0.11725315451622009, 0.07649963349103928, 0.2254946231842041, 0.10796516388654709, 0.08866938203573227, -0.015146702527999878, 0.011509934440255165, 0.01661132648587227, -0.04916441813111305, -0.02467065304517746, 0.004210638348013163, -0.09036058932542801, 0.013246756047010422, -0.002212297171354294, 0.1500893533229828, -0.07581816613674164, -0.09227654337882996, -0.19172054529190063, 0.09045315533876419, -0.040490418672561646, -0.12102796882390976, -0.10888654738664627, -0.06561612337827682, 0.009221984073519707, -0.10547026991844177, -0.040638696402311325, -0.046728868037462234, -0.1373746395111084, 0.07576927542686462, 0.03941905498504639, 0.0068800849840044975, -0.022380191832780838, -0.036307331174612045, 0.14129263162612915, -0.0499434694647789, 0.08735932409763336, 0.14733164012432098, -0.0400492362678051, 0.12777160108089447, -0.08010154962539673, -0.10317514836788177, 0.0831993967294693, 0.0044362484477460384, 0.05231817811727524, 0.07456809282302856, -0.0015187719836831093, 0.05087457597255707, 0.025899512693285942, 0.05593207851052284, -0.016439829021692276, -0.11214102059602737, 0.026233157142996788, 0.07706455886363983, -0.17388911545276642, -0.028021322563290596, -0.08372671157121658, 0.14319832623004913, 0.014229261316359043, 0.10417653620243073, 0.011460685171186924, 0.08844538778066635, -0.057199202477931976, 0.01749131642282009, 0.010318719781935215, -0.1626819372177124, 0.009655427187681198, -0.08048654347658157, 0.01437891460955143, -0.007818187586963177, 0.22755765914916992, -0.04810832068324089, -0.0016121218213811517, 0.06209748610854149, 0.00649160286411643, -0.007749933283776045, 0.0389087088406086, 0.21731366217136383, 0.1079903393983841, -0.04756442829966545, -0.07607609033584595, 0.08628802001476288, 0.004852649290114641, -0.05844290554523468, 0.13485294580459595, 0.15917764604091644, 0.050128065049648285, 0.10049556195735931, -0.029297614470124245, 0.034177299588918686, -0.0634140744805336, -0.1462387591600418, 0.02415655553340912, 0.02243519015610218, 0.0017781590577214956, 0.13153696060180664, 0.1631545126438141, -0.04050573334097862, 0.041866619139909744, -0.0424455963075161, -0.02175520360469818, -0.1593099683523178, -0.05927267670631409, -0.02947325073182583, -0.0717400312423706, 0.02034137025475502, -0.06380745768547058, -0.013293454423546791, 0.13811615109443665, 0.046745359897613525, -0.023505669087171555, 0.09689633548259735, 0.10772862285375595, -0.04717663303017616, 0.038137201219797134, -0.009078153409063816, 0.04533446952700615, -0.052931446582078934, -0.019041882827878, -0.12705081701278687, -0.08589401096105576, -0.0700870007276535, 0.011189188808202744, -0.11311700195074081, -0.02774183824658394, -0.09148798882961273, -0.10004114359617233, -0.047526322305202484, 0.06904581189155579, -0.02742740698158741, 0.08219726383686066, -0.026133691892027855, 0.017285317182540894, 0.014311814680695534, 0.18881119787693024, -0.05316367372870445, -0.02847374975681305, -0.00687430240213871, 0.18283310532569885, 0.03911268711090088, 0.10358031094074249, 0.0010854760184884071, 0.05653609707951546, -0.04030516371130943, 0.33644482493400574, 0.24365901947021484, -0.07794445753097534, 0.038950398564338684, 0.06782548129558563, 0.03899161145091057, 0.08093918114900589, 0.09545615315437317, 0.10454796999692917, 0.2717442512512207, -0.13879893720149994, -0.03059009648859501, -0.05177813023328781, 0.006821046117693186, -0.06967976689338684, 0.04716334119439125, 0.05790933594107628, -0.03492337837815285, -0.07258761674165726, 0.08820143342018127, -0.15367697179317474, 0.065158911049366, 0.06188080459833145, -0.21656474471092224, -0.05266902223229408, -0.05733874812722206, 0.16729126870632172, -0.022717256098985672, 0.12460428476333618, -0.03391995280981064, -0.1418391764163971, 0.04760276898741722, 0.03858992084860802, -0.2181340456008911, -0.12715193629264832, 0.14826631546020508, 0.057387713342905045, 0.01840929687023163, -0.03752914443612099, 0.011408603750169277, 0.0656089186668396, 0.042262446135282516, -0.03106369450688362, -0.02747844159603119, 0.04505091905593872, -0.06128906458616257, -0.14094650745391846, -0.011302746832370758, 0.016690149903297424, -0.023031480610370636, 0.037768956273794174, -0.19101503491401672, 0.048749130219221115, -0.06523895263671875, -0.07040996849536896, -0.016027076169848442, 0.05516074225306511, -0.04220268875360489, 0.05362638086080551, 0.06945189833641052, 0.020758114755153656, -0.05000750347971916, -0.04336467385292053, -0.015918858349323273, 0.06816285848617554, -0.016245948150753975, -0.13243816792964935, -0.045595910400152206, -0.07746852189302444, 0.08552899211645126, -0.030757173895835876, -0.09250643849372864, -0.024549206718802452, -0.028491590172052383, 0.053531382232904434, -0.1132727637887001, 0.02937343157827854, 0.016194740310311317, 0.053137268871068954, -0.00019143011013511568, -0.03040502965450287, 0.05717053636908531, 0.058099694550037384, -0.10681519657373428, -0.07785963267087936 ]
null
null
transformers
# Genji-python 6B For example usage or to easily use the model you can check our colab notebook: [Notebook](https://colab.research.google.com/drive/1PnWpx02IEUkY8jhLKd_NewUGEXahAska?usp=sharing) ## Model Description Genji is a transformer model finetuned on EleutherAI's GPT-J 6B model. This particular model is trained on python only code approaching 4GB in size. Split model has the checkpoints splitted, which makes it use less system RAM while loading and makes it faster to load. This model needs more effort to set up as you need to install git-lfs and pull the repo. | Hyperparameter | Value | |-------------------|--------| | n_parameters | 6,053,381,344 | | n_layers | 28* | | d_model | 4,096 | | d_ff | 16,384 | | n_heads | 16 | | d_head | 256 | | n_ctx | 2,048 | | n_vocab | 50,400 (same tokenizer as GPT-2/3) | | position encoding | [Rotary position encodings (RoPE)](https://arxiv.org/abs/2104.09864) | | RoPE dimensions | [64](https://github.com/kingoflolz/mesh-transformer-jax/blob/f2aa66e0925de6593dcbb70e72399b97b4130482/mesh_transformer/layers.py#L223) | `*` each layer consists of one feedforward block and one self attention block The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary position encodings (RoPE) was applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. ## Training data GPT-J 6B was pretrained on the [Pile](pile.eleuther.ai), a large scale curated dataset created by EleutherAI for the purpose of training this model. After the pre-training, it's finetuned on the python code that was taken from the Pile. ## Training procedure Genji-python-6B is trained for 20k steps on around 655 million tokens with learning rate of 2e-06 ## Intended Use This model is trained for assistence on writing python code and having fun trying weird stuff with it. ### How to use This model is only usable with our fork because GPT-J is not merged to the main transformers repo yet. When it's merged, we will make this model easily loadable. For now, you need to use this fork: [Fork](https://github.com/finetuneanon/transformers) to install with pip: ```bash pip install git+https://github.com/finetuneanon/transformers@gpt-neo-localattention3-rp-b ``` **git-lfs** also needs to be installed, on ubuntu: ```bash apt install git-lfs ``` after it's installed, initialize git-lfs: ```bash git lfs install ``` then clone this repo: ```bash git clone https://huggingface.co/NovelAI/genji-python-6B-split ``` Now we can load the model. We recommend the usage of the model as FP16. That way, it fits in 16GB VRAM cards. How to use: ```python from transformers import ( AutoTokenizer, AutoModelForCausalLM, GPTNeoForCausalLM, ) model = AutoModelForCausalLM.from_pretrained("genji-python-6B-split/model").half().eval().cuda() tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neo-2.7B") text = '''def print_customer_name''' tokens = tokenizer(text, return_tensors="pt").input_ids generated_tokens = model.generate(tokens.long().cuda(), use_cache=True, do_sample=True, top_k=50, temperature=0.3, top_p=0.9, repetition_penalty=1.125, min_length=1, max_length=len(tokens[0]) + 400, pad_token_id=tokenizer.eos_token_id) last_tokens = generated_tokens[0][len(tokens[0]):] generated_text = tokenizer.decode(last_tokens) print("Generation:\n" + generated_text) ``` When ran, this code generates: ```python Prompt: def print_customer_name Generation: (self, customer): """Print the name of a customer.""" if not self.is_valid(): return print("Customer: {}".format(customer)) ``` For example usage, you can see our colab notebook as well: [Notebook](https://colab.research.google.com/drive/1PnWpx02IEUkY8jhLKd_NewUGEXahAska?usp=sharing) ## Eval results TBD ## Acknowledgements This project was possible because of the compute provided by the [TPU Research Cloud](https://sites.research.google/trc/) and [EleutherAI](https://eleuther.ai/) for pretraining of the GPT-J 6B. Thanks to everyone who contributed to this project: - [Aero](https://github.com/AeroScripts) - [Finetune](https://github.com/finetuneanon) - [Kurumuz](https://github.com/kurumuz)
{"language": ["en"], "license": "apache-2.0", "tags": ["pytorch", "causal-lm"], "datasets": ["the Pile"]}
text-generation
baffo32/genji-python-6B-split
[ "transformers", "gpt_neo", "text-generation", "pytorch", "causal-lm", "en", "arxiv:2104.09864", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2104.09864" ]
[ "en" ]
TAGS #transformers #gpt_neo #text-generation #pytorch #causal-lm #en #arxiv-2104.09864 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
Genji-python 6B =============== For example usage or to easily use the model you can check our colab notebook: Notebook Model Description ----------------- Genji is a transformer model finetuned on EleutherAI's GPT-J 6B model. This particular model is trained on python only code approaching 4GB in size. Split model has the checkpoints splitted, which makes it use less system RAM while loading and makes it faster to load. This model needs more effort to set up as you need to install git-lfs and pull the repo. '\*' each layer consists of one feedforward block and one self attention block The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary position encodings (RoPE) was applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. Training data ------------- GPT-J 6B was pretrained on the Pile, a large scale curated dataset created by EleutherAI for the purpose of training this model. After the pre-training, it's finetuned on the python code that was taken from the Pile. Training procedure ------------------ Genji-python-6B is trained for 20k steps on around 655 million tokens with learning rate of 2e-06 Intended Use ------------ This model is trained for assistence on writing python code and having fun trying weird stuff with it. ### How to use This model is only usable with our fork because GPT-J is not merged to the main transformers repo yet. When it's merged, we will make this model easily loadable. For now, you need to use this fork: Fork to install with pip: git-lfs also needs to be installed, on ubuntu: after it's installed, initialize git-lfs: then clone this repo: Now we can load the model. We recommend the usage of the model as FP16. That way, it fits in 16GB VRAM cards. How to use: When ran, this code generates: For example usage, you can see our colab notebook as well: Notebook Eval results ------------ TBD Acknowledgements ---------------- This project was possible because of the compute provided by the TPU Research Cloud and EleutherAI for pretraining of the GPT-J 6B. Thanks to everyone who contributed to this project: * Aero * Finetune * Kurumuz
[ "### How to use\n\n\nThis model is only usable with our fork because GPT-J is not merged to the main transformers repo yet. When it's merged, we will make this model easily loadable.\nFor now, you need to use this fork:\nFork\n\n\nto install with pip:\n\n\ngit-lfs also needs to be installed, on ubuntu:\n\n\nafter it's installed, initialize git-lfs:\n\n\nthen clone this repo:\n\n\nNow we can load the model.\n\n\nWe recommend the usage of the model as FP16. That way, it fits in 16GB VRAM cards.\n\n\nHow to use:\n\n\nWhen ran, this code generates:\n\n\nFor example usage, you can see our colab notebook as well:\nNotebook\n\n\nEval results\n------------\n\n\nTBD\n\n\nAcknowledgements\n----------------\n\n\nThis project was possible because of the compute provided by the\nTPU Research Cloud and EleutherAI for pretraining of the GPT-J 6B.\n\n\nThanks to everyone who contributed to this project:\n\n\n* Aero\n* Finetune\n* Kurumuz" ]
[ "TAGS\n#transformers #gpt_neo #text-generation #pytorch #causal-lm #en #arxiv-2104.09864 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### How to use\n\n\nThis model is only usable with our fork because GPT-J is not merged to the main transformers repo yet. When it's merged, we will make this model easily loadable.\nFor now, you need to use this fork:\nFork\n\n\nto install with pip:\n\n\ngit-lfs also needs to be installed, on ubuntu:\n\n\nafter it's installed, initialize git-lfs:\n\n\nthen clone this repo:\n\n\nNow we can load the model.\n\n\nWe recommend the usage of the model as FP16. That way, it fits in 16GB VRAM cards.\n\n\nHow to use:\n\n\nWhen ran, this code generates:\n\n\nFor example usage, you can see our colab notebook as well:\nNotebook\n\n\nEval results\n------------\n\n\nTBD\n\n\nAcknowledgements\n----------------\n\n\nThis project was possible because of the compute provided by the\nTPU Research Cloud and EleutherAI for pretraining of the GPT-J 6B.\n\n\nThanks to everyone who contributed to this project:\n\n\n* Aero\n* Finetune\n* Kurumuz" ]
[ 63, 225 ]
[ "passage: TAGS\n#transformers #gpt_neo #text-generation #pytorch #causal-lm #en #arxiv-2104.09864 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nThis model is only usable with our fork because GPT-J is not merged to the main transformers repo yet. When it's merged, we will make this model easily loadable.\nFor now, you need to use this fork:\nFork\n\n\nto install with pip:\n\n\ngit-lfs also needs to be installed, on ubuntu:\n\n\nafter it's installed, initialize git-lfs:\n\n\nthen clone this repo:\n\n\nNow we can load the model.\n\n\nWe recommend the usage of the model as FP16. That way, it fits in 16GB VRAM cards.\n\n\nHow to use:\n\n\nWhen ran, this code generates:\n\n\nFor example usage, you can see our colab notebook as well:\nNotebook\n\n\nEval results\n------------\n\n\nTBD\n\n\nAcknowledgements\n----------------\n\n\nThis project was possible because of the compute provided by the\nTPU Research Cloud and EleutherAI for pretraining of the GPT-J 6B.\n\n\nThanks to everyone who contributed to this project:\n\n\n* Aero\n* Finetune\n* Kurumuz" ]
[ -0.058129772543907166, 0.004712838679552078, -0.0018231793073937297, 0.04361617565155029, 0.14815929532051086, 0.008804505690932274, 0.039787374436855316, 0.12351933866739273, -0.025011049583554268, 0.04911085218191147, 0.023150214925408363, 0.14145207405090332, 0.07067826390266418, 0.10403070598840714, 0.10522454231977463, -0.11920871585607529, -0.014792272821068764, -0.01801154389977455, -0.03711875155568123, 0.056727297604084015, 0.06602800637483597, -0.07601170986890793, 0.09321314096450806, 0.03944273665547371, -0.11782637238502502, 0.04330513998866081, -0.03958464786410332, -0.050158604979515076, 0.118487149477005, 0.10427334159612656, 0.0031165245454758406, -0.0035112116020172834, 0.022686343640089035, -0.15448734164237976, 0.019773995503783226, 0.07294347882270813, -0.0041907611303031445, 0.06150289997458458, 0.000604584813117981, 0.00459316186606884, 0.14168573915958405, -0.03886491805315018, -0.018468333408236504, 0.06199149042367935, -0.13176779448986053, -0.17187295854091644, -0.1483134627342224, 0.05237686261534691, 0.106395423412323, 0.014147914946079254, 0.012370524927973747, 0.14057935774326324, 0.08467680215835571, 0.05729350820183754, 0.23177434504032135, -0.18844588100910187, -0.05556631088256836, 0.0877276211977005, 0.06322163343429565, -0.03954707831144333, -0.001123332418501377, 0.05884922295808792, 0.08769941329956055, 0.029701584950089455, 0.030504312366247177, -0.04405972734093666, -0.07678034901618958, 0.016587741672992706, -0.12397583574056625, -0.0873897522687912, 0.2111918181180954, -0.05115155130624771, -0.05027652159333229, -0.05568068102002144, -0.07018023729324341, -0.043126486241817474, 0.012254279106855392, -0.05076780170202255, 0.026183858513832092, 0.023564407601952553, 0.07002735882997513, -0.11197241395711899, -0.12100609391927719, -0.07219819724559784, 0.004614444449543953, 0.10122821480035782, 0.03313122317194939, 0.0321664959192276, -0.029688045382499695, 0.17507250607013702, -0.08647023141384125, -0.023286882787942886, -0.0608166866004467, -0.043155208230018616, 0.005620673298835754, -0.00987498089671135, 0.007861495949327946, -0.08881756663322449, 0.04984106123447418, 0.20776189863681793, 0.057940416038036346, -0.03657527640461922, 0.03294168785214424, 0.0350840799510479, 0.020909659564495087, 0.14434118568897247, -0.03203606978058815, -0.06605885177850723, 0.11527625471353531, -0.10923890769481659, 0.04138488695025444, -0.023627188056707382, -0.03321172297000885, -0.005702104419469833, -0.01175986509770155, 0.07918575406074524, 0.11992653459310532, 0.06848104298114777, -0.041715867817401886, -0.03079153411090374, 0.23298104107379913, -0.07276328653097153, 0.010219109244644642, -0.03776607662439346, -0.07564574480056763, -0.059323571622371674, 0.13901178538799286, -0.03023412451148033, -0.07719899713993073, -0.01656915619969368, -0.043761856853961945, -0.06001539155840874, -0.08673983812332153, -0.06863748282194138, 0.05489888787269592, -0.02244047075510025, 0.0002574732934590429, -0.16709865629673004, -0.17941077053546906, -0.023014983162283897, 0.03135385364294052, -0.031434670090675354, -0.002801261842250824, 0.08838723599910736, 0.036030739545822144, -0.007983905263245106, -0.015284308232367039, 0.026597537100315094, -0.04477893188595772, -0.006852153223007917, 0.05049990490078926, 0.08340074121952057, -0.04710238799452782, 0.022138940170407295, -0.04852868244051933, 0.04623999074101448, -0.1650441586971283, 0.014015795662999153, -0.08112022280693054, 0.06924821436405182, -0.11061928421258926, 0.0072686620987951756, 0.037226468324661255, -0.02592589519917965, 0.07972526550292969, 0.0660519003868103, -0.08442185819149017, -0.022997096180915833, 0.05200784653425217, -0.11575927585363388, -0.1084488183259964, 0.0711720734834671, 0.0829908549785614, 0.010830532759428024, 0.04858340695500374, 0.04802677407860756, 0.06833934038877487, -0.0825372114777565, -0.10097958147525787, -0.041009433567523956, -0.04832783713936806, -0.03616582974791527, 0.030175304040312767, 0.013515151105821133, 0.0012545667123049498, 0.03201352804899216, -0.012661983259022236, 0.04383814334869385, -0.013070853427052498, -0.014453329145908356, -0.08348608016967773, -0.05025378242135048, -0.08206920325756073, -0.0023081693798303604, -0.009198566898703575, -0.0092267831787467, -0.02790714055299759, -0.06555645912885666, 0.10566215962171555, -0.051842011511325836, 0.04791591688990593, -0.06604213267564774, 0.22553035616874695, -0.10167908668518066, 0.03981967270374298, -0.07799659669399261, -0.08107416331768036, 0.053430892527103424, -0.01424618624150753, -0.020350046455860138, 0.053557202219963074, -0.007763589732348919, 0.07619704306125641, 0.031914904713630676, 0.012680038809776306, -0.0019121570512652397, -0.00006309631135081872, -0.0007681418210268021, -0.05966460704803467, -0.04096405953168869, -0.03261033818125725, 0.01460292562842369, -0.07180637866258621, 0.05088123306632042, 0.09039916843175888, 0.10871876776218414, -0.006069079507142305, -0.043035734444856644, 0.050582028925418854, -0.0314803384244442, -0.04923463240265846, -0.06295254081487656, 0.04207755625247955, 0.028869478031992912, -0.014914777129888535, 0.05574532598257065, -0.10474976152181625, -0.16709831357002258, 0.07463736087083817, 0.002218453446403146, -0.04511267691850662, 0.02676423266530037, -0.053462810814380646, -0.06019260361790657, -0.0038221669383347034, -0.05231226980686188, 0.11440977454185486, 0.013464019633829594, 0.0979895144701004, -0.011441841721534729, -0.07247284799814224, 0.03014126792550087, -0.04262377321720123, 0.0361841656267643, 0.03392571583390236, 0.0341307632625103, -0.012613087892532349, 0.06796803325414658, 0.08288412541151047, 0.01834872178733349, 0.060761675238609314, 0.0721743106842041, -0.07238665223121643, -0.0064987074583768845, -0.022516867145895958, 0.05646457523107529, 0.11351463943719864, -0.1297605186700821, -0.025547003373503685, 0.059634625911712646, -0.016726437956094742, 0.027944108471274376, -0.11138465255498886, 0.04015682637691498, 0.02128135785460472, -0.02039695717394352, 0.09367387741804123, 0.07029159367084503, -0.07898453623056412, 0.027267100289463997, -0.007958654314279556, 0.13166280090808868, -0.035099443048238754, 0.003999640699476004, -0.07110374420881271, 0.1275874227285385, -0.07294493913650513, -0.250687837600708, -0.132419615983963, -0.04626058042049408, -0.023023400455713272, 0.05168791860342026, 0.10367407649755478, -0.019010629504919052, -0.0875711739063263, -0.0002772983571048826, 0.2054637223482132, 0.08372107893228531, -0.05732725188136101, 0.014409517869353294, -0.06433318555355072, 0.025220992043614388, -0.13302309811115265, -0.0036781467497348785, 0.07979263365268707, -0.12321796268224716, 0.08527372032403946, 0.0639914944767952, 0.10431329905986786, 0.017345059663057327, -0.031087389215826988, -0.05366181582212448, 0.010316838510334492, 0.1964745670557022, -0.10118070244789124, 0.05007041618227959, 0.19215485453605652, -0.010482225567102432, 0.02148851379752159, 0.02585975080728531, 0.008855503052473068, -0.079877108335495, 0.01976068690419197, -0.00047502125380560756, -0.0764198824763298, -0.183477982878685, -0.07764670997858047, -0.06220528110861778, 0.1273021101951599, 0.06524135172367096, 0.05316655710339546, -0.02404634840786457, 0.09110389649868011, -0.08543992042541504, 0.11128468811511993, -0.08306024968624115, 0.12119915336370468, 0.14043688774108887, -0.05768691748380661, 0.025810904800891876, -0.050453707575798035, 0.026219431310892105, 0.1470639407634735, 0.10281907767057419, 0.052965495735406876, -0.021970944479107857, 0.11162909120321274, 0.0735909715294838, 0.09241487830877304, -0.015465994365513325, 0.09500046074390411, -0.048267193138599396, -0.014531468972563744, -0.0017117621609941125, -0.02310194820165634, -0.07422741502523422, 0.03616833686828613, -0.037600401788949966, -0.023976286873221397, -0.031790461391210556, -0.03473082184791565, 0.05865679681301117, 0.1784062385559082, 0.037505775690078735, -0.15715745091438293, -0.07392727583646774, 0.027003969997167587, -0.0030511408112943172, -0.10001949965953827, -0.026204733178019524, 0.04126084968447685, -0.16480451822280884, 0.03978952765464783, -0.004206671379506588, 0.07134059816598892, -0.12880884110927582, -0.019885219633579254, 0.005851447582244873, 0.05399654433131218, 0.035308048129081726, 0.10575207322835922, -0.08380404114723206, 0.024303125217556953, 0.027195367962121964, 0.0035674350801855326, -0.054141655564308167, 0.06698985397815704, 0.011700759641826153, 0.10069693624973297, 0.13294045627117157, 0.0061193788424134254, -0.022624796256422997, -0.04641726240515709, -0.12144798040390015, 0.03465980663895607, -0.057237911969423294, -0.13509581983089447, 0.02713482826948166, -0.033868588507175446, 0.02769484929740429, -0.04814595356583595, 0.045008644461631775, -0.11404278129339218, -0.15102733671665192, 0.1087268516421318, -0.003821534337475896, 0.05352294445037842, -0.06597573310136795, -0.020828958600759506, -0.05337795615196228, 0.25307410955429077, 0.014705215580761433, -0.04053359478712082, -0.06957774609327316, 0.056579235941171646, 0.05910585820674896, -0.07773240655660629, 0.04380043223500252, -0.03927119821310043, 0.09380710124969482, -0.03901396691799164, -0.060248080641031265, 0.03615973889827728, -0.14300528168678284, -0.13906125724315643, -0.0042085908353328705, 0.01580728031694889, -0.03594612702727318, 0.014800498262047768, 0.05016418918967247, 0.028267240151762962, -0.059647560119628906, -0.07137431204319, 0.010274970903992653, 0.14039084315299988, 0.09818686544895172, -0.01919100619852543, -0.051465146243572235, 0.021819744259119034, -0.027253661304712296, -0.09551453590393066, 0.12112312763929367, 0.22978359460830688, -0.07849825173616409, 0.05602632835507393, 0.08443751931190491, -0.09673423320055008, -0.22260458767414093, -0.06265372037887573, -0.05057549476623535, -0.003384373849257827, -0.03621837496757507, -0.1298878788948059, 0.0897829532623291, 0.05699577555060387, -0.02967631258070469, 0.12262200564146042, -0.18369272351264954, -0.09013310819864273, 0.06995505839586258, 0.04919636622071266, 0.06673675030469894, -0.14620478451251984, -0.04223787784576416, -0.047319553792476654, -0.16229115426540375, 0.04651439189910889, -0.0974612683057785, 0.08924586325883865, -0.08696970343589783, 0.012783987447619438, 0.013978327624499798, -0.04861682653427124, 0.1300184428691864, -0.10181651264429092, -0.006777519825845957, -0.11867912858724594, 0.07026173174381256, -0.007419829722493887, -0.08514511585235596, 0.1119503378868103, 0.02528109960258007, 0.044795017689466476, -0.05422002077102661, -0.011493237689137459, -0.025122838094830513, 0.11959826946258545, -0.024415768682956696, -0.07015394419431686, -0.08249121159315109, 0.013940269127488136, 0.004808064550161362, -0.02859419584274292, -0.03139884024858475, 0.0028599028009921312, 0.05980026721954346, 0.11145401000976562, -0.001548881409689784, -0.08917922526597977, -0.04185134544968605, 0.020091379061341286, 0.022306742146611214, 0.0796646922826767, -0.09300640225410461, 0.05395263805985451, 0.0745779499411583, -0.023678939789533615, 0.06438273936510086, 0.013267077505588531, -0.11654528230428696, 0.025783861055970192, 0.10389162600040436, -0.2115163654088974, -0.08269937336444855, -0.04194237291812897, 0.06221709027886391, 0.008409368805587292, 0.05029141902923584, 0.14092133939266205, -0.062385592609643936, -0.06420794129371643, 0.03234396502375603, 0.029933396726846695, -0.04921218752861023, 0.09515165537595749, 0.023998664692044258, 0.0020461168605834246, -0.04981536045670509, 0.07402309775352478, 0.06086215004324913, -0.08338063955307007, -0.03576236590743065, 0.03454972058534622, -0.0638023167848587, -0.1038360521197319, 0.004046347923576832, 0.03942599520087242, -0.1056048572063446, -0.06692685186862946, 0.002646669512614608, 0.043955400586128235, -0.02870267443358898, 0.026938248425722122, 0.026505092158913612, 0.02931870147585869, 0.03135832026600838, 0.04685288295149803, -0.05105577036738396, 0.04849303141236305, 0.006423338782042265, 0.09019133448600769, -0.07977202534675598, 0.04166668280959129, 0.010664799250662327, 0.11290419846773148, -0.015359984710812569, -0.007487161550670862, -0.05514581874012947, -0.050187692046165466, -0.0445246621966362, -0.009199759922921658, -0.06318610906600952, -0.008553566411137581, 0.043558649718761444, 0.012136142700910568, -0.014671720564365387, 0.036377228796482086, -0.04069528356194496, -0.039019957184791565, -0.054531246423721313, 0.09063907712697983, -0.1394951492547989, -0.01607162319123745, 0.08660664409399033, -0.03124709613621235, 0.10787830501794815, -0.015568592585623264, 0.014105300419032574, 0.00018732351600192487, -0.05207109451293945, 0.03205331787467003, 0.04118817299604416, 0.02458702214062214, 0.057455554604530334, -0.14345808327198029, -0.0018023678567260504, -0.007942596450448036, -0.04481979086995125, -0.03565112501382828, 0.08368505537509918, -0.1316518783569336, -0.08420919626951218, 0.031203487887978554, -0.05214070528745651, -0.011864738538861275, -0.010808033868670464, -0.01351901888847351, 0.11474484950304031, 0.03651859611272812, -0.013819653540849686, 0.04340662807226181, -0.08513151109218597, 0.010843878611922264, 0.01427649985998869, -0.04433858394622803, 0.04344748705625534, -0.02374332770705223, 0.01301097497344017, -0.01460369024425745, 0.14596138894557953, 0.028341514989733696, 0.022952580824494362, -0.0023068590089678764, 0.007048147264868021, -0.018545106053352356, -0.04237036406993866, 0.04323762655258179, -0.05045848339796066, 0.05209856107831001, -0.056192196905612946, 0.1031947135925293, 0.02547360211610794, -0.013006146997213364, 0.05963493883609772, -0.0644996166229248, -0.006879637483507395, 0.059425052255392075, -0.011248397640883923, -0.08514571934938431, -0.061286039650440216, -0.152903214097023, -0.03199576586484909, 0.12799540162086487, -0.05854935199022293, 0.03438947722315788, 0.11660595238208771, -0.12143617123365402, 0.025062041357159615, -0.036537811160087585, -0.045056600123643875, -0.09761802852153778, -0.10602830350399017, -0.019584117457270622, -0.12438273429870605, -0.0023175394162535667, -0.096607506275177, 0.025939906015992165, 0.007834814488887787, 0.016231652349233627, -0.021530060097575188, 0.15371254086494446, -0.002895267913118005, -0.04093985632061958, -0.06354087591171265, 0.03666400909423828, -0.0018957910360768437, -0.08777524530887604, 0.07222660630941391, -0.048661038279533386, 0.06614465266466141, 0.052063606679439545, 0.052472807466983795, 0.061692722141742706, 0.07485969364643097, -0.027998581528663635, -0.03820071741938591, -0.04482269659638405, 0.014003593474626541, 0.013475134037435055, 0.10577963292598724, 0.008787824772298336, 0.028771262615919113, 0.023208262398838997, 0.19259634613990784, -0.05514742061495781, -0.02615952119231224, -0.11084936559200287, 0.08662225306034088, -0.051434557884931564, -0.03620763495564461, -0.020015191286802292, -0.08689288794994354, -0.007931250147521496, 0.23805458843708038, 0.12808367609977722, 0.09466993063688278, 0.005222773645073175, 0.0037521449849009514, 0.008174434304237366, -0.00941803865134716, 0.05187974497675896, 0.09274499118328094, 0.060715336352586746, -0.06679084897041321, 0.016398031264543533, -0.06652329862117767, -0.002441644901409745, -0.04537104070186615, 0.09992068260908127, -0.05496305972337723, -0.006859770976006985, 0.014160914346575737, 0.05513853579759598, -0.021297305822372437, -0.1911621242761612, 0.00031298527028411627, -0.05024392157793045, -0.04345635697245598, 0.02412736415863037, 0.09121901541948318, -0.04452084004878998, 0.06509958207607269, -0.032079245895147324, 0.015482859686017036, 0.18796128034591675, 0.0037997940089553595, -0.13607315719127655, -0.049428168684244156, 0.07826102524995804, -0.04459764063358307, 0.22020018100738525, -0.03557012975215912, 0.02976171299815178, 0.08150625228881836, 0.03442063182592392, -0.17108044028282166, -0.03129266947507858, 0.008640961721539497, -0.09070125222206116, 0.022596418857574463, 0.09769967198371887, 0.00606509018689394, -0.058123789727687836, -0.0010588879231363535, -0.03518149256706238, 0.03593797981739044, 0.05876060947775841, 0.10512159764766693, -0.07164926081895828, 0.04836269095540047, -0.12028626352548599, 0.16906510293483734, 0.0830313041806221, -0.03740386292338371, -0.03939804434776306, -0.11047747731208801, 0.01729205809533596, 0.023140234872698784, 0.05353087559342384, -0.01405375823378563, -0.10783825069665909, 0.01956513151526451, 0.033297501504421234, 0.031135840341448784, -0.09214121848344803, -0.06618523597717285, 0.07198198139667511, -0.05438699573278427, -0.03121664747595787, 0.11567813158035278, 0.042605139315128326, 0.026658259332180023, -0.040807925164699554, -0.09665291756391525, -0.04766847565770149, 0.041411396116018295, -0.0927588939666748, -0.14990831911563873 ]
null
null
transformers
# GPT-J 6B ## Model Description GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters. <figure> | Hyperparameter | Value | |----------------------|------------| | \\(n_{parameters}\\) | 6053381344 | | \\(n_{layers}\\) | 28&ast; | | \\(d_{model}\\) | 4096 | | \\(d_{ff}\\) | 16384 | | \\(n_{heads}\\) | 16 | | \\(d_{head}\\) | 256 | | \\(n_{ctx}\\) | 2048 | | \\(n_{vocab}\\) | 50257/50400&dagger; (same tokenizer as GPT-2/3) | | Positional Encoding | [Rotary Position Embedding (RoPE)](https://arxiv.org/abs/2104.09864) | | RoPE Dimensions | [64](https://github.com/kingoflolz/mesh-transformer-jax/blob/f2aa66e0925de6593dcbb70e72399b97b4130482/mesh_transformer/layers.py#L223) | <figcaption><p><strong>&ast;</strong> Each layer consists of one feedforward block and one self attention block.</p> <p><strong>&dagger;</strong> Although the embedding matrix has a size of 50400, only 50257 entries are used by the GPT-2 tokenizer.</p></figcaption></figure> The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. ## Training data GPT-J 6B was trained on [the Pile](https://pile.eleuther.ai), a large-scale curated dataset created by [EleutherAI](https://www.eleuther.ai). ## Training procedure This model was trained for 402 billion tokens over 383,500 steps on TPU v3-256 pod. It was trained as an autoregressive language model, using cross-entropy loss to maximize the likelihood of predicting the next token correctly. ## Intended Use and Limitations GPT-J learns an inner representation of the English language that can be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating text from a prompt. ### How to use This model can be easily loaded using the `AutoModelForCausalLM` functionality: ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-j-6B") model = AutoModelForCausalLM.from_pretrained("EleutherAI/gpt-j-6B") ``` ### Limitations and Biases The core functionality of GPT-J is taking a string of text and predicting the next token. While language models are widely used for tasks other than this, there are a lot of unknowns with this work. When prompting GPT-J it is important to remember that the statistically most likely next token is often not the token that produces the most "accurate" text. Never depend upon GPT-J to produce factually accurate output. GPT-J was trained on the Pile, a dataset known to contain profanity, lewd, and otherwise abrasive language. Depending upon use case GPT-J may produce socially unacceptable text. See [Sections 5 and 6 of the Pile paper](https://arxiv.org/abs/2101.00027) for a more detailed analysis of the biases in the Pile. As with all language models, it is hard to predict in advance how GPT-J will respond to particular prompts and offensive content may occur without warning. We recommend having a human curate or filter the outputs before releasing them, both to censor undesirable content and to improve the quality of the results. ## Evaluation results <figure> | Model | Public | Training FLOPs | LAMBADA PPL โ†“ | LAMBADA Acc โ†‘ | Winogrande โ†‘ | Hellaswag โ†‘ | PIQA โ†‘ | Dataset Size (GB) | |--------------------------|-------------|----------------|--- |--- |--- |--- |--- |-------------------| | Random Chance | &check; | 0 | ~a lot | ~0% | 50% | 25% | 25% | 0 | | GPT-3 Ada&ddagger; | &cross; | ----- | 9.95 | 51.6% | 52.9% | 43.4% | 70.5% | ----- | | GPT-2 1.5B | &check; | ----- | 10.63 | 51.21% | 59.4% | 50.9% | 70.8% | 40 | | GPT-Neo 1.3B&ddagger; | &check; | 3.0e21 | 7.50 | 57.2% | 55.0% | 48.9% | 71.1% | 825 | | Megatron-2.5B&ast; | &cross; | 2.4e21 | ----- | 61.7% | ----- | ----- | ----- | 174 | | GPT-Neo 2.7B&ddagger; | &check; | 6.8e21 | 5.63 | 62.2% | 56.5% | 55.8% | 73.0% | 825 | | GPT-3 1.3B&ast;&ddagger; | &cross; | 2.4e21 | 5.44 | 63.6% | 58.7% | 54.7% | 75.1% | ~800 | | GPT-3 Babbage&ddagger; | &cross; | ----- | 5.58 | 62.4% | 59.0% | 54.5% | 75.5% | ----- | | Megatron-8.3B&ast; | &cross; | 7.8e21 | ----- | 66.5% | ----- | ----- | ----- | 174 | | GPT-3 2.7B&ast;&ddagger; | &cross; | 4.8e21 | 4.60 | 67.1% | 62.3% | 62.8% | 75.6% | ~800 | | Megatron-11B&dagger; | &check; | 1.0e22 | ----- | ----- | ----- | ----- | ----- | 161 | | **GPT-J 6B&ddagger;** | **&check;** | **1.5e22** | **3.99** | **69.7%** | **65.3%** | **66.1%** | **76.5%** | **825** | | GPT-3 6.7B&ast;&ddagger; | &cross; | 1.2e22 | 4.00 | 70.3% | 64.5% | 67.4% | 78.0% | ~800 | | GPT-3 Curie&ddagger; | &cross; | ----- | 4.00 | 69.3% | 65.6% | 68.5% | 77.9% | ----- | | GPT-3 13B&ast;&ddagger; | &cross; | 2.3e22 | 3.56 | 72.5% | 67.9% | 70.9% | 78.5% | ~800 | | GPT-3 175B&ast;&ddagger; | &cross; | 3.1e23 | 3.00 | 76.2% | 70.2% | 78.9% | 81.0% | ~800 | | GPT-3 Davinci&ddagger; | &cross; | ----- | 3.0 | 75% | 72% | 78% | 80% | ----- | <figcaption><p>Models roughly sorted by performance, or by FLOPs if not available.</p> <p><strong>&ast;</strong> Evaluation numbers reported by their respective authors. All other numbers are provided by running <a href="https://github.com/EleutherAI/lm-evaluation-harness/"><code>lm-evaluation-harness</code></a> either with released weights or with API access. Due to subtle implementation differences as well as different zero shot task framing, these might not be directly comparable. See <a href="https://blog.eleuther.ai/gpt3-model-sizes/">this blog post</a> for more details.</p> <p><strong>โ€ </strong> Megatron-11B provides no comparable metrics, and several implementations using the released weights do not reproduce the generation quality and evaluations. (see <a href="https://github.com/huggingface/transformers/pull/10301">1</a> <a href="https://github.com/pytorch/fairseq/issues/2358">2</a> <a href="https://github.com/pytorch/fairseq/issues/2719">3</a>) Thus, evaluation was not attempted.</p> <p><strong>โ€ก</strong> These models have been trained with data which contains possible test set contamination. The OpenAI GPT-3 models failed to deduplicate training data for certain test sets, while the GPT-Neo models as well as this one is trained on the Pile, which has not been deduplicated against any test sets.</p></figcaption></figure> ## Citation and Related Information ### BibTeX entry To cite this model: ```bibtex @misc{gpt-j, author = {Wang, Ben and Komatsuzaki, Aran}, title = {{GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model}}, howpublished = {\url{https://github.com/kingoflolz/mesh-transformer-jax}}, year = 2021, month = May } ``` To cite the codebase that trained this model: ```bibtex @misc{mesh-transformer-jax, author = {Wang, Ben}, title = {{Mesh-Transformer-JAX: Model-Parallel Implementation of Transformer Language Model with JAX}}, howpublished = {\url{https://github.com/kingoflolz/mesh-transformer-jax}}, year = 2021, month = May } ``` If you use this model, we would love to hear about it! Reach out on [GitHub](https://github.com/kingoflolz/mesh-transformer-jax), Discord, or shoot Ben an email. ## Acknowledgements This project would not have been possible without compute generously provided by Google through the [TPU Research Cloud](https://sites.research.google/trc/), as well as the Cloud TPU team for providing early access to the [Cloud TPU VM](https://cloud.google.com/blog/products/compute/introducing-cloud-tpu-vms) Alpha. Thanks to everyone who have helped out one way or another (listed alphabetically): - [James Bradbury](https://twitter.com/jekbradbury) for valuable assistance with debugging JAX issues. - [Stella Biderman](https://www.stellabiderman.com), [Eric Hallahan](https://twitter.com/erichallahan), [Kurumuz](https://github.com/kurumuz/), and [Finetune](https://github.com/finetuneanon/) for converting the model to be compatible with the `transformers` package. - [Leo Gao](https://twitter.com/nabla_theta) for running zero shot evaluations for the baseline models for the table. - [Laurence Golding](https://github.com/researcher2/) for adding some features to the web demo. - [Aran Komatsuzaki](https://twitter.com/arankomatsuzaki) for advice with experiment design and writing the blog posts. - [Janko Prester](https://github.com/jprester/) for creating the web demo frontend.
{"language": ["en"], "license": "apache-2.0", "tags": ["pytorch", "causal-lm"], "datasets": ["The Pile"]}
text-generation
baffo32/gpt-j-6B-ptmap
[ "transformers", "pytorch", "gptj", "text-generation", "causal-lm", "en", "arxiv:2104.09864", "arxiv:2101.00027", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2104.09864", "2101.00027" ]
[ "en" ]
TAGS #transformers #pytorch #gptj #text-generation #causal-lm #en #arxiv-2104.09864 #arxiv-2101.00027 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
GPT-J 6B ======== Model Description ----------------- GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters. **\*** Each layer consists of one feedforward block and one self attention block. **โ€ ** Although the embedding matrix has a size of 50400, only 50257 entries are used by the GPT-2 tokenizer. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. Training data ------------- GPT-J 6B was trained on the Pile, a large-scale curated dataset created by EleutherAI. Training procedure ------------------ This model was trained for 402 billion tokens over 383,500 steps on TPU v3-256 pod. It was trained as an autoregressive language model, using cross-entropy loss to maximize the likelihood of predicting the next token correctly. Intended Use and Limitations ---------------------------- GPT-J learns an inner representation of the English language that can be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating text from a prompt. ### How to use This model can be easily loaded using the 'AutoModelForCausalLM' functionality: ### Limitations and Biases The core functionality of GPT-J is taking a string of text and predicting the next token. While language models are widely used for tasks other than this, there are a lot of unknowns with this work. When prompting GPT-J it is important to remember that the statistically most likely next token is often not the token that produces the most "accurate" text. Never depend upon GPT-J to produce factually accurate output. GPT-J was trained on the Pile, a dataset known to contain profanity, lewd, and otherwise abrasive language. Depending upon use case GPT-J may produce socially unacceptable text. See Sections 5 and 6 of the Pile paper for a more detailed analysis of the biases in the Pile. As with all language models, it is hard to predict in advance how GPT-J will respond to particular prompts and offensive content may occur without warning. We recommend having a human curate or filter the outputs before releasing them, both to censor undesirable content and to improve the quality of the results. Evaluation results ------------------ Models roughly sorted by performance, or by FLOPs if not available. **\*** Evaluation numbers reported by their respective authors. All other numbers are provided by running [for more details.](URL either with released weights or with API access. Due to subtle implementation differences as well as different zero shot task framing, these might not be directly comparable. See <a href=) **โ€ ** Megatron-11B provides no comparable metrics, and several implementations using the released weights do not reproduce the generation quality and evaluations. (see <a href="URL <a href="URL <a href="URL Thus, evaluation was not attempted.</p> **โ€ก** These models have been trained with data which contains possible test set contamination. The OpenAI GPT-3 models failed to deduplicate training data for certain test sets, while the GPT-Neo models as well as this one is trained on the Pile, which has not been deduplicated against any test sets. and Related Information ### BibTeX entry To cite this model: To cite the codebase that trained this model: If you use this model, we would love to hear about it! Reach out on GitHub, Discord, or shoot Ben an email. Acknowledgements ---------------- This project would not have been possible without compute generously provided by Google through the TPU Research Cloud, as well as the Cloud TPU team for providing early access to the Cloud TPU VM Alpha. Thanks to everyone who have helped out one way or another (listed alphabetically): * James Bradbury for valuable assistance with debugging JAX issues. * Stella Biderman, Eric Hallahan, Kurumuz, and Finetune for converting the model to be compatible with the 'transformers' package. * Leo Gao for running zero shot evaluations for the baseline models for the table. * Laurence Golding for adding some features to the web demo. * Aran Komatsuzaki for advice with experiment design and writing the blog posts. * Janko Prester for creating the web demo frontend.
[ "### How to use\n\n\nThis model can be easily loaded using the 'AutoModelForCausalLM' functionality:", "### Limitations and Biases\n\n\nThe core functionality of GPT-J is taking a string of text and predicting the next token. While language models are widely used for tasks other than this, there are a lot of unknowns with this work. When prompting GPT-J it is important to remember that the statistically most likely next token is often not the token that produces the most \"accurate\" text. Never depend upon GPT-J to produce factually accurate output.\n\n\nGPT-J was trained on the Pile, a dataset known to contain profanity, lewd, and otherwise abrasive language. Depending upon use case GPT-J may produce socially unacceptable text. See Sections 5 and 6 of the Pile paper for a more detailed analysis of the biases in the Pile.\n\n\nAs with all language models, it is hard to predict in advance how GPT-J will respond to particular prompts and offensive content may occur without warning. We recommend having a human curate or filter the outputs before releasing them, both to censor undesirable content and to improve the quality of the results.\n\n\nEvaluation results\n------------------\n\n\n\n\nModels roughly sorted by performance, or by FLOPs if not available.\n\n\n**\\*** Evaluation numbers reported by their respective authors. All other numbers are provided by\nrunning [for more\ndetails.](URL either with released\nweights or with API access. Due to subtle implementation differences as well as different zero shot task framing, these\nmight not be directly comparable. See <a href=)\n\n\n**โ€ ** Megatron-11B provides no comparable metrics, and several implementations using the released weights do not\nreproduce the generation quality and evaluations. (see <a href=\"URL\n<a href=\"URL <a href=\"URL\nThus, evaluation was not attempted.</p>\n**โ€ก** These models have been trained with data which contains possible test set contamination. The OpenAI GPT-3 models\nfailed to deduplicate training data for certain test sets, while the GPT-Neo models as well as this one is\ntrained on the Pile, which has not been deduplicated against any test sets.\n\n\n\n\nand Related Information", "### BibTeX entry\n\n\nTo cite this model:\n\n\nTo cite the codebase that trained this model:\n\n\nIf you use this model, we would love to hear about it! Reach out on GitHub, Discord, or shoot Ben an email.\n\n\nAcknowledgements\n----------------\n\n\nThis project would not have been possible without compute generously provided by Google through the\nTPU Research Cloud, as well as the Cloud TPU team for providing early access to the Cloud TPU VM Alpha.\n\n\nThanks to everyone who have helped out one way or another (listed alphabetically):\n\n\n* James Bradbury for valuable assistance with debugging JAX issues.\n* Stella Biderman, Eric Hallahan, Kurumuz, and Finetune for converting the model to be compatible with the 'transformers' package.\n* Leo Gao for running zero shot evaluations for the baseline models for the table.\n* Laurence Golding for adding some features to the web demo.\n* Aran Komatsuzaki for advice with experiment design and writing the blog posts.\n* Janko Prester for creating the web demo frontend." ]
[ "TAGS\n#transformers #pytorch #gptj #text-generation #causal-lm #en #arxiv-2104.09864 #arxiv-2101.00027 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### How to use\n\n\nThis model can be easily loaded using the 'AutoModelForCausalLM' functionality:", "### Limitations and Biases\n\n\nThe core functionality of GPT-J is taking a string of text and predicting the next token. While language models are widely used for tasks other than this, there are a lot of unknowns with this work. When prompting GPT-J it is important to remember that the statistically most likely next token is often not the token that produces the most \"accurate\" text. Never depend upon GPT-J to produce factually accurate output.\n\n\nGPT-J was trained on the Pile, a dataset known to contain profanity, lewd, and otherwise abrasive language. Depending upon use case GPT-J may produce socially unacceptable text. See Sections 5 and 6 of the Pile paper for a more detailed analysis of the biases in the Pile.\n\n\nAs with all language models, it is hard to predict in advance how GPT-J will respond to particular prompts and offensive content may occur without warning. We recommend having a human curate or filter the outputs before releasing them, both to censor undesirable content and to improve the quality of the results.\n\n\nEvaluation results\n------------------\n\n\n\n\nModels roughly sorted by performance, or by FLOPs if not available.\n\n\n**\\*** Evaluation numbers reported by their respective authors. All other numbers are provided by\nrunning [for more\ndetails.](URL either with released\nweights or with API access. Due to subtle implementation differences as well as different zero shot task framing, these\nmight not be directly comparable. See <a href=)\n\n\n**โ€ ** Megatron-11B provides no comparable metrics, and several implementations using the released weights do not\nreproduce the generation quality and evaluations. (see <a href=\"URL\n<a href=\"URL <a href=\"URL\nThus, evaluation was not attempted.</p>\n**โ€ก** These models have been trained with data which contains possible test set contamination. The OpenAI GPT-3 models\nfailed to deduplicate training data for certain test sets, while the GPT-Neo models as well as this one is\ntrained on the Pile, which has not been deduplicated against any test sets.\n\n\n\n\nand Related Information", "### BibTeX entry\n\n\nTo cite this model:\n\n\nTo cite the codebase that trained this model:\n\n\nIf you use this model, we would love to hear about it! Reach out on GitHub, Discord, or shoot Ben an email.\n\n\nAcknowledgements\n----------------\n\n\nThis project would not have been possible without compute generously provided by Google through the\nTPU Research Cloud, as well as the Cloud TPU team for providing early access to the Cloud TPU VM Alpha.\n\n\nThanks to everyone who have helped out one way or another (listed alphabetically):\n\n\n* James Bradbury for valuable assistance with debugging JAX issues.\n* Stella Biderman, Eric Hallahan, Kurumuz, and Finetune for converting the model to be compatible with the 'transformers' package.\n* Leo Gao for running zero shot evaluations for the baseline models for the table.\n* Laurence Golding for adding some features to the web demo.\n* Aran Komatsuzaki for advice with experiment design and writing the blog posts.\n* Janko Prester for creating the web demo frontend." ]
[ 70, 26, 493, 228 ]
[ "passage: TAGS\n#transformers #pytorch #gptj #text-generation #causal-lm #en #arxiv-2104.09864 #arxiv-2101.00027 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nThis model can be easily loaded using the 'AutoModelForCausalLM' functionality:" ]
[ -0.11835873872041702, 0.10682331770658493, -0.002850749995559454, 0.0836348608136177, 0.13448700308799744, 0.028397077694535255, 0.15984046459197998, 0.10086438059806824, -0.011290029622614384, -0.08193451166152954, 0.20230264961719513, 0.28165391087532043, -0.011612282134592533, 0.09340041875839233, -0.012473366223275661, -0.19713658094406128, 0.05119870603084564, 0.04272010922431946, -0.03725983202457428, 0.09632375836372375, 0.13327711820602417, -0.011265184730291367, 0.10425441712141037, 0.03767145797610283, -0.12012605369091034, 0.010106545872986317, 0.021287601441144943, -0.0874677300453186, 0.10230248421430588, 0.08665767312049866, -0.02956571616232395, 0.043498191982507706, 0.05928967893123627, -0.10378353297710419, 0.009349149651825428, -0.013429262675344944, -0.037673868238925934, 0.0837622731924057, 0.054932624101638794, 0.012084399349987507, 0.13566942512989044, 0.08808135241270065, -0.0510878711938858, 0.019306516274809837, -0.05980720371007919, -0.10482432693243027, -0.05062931030988693, 0.058460310101509094, 0.09009743481874466, 0.12306709587574005, 0.04757358133792877, 0.13410405814647675, -0.0833914652466774, 0.08680199831724167, 0.13640230894088745, -0.3141954839229584, -0.003997656051069498, 0.11832042783498764, -0.003575138049200177, -0.05605553463101387, 0.05621238052845001, 0.031382862478494644, 0.043645571917295456, 0.007847163826227188, 0.03406132012605667, -0.08968830108642578, -0.11450250446796417, 0.02947785146534443, -0.08225185424089432, -0.061980415135622025, 0.30745092034339905, -0.029569676145911217, -0.00229020812548697, -0.015236634761095047, -0.0552775003015995, 0.09682611376047134, -0.03586442396044731, 0.032630644738674164, -0.002951181260868907, 0.09009866416454315, 0.08637088537216187, -0.07008059322834015, -0.11605420708656311, -0.04641002044081688, -0.12012705951929092, 0.08143306523561478, 0.010695001110434532, 0.08760976791381836, -0.11561694741249084, 0.10924571752548218, -0.05945725366473198, -0.0904734879732132, 0.002234636340290308, -0.06876397877931595, 0.11888940632343292, 0.025633802637457848, -0.0296354740858078, -0.019222524017095566, 0.08651165664196014, 0.15797367691993713, 0.07556028664112091, -0.007997654378414154, 0.06364557147026062, 0.0888042002916336, -0.01286181341856718, 0.0948450118303299, -0.07841207832098007, -0.049545157700777054, 0.1274704486131668, -0.04393985867500305, 0.06365209072828293, -0.01164208073168993, -0.18908560276031494, -0.08246148377656937, -0.032088976353406906, 0.1014685332775116, 0.045517295598983765, 0.07589909434318542, -0.019949225708842278, -0.05217723548412323, 0.17971749603748322, -0.06954476982355118, -0.003114549443125725, -0.03785672038793564, -0.021435683593153954, 0.03244129195809364, 0.12556830048561096, 0.017678873613476753, -0.07500908523797989, 0.01685035042464733, -0.047459717839956284, -0.004671772010624409, -0.06864766031503677, -0.03374146297574043, 0.025388021022081375, -0.007510778959840536, 0.05986914783716202, -0.1559217870235443, -0.28077271580696106, 0.01802590675652027, 0.09678936749696732, -0.01044422946870327, -0.14900153875350952, -0.007508326321840286, 0.011196098290383816, -0.0005853918264620006, -0.07010101526975632, 0.09489157795906067, -0.060935962945222855, 0.041502561420202255, -0.04118747636675835, 0.05819639563560486, -0.16054292023181915, 0.06134941056370735, -0.10865064710378647, -0.004395339637994766, -0.06387557834386826, 0.023234419524669647, -0.03184925392270088, 0.11828028410673141, -0.05580819025635719, -0.021432414650917053, 0.03460480272769928, 0.010349159128963947, 0.049788907170295715, 0.13563328981399536, -0.0817376971244812, -0.05694979056715965, 0.10571709275245667, -0.09541473537683487, -0.20292095839977264, 0.03995085507631302, 0.014953340403735638, 0.14357629418373108, 0.07002599537372589, 0.05749589577317238, 0.07195813208818436, -0.03199270740151405, 0.10966325551271439, 0.10059921443462372, -0.11673727631568909, -0.1809401661157608, 0.023227455094456673, -0.013881220482289791, -0.20181411504745483, 0.0835387110710144, -0.046522434800863266, 0.1570766717195511, -0.002333533251658082, -0.05855325609445572, -0.07635900378227234, -0.11534475535154343, -0.050044622272253036, 0.006329620257019997, 0.06626953184604645, 0.015472369268536568, 0.010291563346982002, -0.03197691962122917, 0.08744675666093826, 0.000043915806600125507, -0.0020959689281880856, -0.047529447823762894, 0.05123915523290634, -0.09155258536338806, 0.07183276861906052, -0.13838623464107513, 0.01595098339021206, -0.01334976777434349, 0.013705804012715816, 0.046685799956321716, 0.014006557874381542, 0.05436720699071884, 0.0077024661004543304, 0.0111601697281003, -0.0022086615208536386, 0.163035050034523, 0.022295966744422913, -0.0648278146982193, -0.1155245378613472, 0.025833118706941605, -0.022698314860463142, 0.08287644386291504, -0.04961337894201279, 0.02744140475988388, -0.11361584067344666, 0.025957413017749786, -0.07488079369068146, 0.09124189615249634, 0.02439967356622219, -0.01574082113802433, -0.047163695096969604, -0.016758739948272705, 0.09950657188892365, 0.05258376523852348, -0.04831130430102348, 0.19953054189682007, -0.09059736877679825, 0.21359895169734955, 0.15641409158706665, -0.13210533559322357, 0.018574977293610573, -0.015762021765112877, -0.0399041585624218, -0.017312534153461456, 0.007690963335335255, 0.04366595670580864, 0.05681294575333595, 0.011373949237167835, 0.15829071402549744, -0.05593090131878853, 0.0014234264381229877, -0.0036131320521235466, -0.09101514518260956, 0.03607670217752457, 0.06650220602750778, 0.2123294323682785, -0.13489913940429688, 0.12130627781152725, 0.2484579235315323, -0.06043547764420509, 0.049009598791599274, 0.0027243089862167835, -0.024316415190696716, 0.006885624025017023, -0.05409153550863266, -0.0200066976249218, 0.02430313266813755, -0.11545982211828232, 0.016731806099414825, 0.10298565030097961, -0.03627552092075348, 0.08884487301111221, -0.11466098576784134, -0.037384748458862305, 0.020630503073334694, -0.0178759153932333, 0.005668579135090113, 0.08324778825044632, -0.03078623116016388, 0.09670968353748322, -0.059507403522729874, -0.1030869111418724, 0.09694425761699677, 0.04626161977648735, -0.11283320188522339, 0.159587562084198, -0.12092845141887665, -0.23976749181747437, -0.14214445650577545, -0.10718092322349548, -0.10188575834035873, 0.020424582064151764, 0.09745137393474579, -0.06579265743494034, -0.0947728380560875, -0.019124969840049744, -0.08501385897397995, 0.056319475173950195, -0.01754310540854931, 0.010758140124380589, 0.01809859462082386, 0.009857462719082832, -0.12291926145553589, -0.05812862887978554, 0.014903758652508259, -0.03443613275885582, 0.08775810897350311, -0.09653301537036896, 0.08126289397478104, 0.15094853937625885, 0.03254055231809616, 0.05032864212989807, 0.04178256168961525, 0.2093724012374878, -0.017856251448392868, 0.02065540850162506, 0.2526920437812805, -0.004977460950613022, 0.05720337852835655, 0.11793810874223709, 0.020048337057232857, -0.06976567953824997, 0.006676976103335619, -0.0768413320183754, -0.07622537761926651, -0.20154701173305511, -0.13741549849510193, -0.08946818113327026, 0.02741018682718277, 0.08551978319883347, 0.07542990148067474, 0.14345061779022217, 0.14909572899341583, 0.020657625049352646, 0.08234953135251999, -0.0014710775576531887, 0.0817771703004837, 0.20636120438575745, -0.018925456330180168, 0.12330307811498642, -0.09944140166044235, -0.0676172599196434, 0.12154202163219452, 0.022042563185095787, 0.12749886512756348, 0.04940702021121979, 0.010724437423050404, 0.06198258325457573, 0.12133312225341797, 0.030521290376782417, 0.17402487993240356, -0.02283463254570961, -0.031975895166397095, -0.03552378714084625, -0.043876633048057556, -0.062474340200424194, 0.07161210477352142, -0.09431403130292892, -0.11632230132818222, -0.031330354511737823, 0.02380961738526821, 0.02005777135491371, 0.12436690181493759, 0.02210994064807892, -0.247830331325531, 0.000569407653529197, 0.02855983003973961, 0.028108572587370872, -0.06435885280370712, 0.030588803812861443, -0.11063193529844284, -0.12692253291606903, 0.054222822189331055, -0.0052392794750630856, 0.14365626871585846, -0.010262765921652317, 0.03318396955728531, -0.03145672008395195, 0.0033542511519044638, 0.0323827788233757, 0.17205417156219482, -0.31065890192985535, 0.19913417100906372, 0.00691926758736372, -0.009138659574091434, -0.12348304688930511, 0.03239567577838898, 0.08884251862764359, 0.20299845933914185, 0.10024507343769073, 0.009542117826640606, -0.04246971011161804, -0.05535571277141571, -0.07558462768793106, 0.06408477574586868, -0.0720568522810936, 0.007336003240197897, -0.04207362234592438, -0.04470352828502655, -0.032048389315605164, 0.0050720893777906895, 0.030695589259266853, -0.11696676164865494, -0.15783727169036865, 0.06556129455566406, 0.12504759430885315, 0.013680360279977322, -0.028739523142576218, -0.012501337565481663, -0.07349526137113571, 0.23065412044525146, 0.09453345090150833, -0.11466635763645172, -0.09196051955223083, -0.06916172802448273, 0.07038997858762741, -0.08995500206947327, 0.03511793911457062, -0.06459477543830872, 0.012525217607617378, -0.039455629885196686, -0.18003584444522858, 0.0758054181933403, -0.14668630063533783, -0.007165929768234491, -0.009118587709963322, 0.046278610825538635, -0.07488738745450974, -0.010796661488711834, 0.076527900993824, -0.008374682627618313, -0.12489169090986252, -0.1418396532535553, -0.06075673922896385, 0.07001575082540512, 0.029091689735651016, -0.04462655633687973, -0.0822787657380104, -0.012349444441497326, 0.020159399136900902, -0.059341296553611755, 0.19430553913116455, 0.15278850495815277, -0.06487846374511719, 0.12926052510738373, 0.18112295866012573, -0.07050243020057678, -0.2807278037071228, -0.17191287875175476, -0.07466720789670944, -0.04100330173969269, -0.0368259996175766, -0.14640718698501587, 0.15224817395210266, 0.033833060413599014, -0.06752205640077591, 0.0832279622554779, -0.21519148349761963, -0.11041643470525742, 0.26542168855667114, 0.020405102521181107, 0.27844318747520447, -0.09329714626073837, -0.04938202351331711, -0.11739158630371094, -0.13531538844108582, 0.13611994683742523, -0.07300425320863724, 0.06027743220329285, -0.04600420594215393, 0.09538818150758743, 0.011015405878424644, -0.07362135499715805, 0.10541506111621857, 0.014041684567928314, -0.009698761627078056, -0.12456508725881577, 0.014416780322790146, 0.0426209419965744, -0.022423304617404938, 0.188200905919075, -0.11607349663972855, 0.05236593633890152, -0.123783178627491, -0.04232111945748329, -0.07239893823862076, 0.05035290867090225, 0.024010010063648224, -0.0777537152171135, -0.00351125281304121, -0.05689099803566933, -0.00426630862057209, 0.0004845744406338781, 0.03133980929851532, 0.02029711753129959, -0.02236969955265522, 0.16510894894599915, 0.08644789457321167, -0.2392311990261078, -0.012032603845000267, -0.05035020038485527, -0.053963351994752884, 0.07539908587932587, -0.1566033512353897, 0.03974944353103638, 0.06743326783180237, -0.05252176150679588, 0.080450639128685, 0.04580400139093399, 0.02705203928053379, -0.05129208415746689, 0.09457193315029144, -0.1957828849554062, 0.10711243003606796, -0.05738357454538345, 0.09276438504457474, 0.06441765278577805, 0.013017704710364342, 0.11743146926164627, -0.021473534405231476, -0.04014861583709717, 0.024639200419187546, 0.029611272737383842, -0.06623830646276474, 0.09086316078901291, 0.06850374490022659, -0.011529327370226383, -0.1417563557624817, 0.06279643625020981, 0.007636889815330505, -0.044160228222608566, -0.023751232773065567, 0.08374578505754471, -0.14406614005565643, -0.1302536576986313, -0.0026062042452394962, 0.13903599977493286, -0.24718278646469116, -0.11024577915668488, -0.04342033714056015, -0.07454583048820496, 0.05965345725417137, -0.04785352572798729, 0.08787690103054047, 0.010528718121349812, -0.06072181835770607, -0.07980933785438538, -0.048180919140577316, 0.02314828895032406, 0.048316992819309235, 0.030480505898594856, -0.07226230949163437, -0.029440665617585182, 0.013904592953622341, 0.1329117864370346, -0.04828597232699394, -0.00807257555425167, -0.08271477371454239, 0.029766034334897995, -0.16761066019535065, -0.023535309359431267, -0.12756094336509705, -0.0150812529027462, 0.025612760335206985, -0.03350968286395073, -0.018890822306275368, 0.024868983775377274, -0.11208935081958771, -0.024407921358942986, -0.01844775304198265, 0.03590868413448334, -0.08358070254325867, -0.06496430188417435, 0.04953592270612717, -0.013043106533586979, 0.08951489627361298, 0.09472812712192535, -0.05593834072351456, 0.02708231285214424, -0.12185696512460709, -0.07212118059396744, 0.06572095304727554, 0.05886046960949898, 0.04261736571788788, -0.061406105756759644, 0.026333265006542206, 0.11311865597963333, -0.012504043988883495, -0.006429455243051052, 0.050850048661231995, -0.10111143440008163, 0.02226421982049942, 0.012216605246067047, -0.09868647903203964, -0.015458312816917896, -0.03997707739472389, 0.04747634381055832, 0.051812540739774704, 0.1357453465461731, -0.057187799364328384, 0.04793868958950043, -0.06617368012666702, 0.023314323276281357, -0.0361735001206398, -0.11126599460840225, -0.13999749720096588, -0.09292580932378769, -0.009191961027681828, -0.004646399524062872, 0.22873593866825104, 0.033503320068120956, 0.008869375102221966, -0.005420810077339411, 0.1427110880613327, 0.11693001538515091, -0.04112594574689865, 0.18099132180213928, 0.04600955918431282, 0.061402853578329086, -0.08316465467214584, 0.10501789301633835, 0.04501139000058174, -0.06293465942144394, 0.028455136343836784, -0.06004321575164795, 0.028391685336828232, 0.08122903853654861, 0.00943363830447197, -0.033641085028648376, -0.062150925397872925, -0.11473415046930313, -0.0131229218095541, 0.06800209730863571, -0.014901057817041874, 0.09957042336463928, 0.12549613416194916, -0.0021193521097302437, 0.010260011069476604, 0.01859048195183277, -0.034479912370443344, -0.17148441076278687, -0.17841364443302155, -0.08173108845949173, -0.1414320468902588, -0.017958423122763634, -0.07251622527837753, 0.0029927457217127085, -0.011325445957481861, 0.029194923117756844, -0.04649269953370094, 0.031235070899128914, -0.009949462488293648, -0.03468402475118637, -0.021171187981963158, -0.045998912304639816, -0.007825812324881554, -0.044333502650260925, 0.03130536526441574, -0.09602458029985428, 0.001322574564255774, 0.012322711758315563, 0.04412032291293144, 0.012963107787072659, 0.042330071330070496, -0.025860754773020744, -0.03152494505047798, -0.09026238322257996, 0.026062877848744392, 0.01571333222091198, 0.13054946064949036, -0.004396302159875631, 0.036535926163196564, 0.05172736942768097, 0.1272173523902893, -0.05428940802812576, -0.16425970196723938, -0.08541563898324966, 0.22104328870773315, -0.0070645688101649284, 0.03734443336725235, -0.002494765678420663, 0.05903562158346176, -0.028064550831913948, 0.35606080293655396, 0.26519984006881714, -0.10708588361740112, 0.009933151304721832, -0.020694999024271965, 0.012573782354593277, 0.026057712733745575, 0.1263994574546814, 0.10741213709115982, 0.19151313602924347, -0.07764769345521927, -0.0027661507483571768, -0.09065965563058853, -0.00531063973903656, -0.11743412911891937, 0.06011126562952995, 0.015140819363296032, -0.09023639559745789, 0.03230658918619156, 0.09655725210905075, -0.09822908043861389, 0.2171778678894043, -0.09816792607307434, -0.020814169198274612, -0.07511711865663528, 0.007831359282135963, 0.05871791020035744, 0.009503476321697235, 0.04433993995189667, -0.06351764500141144, 0.012472722679376602, 0.14563710987567902, -0.03400716558098793, -0.23424436151981354, -0.06246502324938774, 0.08082758635282516, 0.016200978308916092, 0.1618577241897583, 0.01157606951892376, 0.07869075238704681, 0.05903826653957367, 0.03491009399294853, -0.15434134006500244, 0.057226285338401794, -0.012193533591926098, -0.07758592069149017, 0.07802433520555496, -0.1229323223233223, -0.018228359520435333, -0.11137751489877701, -0.009394247084856033, -0.10622391849756241, 0.03639669716358185, 0.06685368716716766, -0.02426503598690033, -0.031965214759111404, 0.04192904010415077, -0.07224708795547485, 0.07812051475048065, 0.037035197019577026, -0.03214224800467491, -0.061660222709178925, -0.09519797563552856, 0.05621429532766342, 0.009681515395641327, -0.16568028926849365, -0.06144437938928604, -0.02056412398815155, -0.022068476304411888, 0.012720883823931217, 0.02561783604323864, -0.09361434727907181, -0.039356593042612076, -0.08416986465454102, -0.046693433076143265, -0.10256350785493851, 0.05820683017373085, 0.0708848312497139, 0.024358833208680153, 0.012446936219930649, -0.026178710162639618, 0.006826853379607201, 0.019999846816062927, -0.10606800764799118, -0.1372128129005432 ]
null
null
transformers
# GPT-2 Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large Pretrained model on English language using a causal language modeling (CLM) objective. It was introduced in [this paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) and first released at [this page](https://openai.com/blog/better-language-models/). Disclaimer: The team releasing GPT-2 also wrote a [model card](https://github.com/openai/gpt-2/blob/master/model_card.md) for their model. Content from this model card has been written by the Hugging Face team to complete the information they provided and give specific examples of bias. ## Model description GPT-2 is a transformers model pretrained on a very large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was trained to guess the next word in sentences. More precisely, inputs are sequences of continuous text of a certain length and the targets are the same sequence, shifted one token (word or piece of word) to the right. The model uses internally a mask-mechanism to make sure the predictions for the token `i` only uses the inputs from `1` to `i` but not the future tokens. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating texts from a prompt. ## Intended uses & limitations You can use the raw model for text generation or fine-tune it to a downstream task. See the [model hub](https://huggingface.co/models?filter=gpt2) to look for fine-tuned versions on a task that interests you. ### How to use You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we set a seed for reproducibility: ```python >>> from transformers import pipeline, set_seed >>> generator = pipeline('text-generation', model='gpt2') >>> set_seed(42) >>> generator("Hello, I'm a language model,", max_length=30, num_return_sequences=5) [{'generated_text': "Hello, I'm a language model, a language for thinking, a language for expressing thoughts."}, {'generated_text': "Hello, I'm a language model, a compiler, a compiler library, I just want to know how I build this kind of stuff. I don"}, {'generated_text': "Hello, I'm a language model, and also have more than a few of your own, but I understand that they're going to need some help"}, {'generated_text': "Hello, I'm a language model, a system model. I want to know my language so that it might be more interesting, more user-friendly"}, {'generated_text': 'Hello, I\'m a language model, not a language model"\n\nThe concept of "no-tricks" comes in handy later with new'}] ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import GPT2Tokenizer, GPT2Model tokenizer = GPT2Tokenizer.from_pretrained('gpt2') model = GPT2Model.from_pretrained('gpt2') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import GPT2Tokenizer, TFGPT2Model tokenizer = GPT2Tokenizer.from_pretrained('gpt2') model = TFGPT2Model.from_pretrained('gpt2') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ### Limitations and bias The training data used for this model has not been released as a dataset one can browse. We know it contains a lot of unfiltered content from the internet, which is far from neutral. As the openAI team themselves point out in their [model card](https://github.com/openai/gpt-2/blob/master/model_card.md#out-of-scope-use-cases): > Because large-scale language models like GPT-2 do not distinguish fact from fiction, we donโ€™t support use-cases > that require the generated text to be true. > > Additionally, language models like GPT-2 reflect the biases inherent to the systems they were trained on, so we do > not recommend that they be deployed into systems that interact with humans > unless the deployers first carry out a > study of biases relevant to the intended use-case. We found no statistically significant difference in gender, race, > and religious bias probes between 774M and 1.5B, implying all versions of GPT-2 should be approached with similar > levels of caution around use cases that are sensitive to biases around human attributes. Here's an example of how the model can have biased predictions: ```python >>> from transformers import pipeline, set_seed >>> generator = pipeline('text-generation', model='gpt2') >>> set_seed(42) >>> generator("The White man worked as a", max_length=10, num_return_sequences=5) [{'generated_text': 'The White man worked as a mannequin for'}, {'generated_text': 'The White man worked as a maniser of the'}, {'generated_text': 'The White man worked as a bus conductor by day'}, {'generated_text': 'The White man worked as a plumber at the'}, {'generated_text': 'The White man worked as a journalist. He had'}] >>> set_seed(42) >>> generator("The Black man worked as a", max_length=10, num_return_sequences=5) [{'generated_text': 'The Black man worked as a man at a restaurant'}, {'generated_text': 'The Black man worked as a car salesman in a'}, {'generated_text': 'The Black man worked as a police sergeant at the'}, {'generated_text': 'The Black man worked as a man-eating monster'}, {'generated_text': 'The Black man worked as a slave, and was'}] ``` This bias will also affect all fine-tuned versions of this model. ## Training data The OpenAI team wanted to train this model on a corpus as large as possible. To build it, they scraped all the web pages from outbound links on Reddit which received at least 3 karma. Note that all Wikipedia pages were removed from this dataset, so the model was not trained on any part of Wikipedia. The resulting dataset (called WebText) weights 40GB of texts but has not been publicly released. You can find a list of the top 1,000 domains present in WebText [here](https://github.com/openai/gpt-2/blob/master/domains.txt). ## Training procedure ### Preprocessing The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 1024 consecutive tokens. The larger model was trained on 256 cloud TPU v3 cores. The training duration was not disclosed, nor were the exact details of training. ## Evaluation results The model achieves the following results without any fine-tuning (zero-shot): | Dataset | LAMBADA | LAMBADA | CBT-CN | CBT-NE | WikiText2 | PTB | enwiki8 | text8 | WikiText103 | 1BW | |:--------:|:-------:|:-------:|:------:|:------:|:---------:|:------:|:-------:|:------:|:-----------:|:-----:| | (metric) | (PPL) | (ACC) | (ACC) | (ACC) | (PPL) | (PPL) | (BPB) | (BPC) | (PPL) | (PPL) | | | 35.13 | 45.99 | 87.65 | 83.4 | 29.41 | 65.85 | 1.16 | 1,17 | 37.50 | 75.20 | ### BibTeX entry and citation info ```bibtex @article{radford2019language, title={Language Models are Unsupervised Multitask Learners}, author={Radford, Alec and Wu, Jeff and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya}, year={2019} } ``` <a href="https://huggingface.co/exbert/?model=gpt2"> <img width="300px" src="https://cdn-media.huggingface.co/exbert/button.png"> </a>
{"language": "en", "license": "mit", "tags": ["exbert"]}
text-generation
baffo32/gpt2-ptmap
[ "transformers", "pytorch", "tf", "jax", "tflite", "rust", "gpt2", "text-generation", "exbert", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tf #jax #tflite #rust #gpt2 #text-generation #exbert #en #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
GPT-2 ===== Test the whole generation capabilities here: URL Pretrained model on English language using a causal language modeling (CLM) objective. It was introduced in this paper and first released at this page. Disclaimer: The team releasing GPT-2 also wrote a model card for their model. Content from this model card has been written by the Hugging Face team to complete the information they provided and give specific examples of bias. Model description ----------------- GPT-2 is a transformers model pretrained on a very large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was trained to guess the next word in sentences. More precisely, inputs are sequences of continuous text of a certain length and the targets are the same sequence, shifted one token (word or piece of word) to the right. The model uses internally a mask-mechanism to make sure the predictions for the token 'i' only uses the inputs from '1' to 'i' but not the future tokens. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating texts from a prompt. Intended uses & limitations --------------------------- You can use the raw model for text generation or fine-tune it to a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. ### How to use You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we set a seed for reproducibility: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ### Limitations and bias The training data used for this model has not been released as a dataset one can browse. We know it contains a lot of unfiltered content from the internet, which is far from neutral. As the openAI team themselves point out in their model card: > > Because large-scale language models like GPT-2 do not distinguish fact from fiction, we donโ€™t support use-cases > that require the generated text to be true. > > > Additionally, language models like GPT-2 reflect the biases inherent to the systems they were trained on, so we do > not recommend that they be deployed into systems that interact with humans > unless the deployers first carry out a > study of biases relevant to the intended use-case. We found no statistically significant difference in gender, race, > and religious bias probes between 774M and 1.5B, implying all versions of GPT-2 should be approached with similar > levels of caution around use cases that are sensitive to biases around human attributes. > > > Here's an example of how the model can have biased predictions: This bias will also affect all fine-tuned versions of this model. Training data ------------- The OpenAI team wanted to train this model on a corpus as large as possible. To build it, they scraped all the web pages from outbound links on Reddit which received at least 3 karma. Note that all Wikipedia pages were removed from this dataset, so the model was not trained on any part of Wikipedia. The resulting dataset (called WebText) weights 40GB of texts but has not been publicly released. You can find a list of the top 1,000 domains present in WebText here. Training procedure ------------------ ### Preprocessing The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 1024 consecutive tokens. The larger model was trained on 256 cloud TPU v3 cores. The training duration was not disclosed, nor were the exact details of training. Evaluation results ------------------ The model achieves the following results without any fine-tuning (zero-shot): ### BibTeX entry and citation info <a href="URL <img width="300px" src="URL
[ "### How to use\n\n\nYou can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we\nset a seed for reproducibility:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:", "### Limitations and bias\n\n\nThe training data used for this model has not been released as a dataset one can browse. We know it contains a lot of\nunfiltered content from the internet, which is far from neutral. As the openAI team themselves point out in their\nmodel card:\n\n\n\n> \n> Because large-scale language models like GPT-2 do not distinguish fact from fiction, we donโ€™t support use-cases\n> that require the generated text to be true.\n> \n> \n> Additionally, language models like GPT-2 reflect the biases inherent to the systems they were trained on, so we do\n> not recommend that they be deployed into systems that interact with humans > unless the deployers first carry out a\n> study of biases relevant to the intended use-case. We found no statistically significant difference in gender, race,\n> and religious bias probes between 774M and 1.5B, implying all versions of GPT-2 should be approached with similar\n> levels of caution around use cases that are sensitive to biases around human attributes.\n> \n> \n> \n\n\nHere's an example of how the model can have biased predictions:\n\n\nThis bias will also affect all fine-tuned versions of this model.\n\n\nTraining data\n-------------\n\n\nThe OpenAI team wanted to train this model on a corpus as large as possible. To build it, they scraped all the web\npages from outbound links on Reddit which received at least 3 karma. Note that all Wikipedia pages were removed from\nthis dataset, so the model was not trained on any part of Wikipedia. The resulting dataset (called WebText) weights\n40GB of texts but has not been publicly released. You can find a list of the top 1,000 domains present in WebText\nhere.\n\n\nTraining procedure\n------------------", "### Preprocessing\n\n\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a\nvocabulary size of 50,257. The inputs are sequences of 1024 consecutive tokens.\n\n\nThe larger model was trained on 256 cloud TPU v3 cores. The training duration was not disclosed, nor were the exact\ndetails of training.\n\n\nEvaluation results\n------------------\n\n\nThe model achieves the following results without any fine-tuning (zero-shot):", "### BibTeX entry and citation info\n\n\n<a href=\"URL\n<img width=\"300px\" src=\"URL" ]
[ "TAGS\n#transformers #pytorch #tf #jax #tflite #rust #gpt2 #text-generation #exbert #en #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### How to use\n\n\nYou can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we\nset a seed for reproducibility:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:", "### Limitations and bias\n\n\nThe training data used for this model has not been released as a dataset one can browse. We know it contains a lot of\nunfiltered content from the internet, which is far from neutral. As the openAI team themselves point out in their\nmodel card:\n\n\n\n> \n> Because large-scale language models like GPT-2 do not distinguish fact from fiction, we donโ€™t support use-cases\n> that require the generated text to be true.\n> \n> \n> Additionally, language models like GPT-2 reflect the biases inherent to the systems they were trained on, so we do\n> not recommend that they be deployed into systems that interact with humans > unless the deployers first carry out a\n> study of biases relevant to the intended use-case. We found no statistically significant difference in gender, race,\n> and religious bias probes between 774M and 1.5B, implying all versions of GPT-2 should be approached with similar\n> levels of caution around use cases that are sensitive to biases around human attributes.\n> \n> \n> \n\n\nHere's an example of how the model can have biased predictions:\n\n\nThis bias will also affect all fine-tuned versions of this model.\n\n\nTraining data\n-------------\n\n\nThe OpenAI team wanted to train this model on a corpus as large as possible. To build it, they scraped all the web\npages from outbound links on Reddit which received at least 3 karma. Note that all Wikipedia pages were removed from\nthis dataset, so the model was not trained on any part of Wikipedia. The resulting dataset (called WebText) weights\n40GB of texts but has not been publicly released. You can find a list of the top 1,000 domains present in WebText\nhere.\n\n\nTraining procedure\n------------------", "### Preprocessing\n\n\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a\nvocabulary size of 50,257. The inputs are sequences of 1024 consecutive tokens.\n\n\nThe larger model was trained on 256 cloud TPU v3 cores. The training duration was not disclosed, nor were the exact\ndetails of training.\n\n\nEvaluation results\n------------------\n\n\nThe model achieves the following results without any fine-tuning (zero-shot):", "### BibTeX entry and citation info\n\n\n<a href=\"URL\n<img width=\"300px\" src=\"URL" ]
[ 69, 66, 393, 118, 30 ]
[ "passage: TAGS\n#transformers #pytorch #tf #jax #tflite #rust #gpt2 #text-generation #exbert #en #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### How to use\n\n\nYou can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we\nset a seed for reproducibility:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:" ]
[ -0.0456831268966198, -0.0020836980547755957, -0.0048690494149923325, 0.06944583356380463, 0.13297227025032043, -0.01378298457711935, 0.09417504072189331, 0.08420883119106293, 0.032513681799173355, -0.01342508289963007, 0.2069116234779358, 0.24624812602996826, 0.005484237801283598, 0.117665596306324, -0.035996753722429276, -0.25306788086891174, -0.028685953468084335, 0.048219941556453705, 0.06303611397743225, 0.12564444541931152, 0.08235784620046616, -0.09188155829906464, 0.09324703365564346, 0.021142445504665375, -0.1461545079946518, 0.0028205339331179857, 0.010892256163060665, -0.07841305434703827, 0.10377965122461319, 0.09575073421001434, 0.045376501977443695, 0.08390199393033981, 0.007298699580132961, -0.06534374505281448, 0.04631540924310684, 0.0005943984142504632, -0.03911111503839493, 0.1082080528140068, 0.1401715725660324, -0.12046298384666443, 0.15644411742687225, 0.08068433403968811, -0.08920484781265259, 0.022785767912864685, -0.07815085351467133, 0.0355394072830677, 0.0027269653510302305, 0.023495374247431755, 0.0339370034635067, 0.0665377750992775, -0.024569984525442123, 0.1060929074883461, -0.010402644984424114, 0.11702927201986313, 0.2383047193288803, -0.3413955867290497, -0.0466357059776783, 0.04574940726161003, 0.06998057663440704, 0.01988806575536728, -0.00015099786105565727, 0.04446359723806381, 0.054123666137456894, 0.07936245948076248, 0.17213383316993713, -0.07788325846195221, -0.09458430856466293, 0.016213709488511086, -0.13920027017593384, 0.0027348969597369432, 0.1757872849702835, -0.08438461273908615, -0.011979866772890091, -0.08075214177370071, -0.07834737747907639, -0.025390030816197395, -0.031016772612929344, -0.02815181203186512, -0.04006156697869301, 0.018239961937069893, 0.024273859336972237, -0.11495133489370346, -0.13697753846645355, -0.07478612661361694, -0.11132621765136719, 0.11008098721504211, 0.05811765417456627, 0.05987798050045967, -0.17179885506629944, 0.14502644538879395, 0.06504427641630173, -0.0677405521273613, 0.039209067821502686, -0.10261370241641998, 0.03426093980669975, 0.001175508601590991, -0.08108968287706375, -0.15337815880775452, 0.083906389772892, 0.042365316301584244, -0.024144010618329048, -0.006638612598180771, 0.02389487624168396, 0.10675767809152603, 0.024987874552607536, 0.052922192960977554, -0.10795071721076965, 0.05647118762135506, 0.0437047965824604, -0.087941475212574, -0.06450887024402618, -0.01944383978843689, -0.13493210077285767, -0.045880477875471115, 0.04312686249613762, 0.06790716201066971, 0.025043414905667305, 0.12264122813940048, -0.00389449717476964, -0.09768545627593994, -0.0014054563362151384, -0.07493077963590622, -0.04617961868643761, 0.01755566895008087, -0.07349890470504761, 0.0889374166727066, 0.036735471338033676, -0.03018946945667267, -0.15152095258235931, 0.020619947463274002, -0.06406684964895248, -0.012436684221029282, -0.03970739617943764, -0.10329464823007584, -0.031051944941282272, -0.12147524952888489, 0.009607194922864437, -0.15925103425979614, -0.1971355676651001, -0.0042435890063643456, 0.036798231303691864, -0.04782811924815178, -0.05929749831557274, -0.0743601992726326, -0.02020043507218361, -0.002610812894999981, -0.045028991997241974, 0.10787308216094971, -0.03172915801405907, 0.07020283490419388, -0.06447883695363998, 0.1007995754480362, -0.057892121374607086, 0.06957828998565674, -0.08112705498933792, 0.003643104573711753, -0.000055765343859093264, 0.08642736822366714, 0.025899827480316162, 0.028552642092108727, -0.07245984673500061, -0.05345781520009041, -0.04435994476079941, 0.023081177845597267, 0.021839793771505356, 0.15224656462669373, -0.11741302162408829, -0.08235717564821243, 0.20789408683776855, -0.04265611246228218, -0.11797244846820831, 0.11388792097568512, -0.017464734613895416, 0.0853251963853836, 0.09033653885126114, 0.1328192502260208, 0.08191143721342087, -0.019140878692269325, 0.10829140990972519, 0.11907119303941727, -0.18736813962459564, -0.0972563847899437, 0.013575982302427292, 0.054818447679281235, -0.13138478994369507, 0.07589446008205414, 0.016016853973269463, 0.11604583263397217, -0.07493093609809875, -0.028745684772729874, -0.009440835565328598, -0.05556444078683853, 0.04062214121222496, -0.010202277451753616, 0.12637314200401306, 0.035323865711688995, -0.006162612698972225, -0.056812774389982224, 0.00949816033244133, -0.09286873787641525, 0.019160712137818336, -0.04123730957508087, 0.15207518637180328, -0.006620782893151045, 0.07362411171197891, -0.20519395172595978, 0.01357374619692564, 0.029571544378995895, 0.1325981467962265, 0.023093681782484055, 0.0279728714376688, 0.041439905762672424, 0.04087622091174126, 0.004126064945012331, 0.007335467264056206, 0.2308238446712494, -0.06354545056819916, -0.1034991592168808, -0.10998792946338654, -0.0045299893245100975, -0.03667454794049263, -0.04806048423051834, -0.0884242132306099, 0.00477959867566824, -0.09608867764472961, 0.02720772661268711, -0.036486055701971054, 0.045931000262498856, 0.02064695581793785, 0.0019679043907672167, -0.08815635740756989, 0.012499306350946426, 0.11532462388277054, -0.004422848112881184, -0.1067519262433052, 0.2390732318162918, -0.04941219463944435, 0.10393433272838593, 0.12476363778114319, -0.17043249309062958, -0.06002190709114075, -0.0679902657866478, -0.08029552549123764, -0.01146634016185999, 0.003498801263049245, 0.036577723920345306, 0.2175074815750122, 0.011618208140134811, 0.17231635749340057, -0.08923827111721039, -0.0584070160984993, 0.02067919261753559, -0.07897113263607025, 0.03890563175082207, 0.1385650336742401, 0.07175716757774353, -0.12806105613708496, 0.0791974663734436, 0.051455575972795486, -0.010172804817557335, 0.13269372284412384, 0.0016483378130942583, -0.05148793011903763, 0.005722422618418932, 0.06578224152326584, 0.0016963911475613713, -0.06140865385532379, -0.07703067362308502, -0.04550910368561745, 0.07061934471130371, -0.06059941649436951, 0.07306241244077682, -0.09322769194841385, -0.0066977934911847115, -0.013515549711883068, -0.030623888596892357, 0.08585434406995773, 0.07619258016347885, -0.010885207913815975, 0.13090172410011292, -0.05144412815570831, -0.09668955206871033, 0.058893971145153046, 0.009041323326528072, -0.10380907356739044, 0.22713127732276917, -0.04678576812148094, -0.19849109649658203, -0.10240226984024048, -0.109563447535038, -0.07847656309604645, 0.04849284514784813, 0.04313516616821289, -0.011541961692273617, 0.0009203694644384086, -0.04283217340707779, 0.04980560764670372, -0.04971916228532791, 0.04305637627840042, -0.12575365602970123, -0.02562539465725422, -0.04622254520654678, -0.05600597709417343, -0.046898044645786285, -0.059245478361845016, 0.001412603072822094, 0.1702166199684143, -0.07217616587877274, 0.05949665978550911, 0.140850231051445, -0.0019360658479854465, 0.03285559266805649, -0.03410643711686134, 0.1836210936307907, -0.010261772200465202, 0.06073526293039322, 0.14775465428829193, 0.003611618187278509, 0.0739225447177887, 0.025095444172620773, 0.02738155424594879, -0.04986904561519623, 0.03321247920393944, -0.050459545105695724, -0.05565237998962402, -0.23200632631778717, -0.09337162226438522, -0.08119180798530579, 0.0755140632390976, 0.10052672028541565, 0.048808496445417404, 0.07577016204595566, 0.13436327874660492, -0.00573083758354187, 0.11294741928577423, -0.00007827091030776501, 0.061128851026296616, 0.11514865607023239, -0.005417404696345329, 0.13865140080451965, -0.005062085576355457, -0.11635946482419968, 0.07291775196790695, -0.026774156838655472, 0.22462284564971924, -0.01860327646136284, 0.1231202632188797, 0.07508938759565353, 0.15577727556228638, 0.050469446927309036, 0.16463008522987366, -0.009373290464282036, 0.001333358115516603, -0.046346426010131836, -0.05795679986476898, 0.026798995211720467, 0.038324978202581406, -0.024973643943667412, -0.06264057755470276, -0.03787874057888985, -0.011312952265143394, 0.03402271121740341, 0.16824829578399658, 0.005716128274798393, -0.2882797122001648, 0.007545210421085358, -0.007737204432487488, -0.02007516659796238, -0.03635015711188316, 0.07430419325828552, 0.06326752156019211, -0.11249973624944687, -0.01194804161787033, -0.08094332367181778, 0.1371796578168869, -0.011527551338076591, 0.05540389567613602, -0.03313710168004036, 0.06017614156007767, 0.005124798975884914, 0.12885482609272003, -0.27811262011528015, 0.16001206636428833, -0.015329012647271156, -0.016066256910562515, -0.11191316694021225, -0.026658356189727783, 0.04129677638411522, 0.023206496611237526, 0.17189852893352509, -0.038155827671289444, 0.04572726786136627, -0.06463796645402908, -0.05003022775053978, 0.01594647951424122, 0.059589844197034836, -0.030642949044704437, 0.06552620977163315, 0.01394387986510992, 0.002886950969696045, -0.01966080442070961, 0.02597254514694214, -0.04018811881542206, -0.0865449607372284, 0.026862600818276405, 0.06608191877603531, -0.03528560698032379, 0.040477484464645386, -0.05689864233136177, -0.07893751561641693, 0.129970520734787, -0.0790712758898735, -0.14396008849143982, -0.0924217626452446, 0.014847959391772747, 0.0475994236767292, -0.0882510095834732, -0.024865107610821724, -0.031368598341941833, 0.05096667632460594, -0.050871968269348145, -0.2273024618625641, 0.061300043016672134, -0.09668277204036713, -0.04078951105475426, -0.03432559594511986, 0.09905215352773666, 0.04030878096818924, 0.011756147257983685, 0.006421845406293869, -0.042264219373464584, -0.11553540825843811, -0.13209626078605652, 0.0046353451907634735, -0.017504874616861343, 0.03362683206796646, -0.008996400982141495, -0.07473248243331909, -0.04594697058200836, -0.09583176672458649, 0.03781967610120773, 0.22527575492858887, 0.09270285815000534, -0.09097114950418472, 0.15821032226085663, 0.10084749758243561, -0.11201371252536774, -0.27118876576423645, -0.1650424748659134, 0.013700854033231735, 0.01604587584733963, 0.015574579127132893, -0.18217733502388, 0.06921512633562088, -0.022099537774920464, -0.003365658689290285, 0.13317689299583435, -0.34254270792007446, -0.10973021388053894, 0.18010371923446655, 0.08139258623123169, 0.26287809014320374, -0.12047545611858368, -0.09060553461313248, -0.0021299775689840317, 0.030885521322488785, 0.14647407829761505, -0.049202810972929, 0.10052470862865448, -0.08338654786348343, 0.08808038383722305, 0.03131834790110588, -0.03845607861876488, 0.0342286042869091, 0.027220729738473892, 0.0326840877532959, -0.04782560095191002, 0.03462212532758713, 0.033916693180799484, 0.016824722290039062, 0.11686971038579941, -0.10646171867847443, 0.04708231985569, -0.07949371635913849, -0.07948555797338486, -0.07998527586460114, 0.10760718584060669, 0.030695360153913498, -0.06677734106779099, 0.0061567691154778, -0.05248511955142021, 0.03410567343235016, -0.008964658714830875, -0.020958878099918365, -0.030371498316526413, 0.09227115660905838, 0.22023114562034607, 0.07335418462753296, -0.04141905531287193, -0.08346735686063766, -0.04890907555818558, -0.020970594137907028, 0.08294768631458282, -0.07366227358579636, -0.039143193513154984, 0.13061703741550446, -0.002964962739497423, 0.0802612379193306, 0.08855989575386047, -0.04435764625668526, 0.009120243601500988, 0.045845877379179, -0.24376530945301056, -0.05887395143508911, -0.1057206243276596, -0.1042395830154419, 0.03506390005350113, 0.04710477218031883, 0.10414154082536697, -0.09375108033418655, -0.022646013647317886, -0.003264717059209943, -0.008556176908314228, -0.018142905086278915, 0.12676846981048584, 0.04655159264802933, 0.022899171337485313, -0.11221873015165329, 0.0424782894551754, 0.030628008767962456, -0.03701385110616684, 0.00224868580698967, 0.08378314971923828, -0.1137007400393486, -0.11890754103660583, 0.013258850201964378, 0.09695611149072647, -0.12724798917770386, -0.03320857882499695, -0.04778788238763809, -0.09810950607061386, 0.08142630755901337, 0.10285355150699615, 0.0645652711391449, 0.07510603964328766, -0.06290672719478607, -0.0774494856595993, -0.03022407367825508, 0.06173786148428917, 0.11068619787693024, -0.01806432195007801, -0.08970984816551208, 0.09885019063949585, -0.018824046477675438, 0.14050760865211487, -0.09925612062215805, -0.048462703824043274, -0.11512022465467453, 0.019704220816493034, -0.15419329702854156, 0.05104639008641243, -0.09820358455181122, -0.04886648803949356, -0.006982353050261736, 0.04573216661810875, -0.025795714929699898, 0.012461633421480656, -0.08330702781677246, 0.0023908009752631187, -0.009960049763321877, 0.06297287344932556, -0.09021957218647003, -0.05006841942667961, -0.0137832872569561, -0.008283909410238266, 0.09124518185853958, 0.1440320461988449, -0.12475418299436569, 0.03827706351876259, -0.10605664551258087, -0.013861499726772308, 0.04611070081591606, 0.03386634215712547, 0.013781219720840454, 0.01135291624814272, 0.01371826697140932, 0.03253006935119629, -0.021431969478726387, 0.017616553232073784, -0.010878585278987885, -0.14217698574066162, 0.029029883444309235, 0.005502503830939531, -0.001726388232782483, -0.09874045103788376, 0.009318316355347633, -0.009561270475387573, 0.07907412201166153, 0.10312789678573608, -0.1147085502743721, 0.0966879203915596, -0.08973241597414017, 0.047140732407569885, 0.0058069853112101555, -0.13136546313762665, -0.06980104744434357, -0.08021856844425201, 0.03076818771660328, -0.02887852117419243, 0.16253986954689026, 0.04701385274529457, -0.018450353294610977, -0.03670230507850647, 0.06277105212211609, 0.0670345276594162, -0.00234338641166687, 0.12000090628862381, 0.005603102967143059, -0.03288581967353821, -0.12549549341201782, 0.11998265981674194, 0.047927506268024445, 0.06722826510667801, 0.10509632527828217, -0.01674959622323513, -0.023465503007173538, 0.09720873832702637, -0.004975504241883755, 0.0237262025475502, -0.052950792014598846, -0.10227292031049728, -0.02944936975836754, 0.09110019356012344, 0.004987859167158604, 0.09906454384326935, 0.09791936725378036, -0.023001892492175102, 0.04012066498398781, 0.013021807186305523, -0.06744184345006943, -0.12560652196407318, -0.2765940725803375, -0.03069492056965828, -0.06051483377814293, -0.04343630000948906, -0.10594556480646133, 0.0007112225284799933, -0.02387372776865959, 0.0678243562579155, -0.05819407105445862, 0.07042060047388077, 0.10046105086803436, -0.12820479273796082, 0.11749764531850815, -0.03533663600683212, 0.024479657411575317, -0.058834098279476166, 0.03678968921303749, -0.024724194779992104, -0.036805685609579086, -0.024737728759646416, 0.03367774188518524, 0.03608562424778938, 0.05002037063241005, -0.08925503492355347, -0.08552315086126328, -0.04227042570710182, 0.048211466521024704, -0.006675781216472387, 0.05551886558532715, 0.013350731693208218, -0.0656304582953453, -0.021801026538014412, 0.16946737468242645, -0.029962312430143356, -0.0013383377809077501, -0.04210369288921356, 0.32730087637901306, 0.04351601004600525, 0.014916772954165936, 0.0250420942902565, 0.021530257537961006, -0.06880972534418106, 0.3894796669483185, 0.28233522176742554, -0.08176198601722717, -0.005063761956989765, 0.019804293289780617, 0.00912355538457632, 0.08791843801736832, 0.1495981216430664, 0.03479839488863945, 0.279326856136322, -0.08245712518692017, 0.030450670048594475, -0.07373534142971039, 0.01746196672320366, -0.09379752725362778, 0.03730134665966034, 0.07973787188529968, -0.03826570510864258, -0.005214294884353876, 0.08927322924137115, -0.2975234389305115, 0.011342499405145645, -0.08797866106033325, -0.010327097028493881, -0.04864978790283203, -0.043684426695108414, -0.0011181202717125416, 0.05995131656527519, 0.10754657536745071, -0.06463504582643509, 0.046955227851867676, 0.06925027817487717, -0.01630507968366146, -0.17502687871456146, 0.1024831160902977, 0.04338546097278595, -0.09258635342121124, 0.019810084253549576, -0.016656119376420975, 0.10581497848033905, 0.04085309058427811, 0.033736828714609146, -0.08565524220466614, 0.03343180567026138, -0.021870028227567673, -0.018013227730989456, 0.10291843861341476, 0.008792676962912083, 0.05036616325378418, -0.14014102518558502, -0.0016227398300543427, -0.02373296022415161, 0.02317824959754944, -0.04129055142402649, 0.028422510251402855, -0.04036346822977066, 0.058684878051280975, -0.016791049391031265, 0.05783556401729584, 0.1200762689113617, -0.03784840181469917, -0.02794194035232067, -0.062297478318214417, -0.030364761129021645, -0.017941482365131378, -0.06606432050466537, -0.0489937886595726, -0.11817573755979538, -0.0580105260014534, 0.004469950217753649, -0.0034399277064949274, -0.2646687626838684, 0.0317513607442379, -0.07918622344732285, -0.020082488656044006, -0.08619546890258789, 0.03165614977478981, 0.07599347084760666, 0.03099784255027771, 0.0009400471462868154, 0.09289012104272842, -0.02727808803319931, 0.03551235795021057, -0.10712114721536636, -0.1016051173210144 ]
null
null
transformers
# ByT5 - Base ByT5 is a tokenizer-free version of [Google's T5](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html) and generally follows the architecture of [MT5](https://huggingface.co/google/mt5-base). ByT5 was only pre-trained on [mC4](https://www.tensorflow.org/datasets/catalog/c4#c4multilingual) excluding any supervised training with an average span-mask of 20 UTF-8 characters. Therefore, this model has to be fine-tuned before it is useable on a downstream task. ByT5 works especially well on noisy text data,*e.g.*, `google/byt5-base` significantly outperforms [mt5-base](https://huggingface.co/google/mt5-base) on [TweetQA](https://arxiv.org/abs/1907.06292). Paper: [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) Authors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel* ## Example Inference ByT5 works on raw UTF-8 bytes and can be used without a tokenizer: ```python from transformers import T5ForConditionalGeneration import torch model = T5ForConditionalGeneration.from_pretrained('google/byt5-base') input_ids = torch.tensor([list("Life is like a box of chocolates.".encode("utf-8"))]) + 3 # add 3 for special tokens labels = torch.tensor([list("La vie est comme une boรฎte de chocolat.".encode("utf-8"))]) + 3 # add 3 for special tokens loss = model(input_ids, labels=labels).loss # forward pass ``` For batched inference & training it is however recommended using a tokenizer class for padding: ```python from transformers import T5ForConditionalGeneration, AutoTokenizer model = T5ForConditionalGeneration.from_pretrained('google/byt5-base') tokenizer = AutoTokenizer.from_pretrained('google/byt5-base') model_inputs = tokenizer(["Life is like a box of chocolates.", "Today is Monday."], padding="longest", return_tensors="pt") labels = tokenizer(["La vie est comme une boรฎte de chocolat.", "Aujourd'hui c'est lundi."], padding="longest", return_tensors="pt").input_ids loss = model(**model_inputs, labels=labels).loss # forward pass ``` ## Abstract Most widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units. Encoding text as a sequence of tokens requires a tokenizer, which is typically created as an independent artifact from the model. Token-free models that instead operate directly on raw text (bytes or characters) have many benefits: they can process text in any language out of the box, they are more robust to noise, and they minimize technical debt by removing complex and error-prone text preprocessing pipelines. Since byte or character sequences are longer than token sequences, past work on token-free models has often introduced new model architectures designed to amortize the cost of operating directly on raw text. In this paper, we show that a standard Transformer architecture can be used with minimal modifications to process byte sequences. We carefully characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and show that byte-level models are competitive with their token-level counterparts. We also demonstrate that byte-level models are significantly more robust to noise and perform better on tasks that are sensitive to spelling and pronunciation. As part of our contribution, we release a new set of pre-trained byte-level Transformer models based on the T5 architecture, as well as all code and data used in our experiments. ![model image](https://raw.githubusercontent.com/patrickvonplaten/scientific_images/master/ByT5.png)
{"language": "multilingual", "license": "apache-2.0", "datasets": ["mc4"]}
text2text-generation
baffo32/pyc2py_alpha2
[ "transformers", "jax", "t5", "text2text-generation", "multilingual", "dataset:mc4", "arxiv:1907.06292", "arxiv:2105.13626", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1907.06292", "2105.13626" ]
[ "multilingual" ]
TAGS #transformers #jax #t5 #text2text-generation #multilingual #dataset-mc4 #arxiv-1907.06292 #arxiv-2105.13626 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# ByT5 - Base ByT5 is a tokenizer-free version of Google's T5 and generally follows the architecture of MT5. ByT5 was only pre-trained on mC4 excluding any supervised training with an average span-mask of 20 UTF-8 characters. Therefore, this model has to be fine-tuned before it is useable on a downstream task. ByT5 works especially well on noisy text data,*e.g.*, 'google/byt5-base' significantly outperforms mt5-base on TweetQA. Paper: ByT5: Towards a token-free future with pre-trained byte-to-byte models Authors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel* ## Example Inference ByT5 works on raw UTF-8 bytes and can be used without a tokenizer: For batched inference & training it is however recommended using a tokenizer class for padding: ## Abstract Most widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units. Encoding text as a sequence of tokens requires a tokenizer, which is typically created as an independent artifact from the model. Token-free models that instead operate directly on raw text (bytes or characters) have many benefits: they can process text in any language out of the box, they are more robust to noise, and they minimize technical debt by removing complex and error-prone text preprocessing pipelines. Since byte or character sequences are longer than token sequences, past work on token-free models has often introduced new model architectures designed to amortize the cost of operating directly on raw text. In this paper, we show that a standard Transformer architecture can be used with minimal modifications to process byte sequences. We carefully characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and show that byte-level models are competitive with their token-level counterparts. We also demonstrate that byte-level models are significantly more robust to noise and perform better on tasks that are sensitive to spelling and pronunciation. As part of our contribution, we release a new set of pre-trained byte-level Transformer models based on the T5 architecture, as well as all code and data used in our experiments. !model image
[ "# ByT5 - Base\n\nByT5 is a tokenizer-free version of Google's T5 and generally follows the architecture of MT5.\n\nByT5 was only pre-trained on mC4 excluding any supervised training with an average span-mask of 20 UTF-8 characters. Therefore, this model has to be fine-tuned before it is useable on a downstream task.\n\nByT5 works especially well on noisy text data,*e.g.*, 'google/byt5-base' significantly outperforms mt5-base on TweetQA.\n\nPaper: ByT5: Towards a token-free future with pre-trained byte-to-byte models\n\nAuthors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel*", "## Example Inference\n\nByT5 works on raw UTF-8 bytes and can be used without a tokenizer:\n\n\n\nFor batched inference & training it is however recommended using a tokenizer class for padding:", "## Abstract\n\nMost widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units. Encoding text as a sequence of tokens requires a tokenizer, which is typically created as an independent artifact from the model. Token-free models that instead operate directly on raw text (bytes or characters) have many benefits: they can process text in any language out of the box, they are more robust to noise, and they minimize technical debt by removing complex and error-prone text preprocessing pipelines. Since byte or character sequences are longer than token sequences, past work on token-free models has often introduced new model architectures designed to amortize the cost of operating directly on raw text. In this paper, we show that a standard Transformer architecture can be used with minimal modifications to process byte sequences. We carefully characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and show that byte-level models are competitive with their token-level counterparts. We also demonstrate that byte-level models are significantly more robust to noise and perform better on tasks that are sensitive to spelling and pronunciation. As part of our contribution, we release a new set of pre-trained byte-level Transformer models based on the T5 architecture, as well as all code and data used in our experiments.\n\n!model image" ]
[ "TAGS\n#transformers #jax #t5 #text2text-generation #multilingual #dataset-mc4 #arxiv-1907.06292 #arxiv-2105.13626 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# ByT5 - Base\n\nByT5 is a tokenizer-free version of Google's T5 and generally follows the architecture of MT5.\n\nByT5 was only pre-trained on mC4 excluding any supervised training with an average span-mask of 20 UTF-8 characters. Therefore, this model has to be fine-tuned before it is useable on a downstream task.\n\nByT5 works especially well on noisy text data,*e.g.*, 'google/byt5-base' significantly outperforms mt5-base on TweetQA.\n\nPaper: ByT5: Towards a token-free future with pre-trained byte-to-byte models\n\nAuthors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel*", "## Example Inference\n\nByT5 works on raw UTF-8 bytes and can be used without a tokenizer:\n\n\n\nFor batched inference & training it is however recommended using a tokenizer class for padding:", "## Abstract\n\nMost widely-used pre-trained language models operate on sequences of tokens corresponding to word or subword units. Encoding text as a sequence of tokens requires a tokenizer, which is typically created as an independent artifact from the model. Token-free models that instead operate directly on raw text (bytes or characters) have many benefits: they can process text in any language out of the box, they are more robust to noise, and they minimize technical debt by removing complex and error-prone text preprocessing pipelines. Since byte or character sequences are longer than token sequences, past work on token-free models has often introduced new model architectures designed to amortize the cost of operating directly on raw text. In this paper, we show that a standard Transformer architecture can be used with minimal modifications to process byte sequences. We carefully characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and show that byte-level models are competitive with their token-level counterparts. We also demonstrate that byte-level models are significantly more robust to noise and perform better on tasks that are sensitive to spelling and pronunciation. As part of our contribution, we release a new set of pre-trained byte-level Transformer models based on the T5 architecture, as well as all code and data used in our experiments.\n\n!model image" ]
[ 84, 199, 48, 326 ]
[ "passage: TAGS\n#transformers #jax #t5 #text2text-generation #multilingual #dataset-mc4 #arxiv-1907.06292 #arxiv-2105.13626 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# ByT5 - Base\n\nByT5 is a tokenizer-free version of Google's T5 and generally follows the architecture of MT5.\n\nByT5 was only pre-trained on mC4 excluding any supervised training with an average span-mask of 20 UTF-8 characters. Therefore, this model has to be fine-tuned before it is useable on a downstream task.\n\nByT5 works especially well on noisy text data,*e.g.*, 'google/byt5-base' significantly outperforms mt5-base on TweetQA.\n\nPaper: ByT5: Towards a token-free future with pre-trained byte-to-byte models\n\nAuthors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel*## Example Inference\n\nByT5 works on raw UTF-8 bytes and can be used without a tokenizer:\n\n\n\nFor batched inference & training it is however recommended using a tokenizer class for padding:" ]
[ -0.048178281635046005, 0.012601710855960846, -0.0038410562556236982, 0.05272539332509041, 0.06025351956486702, -0.0038117929361760616, 0.14522899687290192, 0.10922418534755707, -0.10096712410449982, 0.06171932816505432, 0.12820644676685333, 0.02685486152768135, 0.06622815132141113, 0.1683131456375122, 0.006431092973798513, -0.13330243527889252, -0.014924544841051102, -0.05786117911338806, 0.029471902176737785, 0.0860004797577858, 0.05761326104402542, -0.031563956290483475, 0.08437112718820572, -0.02502763271331787, -0.09480558335781097, -0.025848040357232094, 0.015019471757113934, -0.07079367339611053, 0.01996487006545067, 0.0838906392455101, -0.016508400440216064, 0.009442112408578396, -0.03549190238118172, -0.14270450174808502, 0.03470294922590256, 0.09969230741262436, -0.00742028234526515, 0.04478050395846367, 0.064146026968956, -0.03671465814113617, 0.11960510909557343, 0.030472800135612488, -0.06345780938863754, 0.045665353536605835, -0.09594961255788803, 0.02064911462366581, -0.05129434913396835, 0.04969623684883118, 0.05391738936305046, 0.07139209657907486, -0.017869189381599426, 0.10906162858009338, 0.03789011016488075, 0.10847297310829163, 0.06069416180253029, -0.17637136578559875, -0.08480940759181976, 0.022537633776664734, -0.06268584728240967, 0.12061504274606705, -0.05202217772603035, 0.029176387935876846, 0.06515878438949585, -0.0037776962853968143, 0.04708109050989151, -0.02176174707710743, -0.0035507429856806993, -0.015126683749258518, -0.11841028928756714, -0.02008136548101902, 0.17027820646762848, -0.0042566098272800446, -0.0938447117805481, -0.07846850156784058, -0.07234273850917816, 0.06732290238142014, -0.011299588717520237, -0.008489977568387985, -0.010161282494664192, 0.028152301907539368, 0.03774700686335564, -0.12555792927742004, -0.08368438482284546, -0.03238776698708534, -0.041317034512758255, 0.061643458902835846, 0.031389910727739334, 0.03842645511031151, -0.03639543056488037, 0.04447611793875694, -0.09145557135343552, -0.0943308025598526, -0.030626168474555016, -0.09295827895402908, -0.02661866880953312, -0.005249040201306343, 0.0159149132668972, -0.17631861567497253, 0.021536054089665413, 0.034334175288677216, 0.0847846195101738, 0.07885001599788666, -0.04446854442358017, 0.013384287245571613, -0.008710666559636593, -0.00983946118503809, -0.03665083274245262, -0.04347508028149605, 0.10945753753185272, 0.03468236327171326, 0.030477438122034073, -0.055446989834308624, -0.02009974978864193, -0.0315515398979187, -0.05638980120420456, 0.013248682022094727, 0.05530184134840965, 0.04709852114319801, -0.011476622894406319, -0.04612753912806511, 0.1175524964928627, -0.07274065911769867, -0.06536310911178589, 0.00389922340400517, -0.04215897619724274, 0.015733947977423668, 0.03447500616312027, -0.02193155512213707, -0.09817305952310562, -0.03204846382141113, -0.04742101579904556, -0.03623730689287186, -0.07342056185007095, -0.06781835854053497, 0.005280700046569109, 0.04389017075300217, -0.0389065183699131, -0.16749975085258484, -0.18067461252212524, 0.011123488657176495, 0.06674383580684662, -0.04370721057057381, 0.03133396431803703, -0.040179841220378876, -0.09015892446041107, 0.010801908560097218, -0.00871640257537365, 0.11154996603727341, -0.046945709735155106, 0.0411529466509819, 0.07134997099637985, 0.04737868160009384, 0.041800376027822495, -0.00716043496504426, -0.04679718241095543, -0.014707343652844429, -0.1718614399433136, 0.1146065816283226, -0.03992890566587448, 0.007570737972855568, -0.11989482492208481, 0.00281911832280457, -0.11653527617454529, -0.01960928738117218, 0.05189935490489006, 0.14955472946166992, -0.1395435333251953, -0.007586812134832144, 0.1415695995092392, -0.11332186311483383, -0.10254988074302673, 0.15369684994220734, -0.00130421610083431, 0.06346885114908218, 0.04863062500953674, 0.11002811044454575, 0.12232697755098343, -0.11116773635149002, -0.045098692178726196, 0.016081012785434723, -0.019426563754677773, -0.04012083262205124, 0.027543839067220688, 0.019756581634283066, -0.08348530530929565, 0.021002748981118202, -0.00601094588637352, 0.043086521327495575, -0.03154023736715317, -0.05704977363348007, -0.013865934684872627, -0.07474107295274734, -0.03219562768936157, -0.030372299253940582, 0.04664674401283264, -0.022731641307473183, -0.06407414376735687, -0.07196250557899475, 0.05872887000441551, -0.0501149483025074, 0.02644607611000538, -0.05217830836772919, 0.04963556304574013, -0.14296212792396545, 0.053974930197000504, -0.10695262998342514, -0.031244885176420212, 0.031008215621113777, -0.04913543909788132, 0.10868159681558609, -0.0253896564245224, 0.042748503386974335, 0.07755923271179199, -0.029301997274160385, -0.00860985554754734, 0.08012223243713379, -0.04966738075017929, -0.08046707510948181, -0.12422586977481842, 0.0599212646484375, -0.018703505396842957, 0.0061977277509868145, -0.16497230529785156, 0.03415653482079506, -0.009392859414219856, 0.006610882934182882, 0.024161208420991898, -0.023634742945432663, 0.07849976420402527, 0.028299417346715927, -0.02694401890039444, -0.03356429934501648, 0.014979492872953415, 0.027175866067409515, -0.04540230706334114, 0.12905658781528473, -0.13179145753383636, -0.0718509703874588, 0.09247774630784988, 0.03745920956134796, -0.07449101656675339, 0.08914440870285034, 0.003910026978701353, -0.06376229971647263, -0.026192720979452133, 0.016321605071425438, 0.15320904552936554, -0.0061207180842757225, 0.1329038441181183, -0.09083975851535797, -0.04770185425877571, 0.0721496120095253, -0.06975443661212921, -0.03372828662395477, 0.07355393469333649, -0.039787717163562775, -0.08093281090259552, 0.031688570976257324, 0.06708361208438873, -0.05676793307065964, 0.14708185195922852, 0.02048444375395775, -0.06705115735530853, -0.022178275510668755, 0.10887512564659119, 0.05847148597240448, 0.040704380720853806, -0.04615972563624382, -0.0013931316789239645, 0.017960168421268463, 0.04315422102808952, 0.0394100546836853, -0.05559849739074707, 0.05971946194767952, -0.02946360595524311, -0.07145895063877106, -0.02525772899389267, 0.03383933752775192, -0.06447001546621323, 0.05033868923783302, -0.00073193822754547, 0.07999274879693985, 0.015703126788139343, -0.029767297208309174, -0.1175428107380867, 0.15014870464801788, -0.0655607283115387, -0.10871274024248123, -0.12038230895996094, -0.0721512883901596, -0.116078220307827, -0.021865535527467728, 0.04476195201277733, -0.007483253255486488, -0.055212173610925674, -0.15298011898994446, 0.003459867322817445, -0.06413985043764114, -0.021187851205468178, -0.07459114491939545, 0.0045705754309892654, 0.03393610939383507, -0.1060665026307106, 0.043321795761585236, 0.013730977661907673, -0.026741288602352142, 0.07130853086709976, -0.026258079335093498, 0.06064232811331749, 0.04695514962077141, -0.02271087095141411, 0.004409542307257652, -0.045837629586458206, 0.20661930739879608, -0.031205113977193832, 0.059339988976716995, 0.19380933046340942, 0.05246781185269356, 0.08064455538988113, 0.060547079890966415, -0.02415398508310318, -0.06432930380105972, 0.057394061237573624, -0.0030413332860916853, -0.057636044919490814, -0.22194835543632507, 0.004193420987576246, -0.0929059162735939, 0.040463246405124664, 0.016076404601335526, 0.04141121730208397, -0.037579409778118134, 0.05936174839735031, -0.0373399555683136, 0.07384054362773895, 0.014157447032630444, 0.09924940019845963, 0.010001828894019127, 0.028135918080806732, 0.0692431703209877, -0.07435614615678787, 0.03192345052957535, 0.13562218844890594, 0.0214828047901392, 0.19109362363815308, -0.09197250753641129, 0.1424948126077652, 0.056723494082689285, 0.14571329951286316, 0.10372009873390198, 0.006172072142362595, -0.07846154272556305, 0.055637046694755554, -0.027728358283638954, -0.08302129805088043, -0.03608527034521103, 0.05463030934333801, -0.08656727522611618, -0.014751204289495945, 0.050847142934799194, 0.05782925337553024, 0.059292178601026535, 0.22688983380794525, -0.00220266985706985, -0.1635632961988449, -0.07752281427383423, 0.009379602037370205, -0.034387603402137756, -0.07153889536857605, 0.012317776679992676, 0.08102071285247803, -0.0463128425180912, 0.025743374601006508, -0.03174750506877899, 0.08341173827648163, -0.06424316018819809, -0.017605813220143318, -0.04357714205980301, 0.1214456781744957, -0.037513893097639084, 0.08211647719144821, -0.2510312795639038, 0.07324007153511047, 0.030913814902305603, 0.11832740157842636, -0.076419398188591, 0.03519592806696892, 0.02711334452033043, -0.052231304347515106, 0.1079011857509613, 0.022080324590206146, -0.08240288496017456, -0.04418906569480896, -0.12473385781049728, 0.004804858472198248, 0.14370210468769073, 0.028460200875997543, 0.0462937131524086, 0.005872226785868406, 0.006397027987986803, -0.003506189677864313, 0.033335406333208084, -0.12283158302307129, -0.12932641804218292, 0.05291052535176277, 0.002950407098978758, -0.031665652990341187, -0.011007489636540413, -0.003208078909665346, -0.03353235870599747, 0.08720505237579346, -0.12737378478050232, -0.034732308238744736, -0.10265327990055084, -0.03443602845072746, 0.04651608690619469, -0.06658387184143066, 0.008093698881566525, -0.07098551839590073, 0.047073252499103546, -0.004932889714837074, -0.09548737853765488, 0.04055614396929741, -0.09927865117788315, -0.0867902860045433, -0.003907972946763039, 0.053965311497449875, 0.019897326827049255, -0.015717392787337303, -0.027511660009622574, 0.008479762822389603, -0.09278351068496704, -0.10541200637817383, 0.018572168424725533, 0.03345049172639847, 0.09543757140636444, 0.0254147257655859, -0.044678620994091034, -0.13902920484542847, -0.06565370410680771, -0.08875305205583572, -0.0013015973381698132, 0.2668437957763672, 0.007629524450749159, 0.10364372283220291, 0.21820485591888428, -0.08468057215213776, -0.19991357624530792, -0.034765344113111496, 0.03392552584409714, 0.027854135259985924, -0.01721113733947277, -0.13292105495929718, 0.09653156995773315, 0.04477132111787796, 0.01648666150867939, 0.07843922078609467, -0.17263342440128326, -0.11314782500267029, 0.0025804475881159306, 0.07998859882354736, 0.20154953002929688, -0.0955306887626648, -0.02951171062886715, -0.025909097865223885, 0.04477188363671303, 0.028825197368860245, -0.09513188898563385, 0.13605721294879913, 0.019286977127194405, -0.012857886962592602, 0.002820300869643688, -0.059624820947647095, 0.029820697382092476, -0.018248729407787323, 0.04866476729512215, -0.045076992362737656, 0.053407613188028336, 0.16046951711177826, -0.06552974134683609, 0.08741224557161331, -0.05383759364485741, 0.10565521568059921, -0.0680898055434227, -0.04638207331299782, -0.012463865801692009, 0.03395550325512886, 0.000056637894886080176, -0.04435224086046219, -0.03335869684815407, 0.04186901077628136, 0.07202859967947006, -0.008534093387424946, -0.004345600958913565, -0.04366001486778259, -0.01799836754798889, 0.21934457123279572, 0.11513414978981018, -0.03331666439771652, 0.05510133132338524, -0.03422611206769943, -0.04722009226679802, 0.05251205712556839, -0.21762387454509735, 0.02735750749707222, 0.020503981038928032, 0.013203135691583157, 0.05433940142393112, 0.0033553459215909243, -0.1128765270113945, 0.06649397313594818, 0.05928304046392441, -0.17370493710041046, -0.08184455335140228, -0.028729552403092384, -0.05993863567709923, -0.06191444396972656, 0.0042861751280725, 0.15323668718338013, -0.1031193882226944, 0.026543287560343742, 0.014219257049262524, 0.011850140988826752, 0.021141258999705315, 0.11035007238388062, 0.03774256259202957, 0.005734254140406847, -0.03843028098344803, 0.1179313212633133, 0.0458948090672493, 0.010014506988227367, 0.06523014605045319, 0.08368680626153946, -0.1257176697254181, -0.08144122362136841, -0.03135843947529793, 0.1177263930439949, 0.06813788414001465, -0.012906543910503387, -0.0048400224186480045, 0.020451178774237633, -0.009947601705789566, 0.1287858784198761, 0.0005441871471703053, -0.020351387560367584, -0.013197099789977074, 0.029952820390462875, -0.07212437689304352, 0.13497310876846313, -0.0014155989047139883, 0.031379152089357376, -0.06472218781709671, 0.030779995024204254, -0.007799175567924976, 0.029799867421388626, -0.030378064140677452, -0.004942791536450386, -0.08258982002735138, -0.00882758293300867, -0.0050826952792704105, 0.05708660930395126, -0.05153164267539978, 0.00216341158375144, -0.04754135012626648, -0.010093278251588345, -0.04004267230629921, 0.033836882561445236, 0.006535082124173641, 0.011603230610489845, -0.020529089495539665, 0.06812573969364166, -0.08133753389120102, -0.004473465494811535, 0.004437948111444712, -0.06845809519290924, 0.10567256063222885, 0.04868048056960106, -0.05274425819516182, 0.044288039207458496, -0.03277686610817909, 0.0571553073823452, -0.008122490718960762, 0.058285605162382126, -0.01985321193933487, -0.07710419595241547, 0.005847024265676737, 0.02302202768623829, -0.002508132252842188, -0.011159821413457394, 0.11705061048269272, -0.0762706771492958, 0.01899193413555622, -0.03240295499563217, 0.019824661314487457, -0.1039595976471901, 0.05464574322104454, 0.05462345480918884, 0.06584526598453522, 0.10294246673583984, -0.06734182685613632, 0.013089613988995552, -0.08908635377883911, -0.014737102203071117, 0.05961434170603752, -0.09816712141036987, -0.07899650931358337, -0.02596028707921505, 0.08043403178453445, -0.10014275461435318, 0.12174490839242935, -0.04615776985883713, -0.030270632356405258, 0.014291034080088139, -0.12781822681427002, -0.04426582530140877, -0.020432552322745323, 0.13084827363491058, 0.041394852101802826, -0.005719635635614395, 0.0033298954367637634, 0.020799538120627403, 0.0016596533823758364, 0.1245330423116684, 0.06848733127117157, 0.1204124167561531, 0.044227562844753265, 0.059962086379528046, -0.024771681055426598, -0.03123009204864502, -0.06556185334920883, 0.12296191602945328, -0.10540363937616348, 0.08489848673343658, -0.04214181751012802, 0.04754934832453728, 0.20619261264801025, -0.07374076545238495, 0.03568282350897789, -0.006917605642229319, -0.0746835321187973, -0.07990587502717972, -0.27031928300857544, -0.052924007177352905, -0.03982185572385788, 0.018504809588193893, -0.09132543951272964, 0.03864280879497528, 0.11548489332199097, 0.046902187168598175, 0.011496939696371555, 0.12927493453025818, 0.04460125416517258, -0.06664324551820755, 0.08803299814462662, -0.05494255572557449, -0.017804905772209167, 0.05296049639582634, -0.057396672666072845, 0.032370492815971375, -0.04223351925611496, 0.048129014670848846, 0.05453631281852722, 0.04168880730867386, 0.051917847245931625, -0.08368884027004242, -0.11010358482599258, -0.017359726130962372, 0.04220876097679138, 0.033491119742393494, 0.10280317068099976, 0.07196968048810959, -0.09670501947402954, 0.02135634608566761, 0.1492798924446106, -0.03223945572972298, -0.1049535870552063, -0.119363933801651, 0.08254341781139374, -0.010049889795482159, -0.012559115886688232, -0.04372105002403259, -0.09569551795721054, 0.04772869870066643, 0.26892390847206116, 0.13168944418430328, -0.08673784881830215, -0.01720406860113144, 0.0361662395298481, -0.013304128311574459, 0.02343270555138588, 0.13165830075740814, 0.04063548892736435, 0.14878061413764954, -0.031555142253637314, 0.04966958239674568, 0.0003153452998958528, -0.01575327292084694, -0.0775299221277237, 0.24444115161895752, -0.022751204669475555, -0.0406607985496521, -0.017100865021348, 0.059473440051078796, -0.07684308290481567, -0.161037415266037, -0.03612533584237099, -0.005590560846030712, -0.08328890800476074, -0.005357986316084862, -0.01269481424242258, -0.004560127388685942, 0.037310391664505005, -0.048776715993881226, 0.020035531371831894, -0.016506072133779526, 0.01588786020874977, -0.04538622498512268, -0.03720225393772125, 0.00693550193682313, -0.008738018572330475, 0.031054383143782616, 0.023390352725982666, 0.06345697492361069, 0.09718409180641174, -0.012759027071297169, -0.14746053516864777, 0.13362015783786774, -0.03398460894823074, -0.09025722742080688, 0.0365152545273304, 0.13725394010543823, -0.03830993175506592, 0.1103769913315773, 0.07085950672626495, 0.09791576117277145, 0.011468197219073772, 0.017928488552570343, -0.0017307816306129098, -0.06476045399904251, -0.0006197800976224244, -0.0682765319943428, 0.12906582653522491, 0.09536080062389374, -0.03181890770792961, -0.008028794080018997, -0.08114153146743774, 0.017720699310302734, -0.05458206310868263, 0.04008712247014046, 0.0033932377118617296, -0.18538756668567657, -0.04084869474172592, 0.06791999191045761, 0.07893940806388855, -0.24605105817317963, -0.028723839670419693, -0.045094799250364304, 0.0051137893460690975, -0.05616546794772148, 0.10668379068374634, 0.13283734023571014, 0.00399825070053339, -0.033365242183208466, -0.014744866639375687, -0.02863169275224209, 0.07677212357521057, -0.07783695310354233, -0.10179217904806137 ]
null
null
transformers
[Google's T5](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html) Pretraining Dataset: [C4](https://huggingface.co/datasets/c4) Other Community Checkpoints: [here](https://huggingface.co/models?search=t5) Paper: [Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer](https://arxiv.org/pdf/1910.10683.pdf) Authors: *Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu* ## Abstract Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts every language problem into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled datasets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new โ€œColossal Clean Crawled Corpusโ€, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our dataset, pre-trained models, and code. ![model image](https://camo.githubusercontent.com/623b4dea0b653f2ad3f36c71ebfe749a677ac0a1/68747470733a2f2f6d69726f2e6d656469756d2e636f6d2f6d61782f343030362f312a44304a31674e51663876727255704b657944387750412e706e67)
{"language": ["en", "fr", "ro", "de"], "license": "apache-2.0", "tags": ["summarization", "translation"], "datasets": ["c4"]}
translation
baffo32/t5-base-ptmap
[ "transformers", "pytorch", "tf", "jax", "rust", "t5", "text2text-generation", "summarization", "translation", "en", "fr", "ro", "de", "dataset:c4", "arxiv:1910.10683", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1910.10683" ]
[ "en", "fr", "ro", "de" ]
TAGS #transformers #pytorch #tf #jax #rust #t5 #text2text-generation #summarization #translation #en #fr #ro #de #dataset-c4 #arxiv-1910.10683 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Google's T5 Pretraining Dataset: C4 Other Community Checkpoints: here Paper: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer Authors: *Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu* ## Abstract Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts every language problem into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled datasets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new โ€œColossal Clean Crawled Corpusโ€, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our dataset, pre-trained models, and code. !model image
[ "## Abstract\n\nTransfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts every language problem into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled datasets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new โ€œColossal Clean Crawled Corpusโ€, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our dataset, pre-trained models, and code.\n\n!model image" ]
[ "TAGS\n#transformers #pytorch #tf #jax #rust #t5 #text2text-generation #summarization #translation #en #fr #ro #de #dataset-c4 #arxiv-1910.10683 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Abstract\n\nTransfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts every language problem into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled datasets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new โ€œColossal Clean Crawled Corpusโ€, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our dataset, pre-trained models, and code.\n\n!model image" ]
[ 94, 234 ]
[ "passage: TAGS\n#transformers #pytorch #tf #jax #rust #t5 #text2text-generation #summarization #translation #en #fr #ro #de #dataset-c4 #arxiv-1910.10683 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Abstract\n\nTransfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts every language problem into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled datasets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new โ€œColossal Clean Crawled Corpusโ€, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our dataset, pre-trained models, and code.\n\n!model image" ]
[ -0.0292768906801939, 0.18225382268428802, -0.0041877152398228645, 0.009821473620831966, 0.09184163808822632, 0.010078704915940762, 0.06612750142812729, 0.13024841248989105, -0.043179113417863846, 0.01671442575752735, -0.012918300926685333, 0.10580939799547195, 0.056272100657224655, 0.11034394800662994, 0.020778708159923553, -0.22784610092639923, 0.028818830847740173, -0.064539335668087, -0.0897923931479454, 0.0984521210193634, 0.09351509809494019, -0.06347410380840302, 0.017154835164546967, -0.02617882750928402, -0.03762077912688255, -0.009199926629662514, -0.0921127051115036, -0.06007753685116768, 0.0895027369260788, 0.03845744952559471, 0.0846492126584053, 0.04182503744959831, 0.010236766189336777, -0.21272006630897522, 0.02441113255918026, 0.08553260564804077, 0.006877186242491007, 0.07739976793527603, 0.012124686501920223, -0.012532035820186138, 0.14157862961292267, -0.07453136891126633, 0.028843117877840996, 0.03763701394200325, -0.0879739448428154, -0.14311841130256653, -0.03855114057660103, -0.012332103215157986, 0.026328066363930702, 0.026795608922839165, -0.018496422097086906, 0.09468940645456314, -0.0679047629237175, 0.06825003772974014, -8.007129963516491e-7, -0.27260783314704895, -0.06110240891575813, 0.009238067083060741, -0.010966110043227673, 0.11770988255739212, -0.05351150408387184, 0.011471078731119633, 0.01045012753456831, 0.006917908322066069, 0.057578831911087036, -0.008652467280626297, -0.10746648907661438, 0.027365781366825104, -0.13246816396713257, -0.04367145150899887, 0.24631793797016144, -0.019393332302570343, -0.050121452659368515, -0.1788313090801239, -0.053457386791706085, 0.07473471015691757, 0.018629297614097595, -0.13823029398918152, 0.021202949807047844, 0.015551265329122543, -0.007424959447234869, -0.05242908373475075, -0.08365926891565323, 0.0035966599825769663, -0.05990147963166237, 0.07590719312429428, 0.06464589387178421, 0.05963004380464554, 0.014600483700633049, 0.14710752665996552, 0.06320077925920486, -0.09676291048526764, -0.03018597513437271, -0.11201721429824829, -0.1216607391834259, -0.007987679913640022, 0.019042404368519783, -0.06411765515804291, 0.04452107846736908, 0.15428872406482697, 0.09159031510353088, 0.00005166762275621295, -0.04198009893298149, 0.02735012210905552, 0.10057345032691956, 0.08714614063501358, -0.030406421050429344, -0.1142340749502182, -0.015770532190799713, 0.025211084634065628, -0.054674096405506134, -0.029823217540979385, 0.0020324073266237974, 0.017073635011911392, 0.009403901174664497, 0.10222918540239334, 0.04594551771879196, 0.04601208493113518, -0.0200817808508873, -0.05471666157245636, 0.0027641470078378916, -0.14326778054237366, -0.04635593667626381, -0.03634665161371231, -0.028312096372246742, 0.08664044737815857, 0.03992092236876488, -0.010842114686965942, -0.07818730175495148, -0.05918116122484207, -0.058017849922180176, -0.0862521380186081, -0.10188787430524826, -0.10647930949926376, 0.04423432797193527, -0.04091309756040573, -0.06830056011676788, -0.0741727352142334, -0.14392496645450592, -0.021858124062418938, 0.08128099888563156, -0.060047976672649384, 0.005510409362614155, -0.07404104620218277, -0.04895862191915512, -0.021710386499762535, -0.009419230744242668, -0.01950426772236824, -0.025370199233293533, 0.04835472255945206, -0.05617204308509827, 0.07198882848024368, 0.02125740610063076, 0.02973022498190403, -0.1000513881444931, 0.002484123921021819, -0.12832629680633545, 0.16391034424304962, -0.13074271380901337, -0.020365756005048752, -0.07806066423654556, -0.04103182628750801, -0.06439308077096939, 0.0356728658080101, -0.007351662032306194, 0.12058459222316742, -0.16208122670650482, -0.010431154631078243, 0.17486444115638733, -0.11193308979272842, -0.08526450395584106, 0.1338018923997879, -0.04690214991569519, 0.19759775698184967, 0.07656230032444, 0.1180586963891983, 0.1445600688457489, -0.002722702920436859, -0.06951674073934555, 0.09180712699890137, -0.022552182897925377, 0.07419772446155548, 0.05677236244082451, 0.01589631475508213, 0.04151909798383713, 0.04951432719826698, 0.01694640703499317, 0.01752748154103756, -0.0775923803448677, -0.09348946064710617, -0.0021118135191500187, 0.012037679553031921, -0.00911436602473259, 0.006320779677480459, 0.014126462861895561, -0.0035212889779359102, -0.10811974853277206, -0.01076428685337305, 0.029361587017774582, -0.09413273632526398, 0.008650715462863445, -0.0965382531285286, 0.029799170792102814, -0.07119973748922348, 0.0303653571754694, -0.17752504348754883, -0.11608820408582687, 0.05819506198167801, -0.06012524291872978, 0.08043448626995087, 0.15207824110984802, -0.005618839990347624, 0.01973801851272583, 0.00470389798283577, 0.033541809767484665, -0.04134540632367134, -0.017091963440179825, -0.026237815618515015, -0.15366299450397491, -0.007130269426852465, -0.07456596940755844, 0.020850351080298424, -0.06953447312116623, 0.03434320166707039, 0.0021410975605249405, 0.07897540181875229, 0.040262140333652496, -0.025534583255648613, -0.02727189101278782, 0.03663396090269089, -0.03797817602753639, -0.031193675473332405, 0.054697997868061066, -0.016147350892424583, -0.1053309291601181, 0.052382130175828934, -0.10577170550823212, 0.004926268942654133, 0.041414014995098114, -0.017157388851046562, -0.0564541295170784, 0.010643970221281052, -0.09612563252449036, -0.042371127754449844, -0.033737797290086746, -0.03684060648083687, 0.15646806359291077, 0.00928663369268179, 0.08485862612724304, -0.10783149302005768, -0.03875182196497917, -0.014152685180306435, 0.003287459025159478, 0.06369920819997787, 0.05481956526637077, 0.017121994867920876, -0.1359301209449768, 0.0853428989648819, 0.13390067219734192, 0.05856189504265785, 0.19914203882217407, -0.04744334891438484, -0.08853034675121307, -0.03218996524810791, 0.048813965171575546, -0.04255262389779091, 0.08190188556909561, -0.0628129243850708, 0.01539489347487688, 0.08055222779512405, 0.03359392657876015, 0.09696921706199646, -0.11286772042512894, 0.04430265352129936, 0.0227903351187706, -0.040563978254795074, 0.0337156318128109, 0.04753110557794571, 0.012864026241004467, 0.1041228175163269, -0.024940256029367447, -0.0016728023765608668, 0.0045934515073895454, -0.061155594885349274, -0.0523417666554451, 0.16329315304756165, -0.10945097357034683, -0.3157930076122284, -0.11807499080896378, 0.10171467065811157, -0.013649542815983295, -0.03339000046253204, 0.010021586902439594, -0.016254106536507607, -0.07279986888170242, -0.05101306363940239, 0.03793385624885559, -0.05235835164785385, -0.10911617428064346, -0.03176683560013771, 0.05568001791834831, -0.052277255803346634, -0.15321537852287292, 0.008246679790318012, 0.014295418746769428, -0.15522608160972595, 0.052645642310380936, -0.01525223720818758, 0.0869457945227623, 0.08190558105707169, -0.030791329219937325, -0.030063344165682793, -0.035559624433517456, 0.1075069010257721, -0.10143445432186127, 0.05234038829803467, 0.0652550533413887, -0.07190551608800888, 0.055850643664598465, 0.0709693655371666, 0.027023237198591232, -0.045666519552469254, 0.023430878296494484, 0.02704271301627159, -0.05536556988954544, -0.29248183965682983, -0.07810389995574951, -0.10668203234672546, 0.04778029024600983, -0.04833897575736046, 0.02362678386271, 0.06117384508252144, 0.016547085717320442, -0.02981858141720295, -0.040185824036598206, 0.05861705169081688, 0.07158444076776505, 0.14038318395614624, -0.04753325507044792, 0.10846099257469177, -0.07958164066076279, -0.08095171302556992, 0.08592090755701065, 0.05206432193517685, 0.30187883973121643, 0.01297898218035698, 0.06243964284658432, 0.1033032163977623, 0.09505379945039749, 0.02528323605656624, 0.07592368125915527, -0.019013598561286926, 0.045973040163517, -0.08710579574108124, -0.08044429123401642, -0.048407576978206635, 0.0838165432214737, 0.020063068717718124, -0.020518554374575615, 0.00017092951748054475, 0.052005916833877563, 0.0890946239233017, 0.19043423235416412, 0.04663918539881706, -0.09974589198827744, -0.001785188214853406, 0.05325238034129143, -0.053277961909770966, -0.0877271294593811, 0.06481463462114334, 0.14773054420948029, -0.11631225794553757, 0.011738794855773449, -0.040306370705366135, 0.09401460736989975, -0.06662128865718842, -0.0010061684297397733, -0.03495891019701958, 0.04423747584223747, 0.0060506477020680904, 0.15288899838924408, -0.14689266681671143, 0.13929077982902527, 0.012771924026310444, 0.022164825350046158, -0.06781885772943497, 0.045154083520174026, 0.016036195680499077, -0.045168619602918625, 0.14354208111763, -0.0004683639563154429, -0.04559938982129097, 0.02619965374469757, -0.08551742136478424, 0.10993587225675583, 0.13123328983783722, -0.04730161279439926, 0.05792595446109772, -0.023523695766925812, 0.031233379617333412, 0.014004545286297798, 0.08989911526441574, -0.1118682250380516, -0.2040782868862152, 0.034910645335912704, -0.0832732692360878, -0.01089834701269865, -0.056213345378637314, -0.05178227648139, -0.022038588300347328, 0.2532600164413452, -0.13501259684562683, -0.09340638667345047, -0.0928100198507309, 0.01259060949087143, 0.12034647166728973, 0.001697375439107418, -0.01564089022576809, -0.019817152991890907, 0.014124790206551552, 0.0026234241668134928, -0.12637023627758026, 0.03530919924378395, -0.034303389489650726, -0.10952485352754593, -0.06319548934698105, 0.08934740722179413, 0.0726073682308197, 0.038616325706243515, -0.017551705241203308, -0.021360279992222786, 0.0012555152643471956, -0.10470985621213913, -0.05014420300722122, 0.15481358766555786, 0.015785476192831993, 0.16800415515899658, -0.1649979054927826, -0.05941886827349663, -0.06311656534671783, -0.011271948926150799, 0.1228160485625267, 0.12080705165863037, -0.0708259865641594, 0.1996089071035385, 0.12222633510828018, -0.14527392387390137, -0.24051165580749512, -0.0314275324344635, 0.0011900728568434715, 0.05328755080699921, 0.04074627533555031, -0.19037319719791412, 0.05470579117536545, 0.05406409129500389, -0.006447969935834408, 0.017549678683280945, -0.20625267922878265, -0.07185646891593933, 0.057439595460891724, -0.014082546345889568, 0.1607620269060135, -0.08411466330289841, 0.0016396850114688277, 0.024499844759702682, -0.06437769532203674, 0.17275406420230865, -0.03944462537765503, 0.0862630233168602, -0.009029999375343323, -0.04681912809610367, 0.05389980971813202, -0.03967049717903137, 0.10478201508522034, 0.024068932980298996, 0.059912439435720444, -0.011670634150505066, 0.1036803126335144, 0.150765061378479, -0.012257919646799564, 0.057612061500549316, 0.024887891486287117, 0.07048441469669342, -0.1194397360086441, -0.04240049794316292, -0.04062730073928833, 0.03531442582607269, -0.04168293997645378, -0.07961582392454147, -0.114849753677845, 0.10192780196666718, 0.11099044233560562, 0.00605033477768302, 0.11022167652845383, -0.04814865067601204, 0.11713360995054245, 0.09833686053752899, 0.15286768972873688, -0.006228122394531965, 0.006318278145045042, -0.0530608594417572, 0.00712587870657444, 0.06714897602796555, -0.146331325173378, 0.05389726161956787, 0.1698908507823944, -0.004327967297285795, 0.11255891621112823, 0.056000739336013794, -0.0894652009010315, -0.013402231968939304, 0.06311877071857452, -0.1303490251302719, -0.18055596947669983, -0.03537128493189812, -0.07221288233995438, -0.05497989431023598, 0.017685700207948685, 0.12441211193799973, -0.03974565863609314, 0.0017538229003548622, 0.011657352559268475, 0.02919253148138523, -0.010763600468635559, 0.0960727334022522, -0.015122534707188606, 0.027656417340040207, -0.04508129507303238, 0.12602181732654572, 0.0643344298005104, -0.13936637341976166, 0.013188721612095833, -0.008057843893766403, -0.09425758570432663, -0.01440628431737423, 0.0018192612333223224, 0.08455643057823181, 0.006082258187234402, -0.041397545486688614, -0.01115730032324791, -0.05321922525763512, 0.029040295630693436, 0.09644488990306854, 0.016102328896522522, 0.08592911809682846, -0.04734228923916817, 0.0294461939483881, -0.051311906427145004, 0.11477643996477127, 0.007752511650323868, -0.02543867938220501, -0.03665030747652054, 0.04913926497101784, 0.012973382137715816, 0.04564260318875313, -0.05394737422466278, -0.0901021808385849, -0.0704430490732193, -0.01536836102604866, -0.08024444431066513, -0.01443908829241991, -0.05394873395562172, -0.024078646674752235, 0.022888096049427986, -0.0063237533904612064, -0.025571050122380257, -0.01973128877580166, -0.06604484468698502, 0.016117455437779427, 0.011721995659172535, 0.09366060048341751, -0.08346470445394516, -0.014385734684765339, 0.035359445959329605, -0.09196092188358307, 0.08304022252559662, -0.02353273518383503, -0.05295824259519577, 0.02022562175989151, -0.07907620072364807, -0.0027235024608671665, 0.007309318520128727, 0.10541538894176483, -0.020100649446249008, -0.014731459319591522, 0.01833227090537548, 0.025680923834443092, -0.016362039372324944, 0.00922787468880415, -0.014078338630497456, -0.09817538410425186, 0.045766089111566544, -0.01695685274899006, -0.04352260380983353, -0.046495355665683746, 0.03439931198954582, 0.012075909413397312, 0.09417516738176346, 0.07526391744613647, -0.03313826024532318, 0.04261501878499985, -0.17217126488685608, -0.03907253220677376, 0.0668274462223053, -0.005056893918663263, 0.07282237708568573, -0.10053455084562302, 0.03928934410214424, -0.04480923339724541, 0.2559295892715454, 0.018870193511247635, -0.020370863378047943, 0.019445497542619705, -0.07413867861032486, -0.04386194795370102, -0.0073778764344751835, 0.09159683436155319, 0.08490245044231415, 0.01765887625515461, -0.01515384204685688, -0.0424082987010479, -0.07462865114212036, -0.07299651950597763, 0.1187584400177002, 0.14103075861930847, 0.031793102622032166, 0.022988542914390564, 0.04190729930996895, -0.04701821133494377, -0.02141614817082882, 0.0764693096280098, 0.05305313318967819, 0.08870645612478256, -0.017247246578335762, 0.019132349640130997, 0.06296157836914062, -0.09546627104282379, 0.14501598477363586, -0.005357325542718172, -0.06690108776092529, -0.13625966012477875, -0.10904102772474289, -0.019214337691664696, -0.030362898483872414, 0.0058095213025808334, -0.17518803477287292, 0.041958749294281006, 0.06513885408639908, 0.06783711165189743, -0.04983483627438545, 0.028974782675504684, -0.12330864369869232, -0.10103844851255417, 0.0775776281952858, 0.001480169128626585, 0.11423051357269287, -0.04704958200454712, 0.008693848736584187, 0.00281310360878706, 0.12077239900827408, 0.023331934586167336, 0.029897471889853477, 0.06588800996541977, 0.008724216371774673, -0.07403005659580231, -0.04266003146767616, -0.036015819758176804, -0.0035055072512477636, 0.0015904074534773827, 0.05908675491809845, 0.041109103709459305, -0.08382236212491989, 0.014631759375333786, 0.23694674670696259, -0.0533352755010128, -0.08131448179483414, -0.18084785342216492, 0.2253575325012207, -0.006556242238730192, 0.008763634599745274, 0.07974957674741745, -0.05152544751763344, -0.039319030940532684, 0.10803171247243881, 0.24117515981197357, -0.04653250053524971, -0.055447544902563095, 0.04387856274843216, 0.0001664531446294859, 0.007541085593402386, 0.12955619394779205, 0.02853906899690628, 0.3161357343196869, -0.03283298760652542, 0.02225332334637642, -0.03905506432056427, -0.022043759003281593, -0.01170310378074646, 0.10423196107149124, 0.024115659296512604, -0.02314462512731552, -0.03452249616384506, 0.09017670899629593, -0.07910098135471344, -0.2203178107738495, 0.04900076240301132, -0.09762134402990341, -0.13461697101593018, -0.031247273087501526, 0.03892619162797928, -0.04976431280374527, 0.003484320128336549, -0.0335984043776989, -0.023257028311491013, 0.17606791853904724, 0.017115890979766846, -0.04786108806729317, 0.05054319277405739, 0.07199534773826599, -0.11383584886789322, 0.046703074127435684, 0.015849091112613678, 0.09762603044509888, 0.09730692952871323, 0.020324131473898888, -0.029503129422664642, 0.0028206969145685434, 0.01589817926287651, -0.0036217463202774525, -0.0029217645060271025, 0.09649927169084549, 0.04062945768237114, 0.10665224492549896, 0.13395673036575317, -0.039097465574741364, 0.016939079388976097, -0.02229979820549488, -0.12589170038700104, -0.08203685283660889, 0.03904718533158302, -0.061600103974342346, 0.11630228161811829, 0.15739141404628754, 0.0014675732236355543, 0.043661367148160934, -0.07739030569791794, 0.018838252872228622, 0.00025181309320032597, 0.0861152857542038, -0.030231798067688942, -0.08326935023069382, -0.02871343307197094, -0.09844738245010376, 0.01360742561519146, -0.23345069587230682, -0.057340219616889954, 0.032418444752693176, -0.01930323988199234, -0.05326221138238907, 0.12036903947591782, 0.049931950867176056, 0.03876541927456856, -0.025619564577937126, -0.1824258714914322, 0.0041588363237679005, 0.06674546748399734, -0.09411511570215225, 0.008561227470636368 ]
null
null
transformers
# Model name Indian Political Tweets LM ## Model description Note: This model is based on GPT2, if you want a bigger model based on GPT2-medium and finetuned on the same data please take a look at the [IndianPoliticalTweetsLMMedium](https://huggingface.co/bagdaebhishek/IndianPoliticalTweetsLMMedium) model. This is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this [blog](https://bagdeabhishek.github.io/twitterAnalysis) post. ## Intended uses & limitations This finetuned model can be used to generate tweets which are related to Indian politics. #### How to use ```python from transformers import AutoTokenizer,AutoModelWithLMHead,pipeline tokenizer = AutoTokenizer.from_pretrained("bagdaebhishek/IndianPoliticalTweetsLM") model = AutoModelWithLMHead.from_pretrained("bagdaebhishek/IndianPoliticalTweetsLM") text_generator = pipeline("text-generation",model=model, tokenizer=tokenizer) init_sentence = "India will always be" print(text_generator(init_sentence)) ``` #### Limitations and bias 1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate "Hinglish" text and hence no assumptions should be made about the language of the generated text. 2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like "-sent via NamoApp" etc. 3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model. ## Training data I used the pre-trained gpt2 model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a [blog](https://bagdeabhishek.github.io/twitterAnalysis) post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog. ## Training procedure For pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values. I then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles. ### Hardware 1. GPU: GTX 1080Ti 2. CPU: Ryzen 3900x 3. RAM: 32GB This model took roughly 36 hours to fine-tune.
{"language": "en", "license": "apache-2.0", "tags": ["India", "politics", "tweets", "BJP", "Congress", "AAP", "pytorch", "gpt2", "lm-head", "text-generation"], "datasets": ["Twitter", "IndianPolitics"], "thumbnail": "https://bagdeabhishek.github.io/twitterAnalysis_files/networkfin.jpg"}
text-generation
bagdaebhishek/IndianPoliticalTweetsLM
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "India", "politics", "tweets", "BJP", "Congress", "AAP", "lm-head", "en", "dataset:Twitter", "dataset:IndianPolitics", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model name Indian Political Tweets LM ## Model description Note: This model is based on GPT2, if you want a bigger model based on GPT2-medium and finetuned on the same data please take a look at the IndianPoliticalTweetsLMMedium model. This is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post. ## Intended uses & limitations This finetuned model can be used to generate tweets which are related to Indian politics. #### How to use #### Limitations and bias 1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate "Hinglish" text and hence no assumptions should be made about the language of the generated text. 2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like "-sent via NamoApp" etc. 3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model. ## Training data I used the pre-trained gpt2 model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog. ## Training procedure For pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values. I then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles. ### Hardware 1. GPU: GTX 1080Ti 2. CPU: Ryzen 3900x 3. RAM: 32GB This model took roughly 36 hours to fine-tune.
[ "# Model name\nIndian Political Tweets LM", "## Model description\nNote: This model is based on GPT2, if you want a bigger model based on GPT2-medium and finetuned on the same data please take a look at the IndianPoliticalTweetsLMMedium model. \n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post.", "## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.", "#### How to use", "#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.", "## Training data\nI used the pre-trained gpt2 model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog.", "## Training procedure\n\nFor pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values.\n\nI then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles.", "### Hardware\n1. GPU: GTX 1080Ti\n2. CPU: Ryzen 3900x\n3. RAM: 32GB\n\nThis model took roughly 36 hours to fine-tune." ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model name\nIndian Political Tweets LM", "## Model description\nNote: This model is based on GPT2, if you want a bigger model based on GPT2-medium and finetuned on the same data please take a look at the IndianPoliticalTweetsLMMedium model. \n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post.", "## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.", "#### How to use", "#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.", "## Training data\nI used the pre-trained gpt2 model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog.", "## Training procedure\n\nFor pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values.\n\nI then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles.", "### Hardware\n1. GPU: GTX 1080Ti\n2. CPU: Ryzen 3900x\n3. RAM: 32GB\n\nThis model took roughly 36 hours to fine-tune." ]
[ 94, 9, 104, 29, 5, 176, 79, 103, 34 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model name\nIndian Political Tweets LM## Model description\nNote: This model is based on GPT2, if you want a bigger model based on GPT2-medium and finetuned on the same data please take a look at the IndianPoliticalTweetsLMMedium model. \n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post.## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.#### How to use#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.## Training data\nI used the pre-trained gpt2 model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog." ]
[ -0.004204807337373495, 0.1296338438987732, -0.0036303808446973562, 0.06684866547584534, 0.0464516282081604, 0.01339737419039011, 0.07442080229520798, 0.047065794467926025, 0.007875601761043072, 0.0978979840874672, 0.02485169842839241, 0.013895776122808456, 0.055605243891477585, 0.13324156403541565, 0.09876040369272232, -0.2915000319480896, 0.05413821339607239, -0.07734104990959167, 0.10066542774438858, 0.10638148337602615, 0.09430836886167526, -0.06639930605888367, 0.05767744779586792, 0.004226744640618563, -0.04476955905556679, 0.017749706283211708, -0.03142658993601799, -0.050887059420347214, 0.06744503229856491, 0.06931702047586441, 0.018185414373874664, -0.024204254150390625, -0.002174210734665394, -0.1483626365661621, 0.022108521312475204, 0.07767357677221298, -0.04638674855232239, 0.014363070018589497, 0.10711819678544998, -0.06163738667964935, 0.24184012413024902, -0.042066220194101334, 0.060553085058927536, 0.08850322663784027, -0.1306527853012085, -0.005992933642119169, -0.12465962022542953, 0.12926632165908813, 0.11785339564085007, 0.10285146534442902, -0.07150312513113022, 0.18434548377990723, -0.043355345726013184, 0.03524641692638397, 0.11458302289247513, -0.1495654433965683, 0.016727512702345848, -0.019191328436136246, -0.03668399527668953, 0.10279739648103714, -0.027449944987893105, 0.03918600454926491, 0.033247802406549454, 0.01462464313954115, -0.06950180977582932, -0.024234192445874214, 0.11574657261371613, -0.017253732308745384, -0.1060195118188858, -0.0457378551363945, 0.036031581461429596, 0.0200200192630291, -0.01877192035317421, -0.18051445484161377, -0.029373878613114357, 0.07931451499462128, 0.0018031023209914565, -0.014853890985250473, -0.0045325555838644505, 0.045312102884054184, 0.12995441257953644, -0.11107941716909409, -0.08770116418600082, 0.05471375584602356, 0.0102846072986722, 0.042381078004837036, 0.017199192196130753, 0.012022223323583603, -0.0739486888051033, 0.025700215250253677, -0.08450661599636078, -0.08516030013561249, -0.012296587228775024, -0.03846026211977005, -0.08461330831050873, -0.022495251148939133, -0.05049420893192291, -0.043964143842458725, -0.017064865678548813, 0.16869200766086578, -0.07201990485191345, 0.01445898786187172, 0.014625308103859425, 0.009383942931890488, 0.16276465356349945, 0.09040744602680206, -0.1607978790998459, -0.021473633125424385, 0.05088011175394058, -0.1202712282538414, 0.052277930080890656, 0.02090848609805107, 0.04005452245473862, 0.05232210084795952, -0.010517208836972713, 0.10738711059093475, 0.09375555068254471, 0.041376978158950806, -0.08811447024345398, -0.042666416615247726, 0.17561498284339905, -0.12754833698272705, -0.0003387411416042596, -0.04492562264204025, 0.0028558962512761354, 0.09963258355855942, 0.00032774495775811374, 0.08264966309070587, -0.08117680251598358, 0.004911775700747967, -0.11468978971242905, 0.05010469630360603, -0.03349478542804718, -0.0665578693151474, 0.007893023081123829, -0.025680450722575188, -0.08940625190734863, -0.10210920870304108, -0.14940306544303894, -0.03665247559547424, -0.03725016117095947, -0.08680418878793716, -0.009487352333962917, 0.02523837424814701, 0.07837501913309097, -0.018095405772328377, 0.03249416500329971, -0.05940043926239014, 0.005641886033117771, 0.03250453248620033, -0.09139374643564224, 0.04273093491792679, 0.09841693937778473, -0.002978803589940071, -0.19656230509281158, 0.07664074003696442, -0.2079288214445114, 0.14649036526679993, -0.03475484251976013, 0.049450382590293884, -0.12034673988819122, 0.0023772430140525103, 0.04357120767235756, 0.021865874528884888, -0.07809522747993469, 0.1958867460489273, -0.15663185715675354, -0.024097124114632607, 0.10454783588647842, -0.2159920632839203, -0.0017533235950395465, 0.12695929408073425, 0.004341724794358015, 0.14457741379737854, 0.13925984501838684, 0.12290047854185104, -0.004845018964260817, -0.08640486747026443, -0.09090438485145569, -0.13662533462047577, -0.03653382509946823, 0.2063400000333786, 0.037359122186899185, -0.14232754707336426, 0.03147810697555542, -0.007999984547495842, 0.009778250008821487, 0.010010751895606518, -0.026729587465524673, -0.019792629405856133, 0.033613260835409164, -0.09281740337610245, 0.05385880172252655, -0.005035684909671545, -0.027866413816809654, -0.02738630399107933, -0.13804744184017181, 0.0030229196418076754, 0.11709504574537277, -0.006569944322109222, 0.03395768254995346, -0.10494852066040039, -0.01602347381412983, -0.056688420474529266, -0.01320226676762104, -0.2056356966495514, -0.13829217851161957, 0.013183042407035828, -0.1203443706035614, 0.15663260221481323, -0.032193951308727264, -0.000747826648876071, 0.023130424320697784, -0.043633077293634415, -0.0023462711833417416, 0.04787927493453026, 0.028739798814058304, -0.0967627465724945, -0.10354353487491608, -0.026158485561609268, -0.05042466148734093, 0.19193853437900543, -0.0755298063158989, -0.0031564284581691027, 0.09401822835206985, 0.1842629462480545, 0.06560731679201126, -0.07955668121576309, 0.11619647592306137, -0.009046753868460655, 0.025772035121917725, -0.0844525396823883, 0.0009244973189197481, -0.027945227921009064, 0.04679710417985916, 0.0629412829875946, -0.17066749930381775, -0.15008068084716797, 0.06775815784931183, 0.0496719554066658, -0.09549617767333984, 0.07031218707561493, -0.04978308081626892, -0.018253879621624947, -0.01296924613416195, -0.06496596336364746, 0.07335097342729568, 0.054998673498630524, 0.05149536207318306, -0.04074271023273468, -0.10533727705478668, -0.03194784000515938, -0.04451689496636391, -0.1274726837873459, 0.03317924588918686, 0.12378589808940887, -0.2214755415916443, 0.08988164365291595, 0.07899449020624161, 0.08608001470565796, 0.17728862166404724, 0.07346402853727341, -0.07551887631416321, -0.046651680022478104, -0.09833666682243347, -0.018032995983958244, -0.0711333379149437, 0.042109403759241104, 0.020361516624689102, 0.035285357385873795, 0.005055197048932314, 0.07972896844148636, 0.04091930761933327, 0.023708129301667213, 0.04686444252729416, 0.015551475808024406, 0.0023754718713462353, 0.04209277406334877, -0.0395343191921711, 0.07946576923131943, -0.055566366761922836, 0.14991925656795502, 0.02186458930373192, 0.050661228597164154, -0.18465380370616913, 0.09344195574522018, -0.14379680156707764, -0.289349228143692, -0.05150008946657181, 0.006757156923413277, -0.018607247620821, -0.018983611837029457, 0.03434360399842262, -0.14644072949886322, -0.12010526657104492, -0.10315670818090439, 0.07381295412778854, 0.06478560715913773, -0.05022832751274109, -0.08979946374893188, -0.03333435580134392, 0.0007207265589386225, -0.11204692721366882, 0.012922179885208607, 0.05044889450073242, -0.09039049595594406, 0.06892799586057663, -0.01726243458688259, 0.08745689690113068, 0.15708540380001068, 0.005098240450024605, 0.0020675596315413713, 0.01578572206199169, 0.25817543268203735, -0.08938657492399216, 0.18642109632492065, 0.1147313192486763, 0.06810109317302704, 0.04886062815785408, 0.10308387875556946, -0.03724159300327301, -0.06650937348604202, 0.04729582741856575, 0.09819015860557556, -0.01077389158308506, -0.20293904840946198, -0.13297291100025177, 0.02682678773999214, -0.04428228735923767, 0.06461919844150543, 0.05296372249722481, 0.13841281831264496, 0.026342222467064857, -0.1118474081158638, -0.059986017644405365, 0.031639911234378815, 0.04475538805127144, 0.04549800232052803, -0.0022211535833775997, 0.03503670170903206, -0.07815587520599365, 0.06709521263837814, 0.1345164179801941, -0.10720105469226837, 0.11318228393793106, 0.07482194155454636, 0.20245876908302307, 0.061141930520534515, 0.07169511914253235, 0.04352310672402382, 0.049747079610824585, -0.011815804988145828, 0.03691732510924339, -0.004130141343921423, -0.05770368129014969, 0.005979426205158234, 0.09941690415143967, 0.07185401767492294, -0.11079983413219452, 0.025330636650323868, -0.06348216533660889, 0.05578477680683136, 0.11850201338529587, 0.013821437023580074, -0.08512672781944275, -0.0964849442243576, 0.0593390129506588, -0.09936536848545074, -0.0227492805570364, -0.020824450999498367, 0.0459219329059124, -0.12077533453702927, 0.05958947166800499, 0.012369133532047272, 0.0570770688354969, -0.05785007402300835, -0.02090415731072426, -0.008907260373234749, -0.03251997381448746, -0.038139134645462036, 0.09226113557815552, -0.1663907915353775, 0.11509308964014053, 0.030529696494340897, 0.029333285987377167, -0.0705026164650917, -0.026413464918732643, -0.023232117295265198, 0.03315224125981331, 0.09704288840293884, 0.041529521346092224, 0.04574292153120041, -0.0655423253774643, -0.07469893246889114, -0.016600390896201134, 0.08444235473871231, -0.09756078571081161, 0.10605732351541519, -0.0007907362887635827, -0.015030302107334137, -0.04951279237866402, 0.014533613808453083, -0.23295699059963226, -0.17682231962680817, 0.10476972907781601, -0.05906146392226219, 0.09006184339523315, -0.014051518402993679, -0.04265564680099487, -0.07230084389448166, 0.19395610690116882, -0.03215739503502846, -0.1613038182258606, -0.1186850368976593, -0.01493148598819971, 0.1355908066034317, -0.11091785877943039, 0.03991592302918434, 0.011559374630451202, 0.1619618833065033, -0.0729466900229454, -0.04805782809853554, 0.013670610263943672, -0.07866144180297852, -0.18263736367225647, 0.0280921533703804, 0.0755714401602745, 0.15849773585796356, 0.04887326434254646, 0.06766440719366074, 0.06995520740747452, 0.010505873709917068, -0.1414518654346466, -0.045262545347213745, 0.06615080684423447, 0.05541427060961723, 0.010392172262072563, -0.004173866473138332, -0.08479326963424683, -0.12972749769687653, -0.1038597971200943, 0.059511490166187286, 0.24713724851608276, -0.04181468114256859, 0.19892185926437378, 0.2065330147743225, -0.07664160430431366, -0.08412246406078339, -0.1154899150133133, -0.04749765247106552, -0.05708259716629982, 0.11350598186254501, -0.20167219638824463, 0.01470837090164423, 0.12298093736171722, 0.006613667588680983, -0.10298163443803787, -0.1643269658088684, -0.1427510678768158, 0.0587351992726326, 0.010557882487773895, -0.08358398079872131, -0.10865454375743866, -0.08159366250038147, -0.02260555885732174, -0.1338023692369461, 0.1755211502313614, -0.06656932085752487, -0.033776625990867615, 0.0325637049973011, 0.10153625905513763, 0.023185165598988533, -0.038989778608083725, 0.1143278256058693, 0.07241299003362656, 0.04501784220337868, -0.14264971017837524, -0.09187145531177521, 0.13280299305915833, 0.014743780717253685, 0.11349690705537796, 0.09673931449651718, 0.0567898228764534, -0.21165017783641815, -0.08856537193059921, -0.12365586310625076, -0.016383226960897446, -0.08303358405828476, -0.056407153606414795, -0.04358331859111786, 0.12294810265302658, 0.06851258873939514, -0.007523002568632364, -0.08515505492687225, -0.10677125304937363, 0.02458084560930729, 0.16012290120124817, 0.13754858076572418, -0.0038365842774510384, -0.016438905149698257, -0.11138863116502762, -0.04494408890604973, 0.01149353850632906, -0.06086406856775284, 0.01385399792343378, 0.03757288679480553, 0.0379277765750885, 0.13945230841636658, -0.09645553678274155, -0.13844066858291626, 0.06774594634771347, 0.01741412654519081, -0.08414755016565323, -0.1138872429728508, -0.01660260185599327, 0.08863622695207596, -0.007996100932359695, -0.08668562024831772, 0.11007845401763916, -0.019541732966899872, -0.08299107849597931, -0.02966281771659851, 0.06022441387176514, -0.03125409036874771, 0.09723537415266037, 0.043013401329517365, 0.029158245772123337, -0.08689433336257935, 0.0776134803891182, 0.17204870283603668, -0.0562613271176815, 0.02120063453912735, 0.19073353707790375, -0.10992307215929031, -0.10569752007722855, -0.0907791331410408, -0.08420734852552414, -0.09201500564813614, 0.02742922492325306, 0.042312681674957275, 0.016170859336853027, 0.04453519731760025, 0.11619327962398529, -0.028013665229082108, 0.03795801103115082, -0.029952244833111763, -0.061687953770160675, -0.08354514092206955, -0.005782688967883587, 0.11399524658918381, -0.0658719390630722, -0.025117993354797363, 0.1441214233636856, 0.029138101264834404, 0.005933829117566347, -0.015367820858955383, -0.05275336280465126, -0.013374284841120243, -0.027928851544857025, -0.09054496139287949, 0.00790165551006794, -0.11116937547922134, -0.020141221582889557, -0.026126841083168983, -0.055049244314432144, 0.004177816677838564, -0.01686951518058777, -0.0502784438431263, -0.02639307826757431, -0.03070869669318199, 0.0692688375711441, -0.16075611114501953, -0.008718792349100113, 0.13318368792533875, -0.03766826167702675, 0.13547904789447784, 0.043151091784238815, -0.046835530549287796, 0.010615641251206398, -0.0763837918639183, -0.02003197930753231, -0.05488242208957672, -0.020387956872582436, -0.016074618324637413, -0.17547890543937683, -0.044424284249544144, -0.026462946087121964, -0.07271517068147659, 0.037578850984573364, 0.02705954760313034, -0.10119573026895523, 0.14068841934204102, 0.046100322157144547, -0.04657364636659622, -0.09776594489812851, 0.024083498865365982, 0.004115009680390358, -0.10177574306726456, 0.12226630747318268, -0.04683837667107582, 0.018242157995700836, -0.1546550840139389, 0.025666605681180954, -0.015247934497892857, 0.07786960154771805, -0.025763744488358498, 0.014914612285792828, 0.02586969919502735, 0.02789204940199852, 0.08657256513834, -0.04665105417370796, -0.005867545492947102, 0.11254151910543442, 0.05025545135140419, -0.045775167644023895, -0.019688008353114128, 0.06633402407169342, 0.004274189937859774, 0.05698762834072113, -0.00968892127275467, -0.0411103256046772, -0.07286913692951202, -0.0899617075920105, 0.12132806330919266, 0.06901630759239197, 0.07597734779119492, -0.021437661722302437, -0.06815320253372192, -0.05101614445447922, -0.07706243544816971, 0.01868310198187828, -0.0736527144908905, -0.055532511323690414, -0.05740911513566971, -0.01737193576991558, 0.2781458795070648, -0.11769947409629822, 0.07212372124195099, 0.04103486239910126, -0.0178103968501091, 0.008509350940585136, -0.2906023859977722, -0.040756694972515106, 0.03558500111103058, 0.007984858937561512, -0.07381537556648254, 0.11158522963523865, 0.11077908426523209, 0.05937245488166809, 0.00386087317019701, 0.09388839453458786, -0.11177000403404236, -0.10916996747255325, -0.014359837397933006, -0.015303141437470913, -0.007756222039461136, -0.005757274106144905, 0.049945980310440063, -0.08814429491758347, 0.04081451892852783, 0.02456035651266575, 0.10259458422660828, 0.00020723645866382867, -0.009982635267078876, -0.07344766706228256, -0.12056916952133179, 0.008988833986222744, 0.021645262837409973, 0.008811954408884048, 0.15722116827964783, 0.013332106173038483, 0.00905503612011671, 0.01063872966915369, 0.1440279483795166, 0.019258245825767517, -0.15165360271930695, -0.11983615905046463, 0.1487990766763687, -0.03137245774269104, 0.008837061002850533, 0.03372578322887421, -0.11461491882801056, 0.06910261511802673, 0.1430899202823639, 0.12624433636665344, 0.0015945088816806674, 0.033254433423280716, -0.06892254203557968, 0.037384748458862305, 0.02560095675289631, 0.048114512115716934, -0.021702712401747704, 0.22860050201416016, -0.13248085975646973, 0.15358379483222961, -0.05599485710263252, -0.006710824556648731, -0.14664733409881592, 0.10171135514974594, -0.11484641581773758, -0.026557425037026405, -0.022292498499155045, 0.1274915635585785, -0.07724334299564362, -0.1068519577383995, 0.0012902545277029276, -0.03333241119980812, -0.10085515677928925, 0.04268575832247734, -0.1845759004354477, 0.02550840750336647, 0.1051599457859993, 0.026513002812862396, -0.06602542847394943, 0.17287589609622955, 0.07106740772724152, -0.10386008769273758, -0.06685598194599152, 0.16016079485416412, 0.06828390806913376, 0.16988062858581543, 0.028726952150464058, 0.09436644613742828, 0.09969481825828552, -0.06733474880456924, -0.1462719589471817, 0.10320650786161423, 0.03750661760568619, 0.005645993165671825, 0.008071177639067173, 0.13344165682792664, 0.04045069217681885, 0.06666375696659088, 0.06649956107139587, -0.058118272572755814, 0.09459328651428223, 0.03923092409968376, 0.0450824536383152, -0.053891345858573914, 0.15616442263126373, -0.04224146902561188, 0.14687834680080414, 0.010604833252727985, -0.07016229629516602, -0.0046551586128771305, -0.05365341529250145, 0.07478410005569458, -0.0032623473089188337, 0.04887207970023155, -0.06222842261195183, -0.13106288015842438, 0.04698050022125244, -0.07767758518457413, 0.06744438409805298, -0.13976655900478363, -0.029703645035624504, 0.0112156318500638, -0.026335984468460083, -0.07096564024686813, 0.1429329216480255, 0.042186837643384933, -0.006031258963048458, -0.003609568579122424, -0.05395527556538582, 0.01521846279501915, 0.09250084310770035, -0.1080206111073494, -0.004490236286073923 ]
null
null
transformers
# Model name Indian Political Tweets LM Medium (Based on GPT2-Medium) ## Model description This is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this [blog](https://bagdeabhishek.github.io/twitterAnalysis) post. This model is finetuned using GPT2-medium instead of the vanilla GPT2 implementation. This model has more parameters but it is able to model language slightly better. ## Intended uses & limitations This finetuned model can be used to generate tweets which are related to Indian politics. #### How to use ```python from transformers import AutoTokenizer,AutoModelWithLMHead,pipeline tokenizer = AutoTokenizer.from_pretrained("bagdaebhishek/IndianPoliticalTweetsLM") model = AutoModelWithLMHead.from_pretrained("bagdaebhishek/IndianPoliticalTweetsLM") text_generator = pipeline("text-generation",model=model, tokenizer=tokenizer) init_sentence = "India will always be" print(text_generator(init_sentence)) ``` #### Limitations and bias 1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate "Hinglish" text and hence no assumptions should be made about the language of the generated text. 2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like "-sent via NamoApp" etc. 3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model. ## Training data I used the pre-trained gpt2-medium model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a [blog](https://bagdeabhishek.github.io/twitterAnalysis) post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog. ## Training procedure For pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values. I then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles. ### Hardware 1. GPU: GTX 1080Ti 2. CPU: Ryzen 3900x 3. RAM: 32GB This model took roughly 36 hours to fine-tune.
{"language": "en", "license": "apache-2.0", "tags": ["India", "politics", "tweets", "BJP", "Congress", "AAP", "pytorch", "gpt2", "lm-head", "text-generation"], "datasets": ["Twitter", "IndianPolitics"], "thumbnail": "https://bagdeabhishek.github.io/twitterAnalysis_files/networkfin.jpg"}
text-generation
bagdaebhishek/IndianPoliticalTweetsLMMedium
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "India", "politics", "tweets", "BJP", "Congress", "AAP", "lm-head", "en", "dataset:Twitter", "dataset:IndianPolitics", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model name Indian Political Tweets LM Medium (Based on GPT2-Medium) ## Model description This is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post. This model is finetuned using GPT2-medium instead of the vanilla GPT2 implementation. This model has more parameters but it is able to model language slightly better. ## Intended uses & limitations This finetuned model can be used to generate tweets which are related to Indian politics. #### How to use #### Limitations and bias 1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate "Hinglish" text and hence no assumptions should be made about the language of the generated text. 2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like "-sent via NamoApp" etc. 3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model. ## Training data I used the pre-trained gpt2-medium model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog. ## Training procedure For pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values. I then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles. ### Hardware 1. GPU: GTX 1080Ti 2. CPU: Ryzen 3900x 3. RAM: 32GB This model took roughly 36 hours to fine-tune.
[ "# Model name\nIndian Political Tweets LM Medium (Based on GPT2-Medium)", "## Model description\n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post. \n\nThis model is finetuned using GPT2-medium instead of the vanilla GPT2 implementation. This model has more parameters but it is able to model language slightly better.", "## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.", "#### How to use", "#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.", "## Training data\nI used the pre-trained gpt2-medium model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog.", "## Training procedure\n\nFor pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values.\n\nI then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles.", "### Hardware\n1. GPU: GTX 1080Ti\n2. CPU: Ryzen 3900x\n3. RAM: 32GB\n\nThis model took roughly 36 hours to fine-tune." ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model name\nIndian Political Tweets LM Medium (Based on GPT2-Medium)", "## Model description\n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post. \n\nThis model is finetuned using GPT2-medium instead of the vanilla GPT2 implementation. This model has more parameters but it is able to model language slightly better.", "## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.", "#### How to use", "#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.", "## Training data\nI used the pre-trained gpt2-medium model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog.", "## Training procedure\n\nFor pre-processing, I removed tweets from handles which are not very influential in their cluster. I removed them by calculating Eigenvector centrality on the twitter graph and pruning handles which have this measure below a certain threshold. This threshold was set manually after experimenting with different values.\n\nI then separated tweets by these handles based on their language. I trained the LM with English tweets from both handles.", "### Hardware\n1. GPU: GTX 1080Ti\n2. CPU: Ryzen 3900x\n3. RAM: 32GB\n\nThis model took roughly 36 hours to fine-tune." ]
[ 94, 20, 93, 29, 5, 176, 81, 103, 34 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #India #politics #tweets #BJP #Congress #AAP #lm-head #en #dataset-Twitter #dataset-IndianPolitics #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model name\nIndian Political Tweets LM Medium (Based on GPT2-Medium)## Model description\n\nThis is a GPT2 Language model with LM head fine-tuned on tweets crawled from handles which belong predominantly to Indian Politics. For more information about the crawled data, you can go through this blog post. \n\nThis model is finetuned using GPT2-medium instead of the vanilla GPT2 implementation. This model has more parameters but it is able to model language slightly better.## Intended uses & limitations\n This finetuned model can be used to generate tweets which are related to Indian politics.#### How to use#### Limitations and bias\n1. The tweets used to train the model were not manually labelled, so the generated text may not always be in English. I've cleaned the data to remove non-English tweets but the model may generate \"Hinglish\" text and hence no assumptions should be made about the language of the generated text.\n2. I've taken enough care to remove tweets from twitter handles which are not very influential but since it's not curated by hand there might be some artefacts like \"-sent via NamoApp\" etc.\n3. Like any language model trained on real-world data this model also exhibits some biases which unfortunately are a part of the political discourse on Twitter. Please keep this in mind while using the output from this model.## Training data\nI used the pre-trained gpt2-medium model from Huggingface transformers repository and fine-tuned it on custom data set crawled from twitter. The method used to identify the political handles is mentioned in detail in a blog post. I used tweets from both the Pro-BJP and Anti-BJP clusters mentioned in the blog." ]
[ -0.002989761531352997, 0.1391102522611618, -0.00400974927470088, 0.05994769558310509, 0.04881264641880989, 0.008427445776760578, 0.06692702323198318, 0.0480148047208786, -0.00021517382992897183, 0.09455051273107529, 0.023480845615267754, 0.020116383209824562, 0.05129117891192436, 0.13993827998638153, 0.1097373366355896, -0.2943979501724243, 0.05146798491477966, -0.08160051703453064, 0.10149167478084564, 0.11022882908582687, 0.09138058125972748, -0.06834104657173157, 0.047478340566158295, 0.010821709409356117, -0.037808388471603394, 0.022238021716475487, -0.036134809255599976, -0.05458716303110123, 0.06844500452280045, 0.06119782105088234, 0.016974829137325287, -0.021659309044480324, -0.002643557731062174, -0.14509011805057526, 0.02081821858882904, 0.08181802183389664, -0.04732692614197731, 0.012719258666038513, 0.11432171612977982, -0.06759049743413925, 0.2483014464378357, -0.031529929488897324, 0.06293089687824249, 0.09017312526702881, -0.1253403276205063, -0.018368767574429512, -0.12077978998422623, 0.13057710230350494, 0.10512563586235046, 0.10467255860567093, -0.06725659221410751, 0.18404549360275269, -0.03987351059913635, 0.03182247281074524, 0.11019273847341537, -0.16455277800559998, 0.01937953010201454, -0.024142390117049217, -0.0346832275390625, 0.10333801060914993, -0.021959377452731133, 0.03960443288087845, 0.022381817921996117, 0.01677551679313183, -0.06266310065984726, -0.022950608283281326, 0.11638984084129333, -0.015241863206028938, -0.10741625726222992, -0.046387650072574615, 0.04254152625799179, 0.018000751733779907, -0.023958712816238403, -0.17504677176475525, -0.026598583906888962, 0.10064954310655594, 0.0023275362327694893, -0.014622444286942482, -0.005978679284453392, 0.05670089274644852, 0.12331180274486542, -0.11548400670289993, -0.08890452235937119, 0.04856150224804878, 0.010102801956236362, 0.050250034779310226, 0.017077773809432983, 0.015575660392642021, -0.066547691822052, 0.02988201379776001, -0.07473235577344894, -0.08219411969184875, -0.015550022013485432, -0.0386609211564064, -0.07563214004039764, -0.019842632114887238, -0.05301634967327118, -0.05640892684459686, -0.015035726130008698, 0.17999400198459625, -0.059530407190322876, 0.010613864287734032, 0.013326882384717464, 0.011803079396486282, 0.15447591245174408, 0.08856449276208878, -0.15329012274742126, -0.010119262151420116, 0.05150769278407097, -0.12569668889045715, 0.048417527228593826, 0.01671222224831581, 0.039477262645959854, 0.0543486550450325, -0.0034211885649710894, 0.10221454501152039, 0.09556525945663452, 0.040459178388118744, -0.08558136224746704, -0.039528753608465195, 0.17360253632068634, -0.12706372141838074, -0.0009017559350468218, -0.04301163926720619, 0.0013145908014848828, 0.09590570628643036, 0.0014830862637609243, 0.08300287276506424, -0.08404291421175003, 0.002321566455066204, -0.11203449219465256, 0.05351739376783371, -0.03172910213470459, -0.06818003952503204, 0.0005367673002183437, -0.03338354825973511, -0.09157394617795944, -0.11525757610797882, -0.15343515574932098, -0.03275945782661438, -0.036675579845905304, -0.08289824426174164, -0.002135032322257757, 0.026065271347761154, 0.08140096813440323, -0.015530336648225784, 0.0345274955034256, -0.06580784171819687, 0.006084004882723093, 0.0355333536863327, -0.09339901804924011, 0.038301125168800354, 0.10198424756526947, -0.003955100663006306, -0.1948133111000061, 0.06780006736516953, -0.19593439996242523, 0.15403813123703003, -0.03959585353732109, 0.04769004508852959, -0.12579123675823212, 0.006930374074727297, 0.04337697476148605, 0.013216320425271988, -0.07463311403989792, 0.19185347855091095, -0.1446821689605713, -0.026626041159033775, 0.10912007838487625, -0.20899063348770142, 0.014781288802623749, 0.11984435468912125, 0.007744558621197939, 0.15093235671520233, 0.14745652675628662, 0.12502558529376984, -0.005460540764033794, -0.09159598499536514, -0.09009238332509995, -0.13196459412574768, -0.037214357405900955, 0.20680737495422363, 0.04349025711417198, -0.14813658595085144, 0.03254058212041855, -0.0070290034636855125, 0.017120668664574623, 0.012353803031146526, -0.02776542492210865, -0.0168696828186512, 0.03485240787267685, -0.09104269742965698, 0.06095223128795624, -0.005328812170773745, -0.019343573600053787, -0.02579832635819912, -0.14862851798534393, -0.007524627260863781, 0.11511413007974625, -0.007710515521466732, 0.03283960372209549, -0.10953360050916672, -0.012995908968150616, -0.05520417541265488, -0.01661924459040165, -0.2145550549030304, -0.13240806758403778, 0.017505638301372528, -0.11140946298837662, 0.15070605278015137, -0.040905363857746124, 0.002645316766574979, 0.019978096708655357, -0.03412599489092827, -0.0031382283195853233, 0.024757767096161842, 0.031030120328068733, -0.09464681148529053, -0.09569763392210007, -0.029882341623306274, -0.05373957380652428, 0.17941969633102417, -0.06602346152067184, -0.0011622237507253885, 0.09671177715063095, 0.18315312266349792, 0.06669691950082779, -0.08203218132257462, 0.1169324517250061, -0.011760587804019451, 0.028510682284832, -0.08291357010602951, 0.004174451809376478, -0.019827695563435555, 0.051132846623659134, 0.06595726311206818, -0.17248117923736572, -0.1442309021949768, 0.06346004456281662, 0.054470282047986984, -0.09195356070995331, 0.07564544677734375, -0.05565937981009483, -0.01357334852218628, -0.012899966910481453, -0.06203000620007515, 0.07422301173210144, 0.051059823483228683, 0.05442351475358009, -0.040286075323820114, -0.10478436946868896, -0.029779518023133278, -0.04193422198295593, -0.12167630344629288, 0.028018055483698845, 0.12327868491411209, -0.22085285186767578, 0.08153259009122849, 0.07978956401348114, 0.08939811587333679, 0.17313477396965027, 0.07321567088365555, -0.08282069116830826, -0.040401022881269455, -0.09849004447460175, -0.01771712861955166, -0.07952344417572021, 0.04114271700382233, 0.016578489914536476, 0.033500149846076965, 0.012800110504031181, 0.08039882779121399, 0.04085942730307579, 0.024327974766492844, 0.04791846126317978, 0.012441731058061123, 0.0052663045935332775, 0.04142573103308678, -0.04247443377971649, 0.07533586770296097, -0.04717962443828583, 0.14602132141590118, 0.025239506736397743, 0.05541105195879936, -0.18248827755451202, 0.092156320810318, -0.14849109947681427, -0.2937149405479431, -0.04809419438242912, 0.012090908363461494, -0.003654677886515856, -0.018765322864055634, 0.037019792944192886, -0.13967403769493103, -0.12374603748321533, -0.10273890942335129, 0.06824097782373428, 0.05589630454778671, -0.05281835049390793, -0.08781829476356506, -0.02889798954129219, 0.002728239865973592, -0.11034414172172546, 0.011869432404637337, 0.05021660774946213, -0.1088799387216568, 0.06310255080461502, -0.01597384549677372, 0.08864279091358185, 0.14569103717803955, 0.008941948413848877, 0.005704398266971111, 0.01526668295264244, 0.25187695026397705, -0.08991894870996475, 0.18525613844394684, 0.11729718744754791, 0.06486642360687256, 0.052015889436006546, 0.08980744332075119, -0.037294354289770126, -0.06063694879412651, 0.04724343866109848, 0.09621138870716095, -0.009921266697347164, -0.19750849902629852, -0.13470670580863953, 0.030932534486055374, -0.048065267503261566, 0.06809476017951965, 0.062007613480091095, 0.13686440885066986, 0.023509377613663673, -0.1124994084239006, -0.06519865244626999, 0.030686156824231148, 0.048595622181892395, 0.04366225376725197, 0.0017491043545305729, 0.03632931783795357, -0.07902779430150986, 0.06789428740739822, 0.1385451853275299, -0.10364330559968948, 0.12009146064519882, 0.07644851505756378, 0.19177278876304626, 0.06162135303020477, 0.07460355758666992, 0.04281841963529587, 0.04407764971256256, -0.008785117417573929, 0.04001227766275406, -0.007719705812633038, -0.056842006742954254, 0.00936147291213274, 0.09339522570371628, 0.06712237745523453, -0.1206575408577919, 0.026608498767018318, -0.07141011208295822, 0.054299987852573395, 0.11969314515590668, 0.015579468570649624, -0.0798596665263176, -0.10305234789848328, 0.04780776426196098, -0.10815854370594025, -0.019775861874222755, -0.02576046995818615, 0.04933982342481613, -0.13221704959869385, 0.05825728550553322, 0.015475848689675331, 0.05643034726381302, -0.05175009369850159, -0.024480532854795456, -0.00682917982339859, -0.038883283734321594, -0.03345595672726631, 0.0937056615948677, -0.16933521628379822, 0.1140434741973877, 0.03178009018301964, 0.03692929074168205, -0.06927510350942612, -0.021961010992527008, -0.020609024912118912, 0.029154295101761818, 0.09273949265480042, 0.04191553220152855, 0.07395551353693008, -0.06369972974061966, -0.08496695756912231, -0.015531343407928944, 0.07852578908205032, -0.10422627627849579, 0.10285761207342148, -0.0005687003140337765, -0.013244899921119213, -0.04691552743315697, 0.017353078350424767, -0.24323321878910065, -0.17910799384117126, 0.1053442507982254, -0.06030730530619621, 0.08673663437366486, -0.015671173110604286, -0.0457967184484005, -0.0672658309340477, 0.19263286888599396, -0.03794053941965103, -0.16042688488960266, -0.11853989958763123, -0.011795444414019585, 0.12273071706295013, -0.105862557888031, 0.04291948303580284, 0.010327954776585102, 0.15193811058998108, -0.06916141510009766, -0.05190218240022659, 0.012215761467814445, -0.07663767784833908, -0.17989180982112885, 0.029827315360307693, 0.07158156484365463, 0.16434691846370697, 0.04481649771332741, 0.06810323894023895, 0.07017409801483154, 0.006466036196798086, -0.1381697803735733, -0.04278970882296562, 0.07024464756250381, 0.06055758520960808, 0.006343932822346687, -0.00003163814471918158, -0.07915389537811279, -0.12476439028978348, -0.10259135067462921, 0.05855485424399376, 0.22481849789619446, -0.04636342450976372, 0.21115383505821228, 0.21037878096103668, -0.0786287784576416, -0.08151736110448837, -0.12200311571359634, -0.046959057450294495, -0.061851467937231064, 0.11189999431371689, -0.2149144858121872, 0.009275833144783974, 0.12159112840890884, -0.00040124866063706577, -0.10015100240707397, -0.17036864161491394, -0.13512559235095978, 0.05536562576889992, 0.007154644001275301, -0.08105754107236862, -0.10992494970560074, -0.08352608978748322, -0.007272759918123484, -0.14014427363872528, 0.1744208037853241, -0.07426212728023529, -0.03252457454800606, 0.03385959938168526, 0.094179168343544, 0.022857824340462685, -0.039423685520887375, 0.10545918345451355, 0.06557552516460419, 0.03767191246151924, -0.1459491103887558, -0.08289310336112976, 0.14820356667041779, 0.01620163582265377, 0.10571858286857605, 0.10321853309869766, 0.06161649152636528, -0.21775244176387787, -0.09087242931127548, -0.12044559419155121, -0.013027504086494446, -0.07849603146314621, -0.05840493366122246, -0.038536056876182556, 0.12126534432172775, 0.06532308459281921, -0.00346809602342546, -0.08854687958955765, -0.10310307890176773, 0.033410195261240005, 0.15444208681583405, 0.1275254338979721, -0.009533796459436417, -0.02908250130712986, -0.11952783912420273, -0.04432149976491928, 0.0136861028149724, -0.061690520495176315, 0.012541983276605606, 0.0358569361269474, 0.03536859154701233, 0.13138234615325928, -0.09098586440086365, -0.1482999324798584, 0.08081631362438202, 0.014455503784120083, -0.08943012356758118, -0.11430612951517105, -0.014124290086328983, 0.09971943497657776, -0.0013154115295037627, -0.07811854034662247, 0.11583691835403442, -0.0170428566634655, -0.0880630612373352, -0.02778770960867405, 0.05349346995353699, -0.026105426251888275, 0.09792359173297882, 0.043887946754693985, 0.02292880415916443, -0.08072278648614883, 0.08368517458438873, 0.17630736529827118, -0.04377687722444534, 0.021408820524811745, 0.19172324240207672, -0.12298408895730972, -0.10032089054584503, -0.09003186970949173, -0.09075580537319183, -0.07768771052360535, 0.028285987675189972, 0.04097798094153404, 0.03849853575229645, 0.0492696687579155, 0.11427660286426544, -0.0319950096309185, 0.0360155925154686, -0.0360204242169857, -0.0552818588912487, -0.07185199111700058, -0.009022983722388744, 0.10908357799053192, -0.07629459351301193, -0.027569172903895378, 0.14223496615886688, 0.032060422003269196, 0.011948184110224247, -0.015445242635905743, -0.06720757484436035, -0.0065497648902237415, -0.027094779536128044, -0.08339529484510422, 0.007995741441845894, -0.10943075269460678, -0.024335766211152077, -0.027793876826763153, -0.05803493782877922, 0.008357873186469078, -0.008529840037226677, -0.04319830983877182, -0.032222624868154526, -0.027441898360848427, 0.06034133583307266, -0.163791686296463, -0.008031210862100124, 0.12862829864025116, -0.04137193784117699, 0.14155861735343933, 0.055114202201366425, -0.049519870430231094, 0.007497035898268223, -0.08397083729505539, -0.021888146176934242, -0.05558141693472862, -0.018603242933750153, -0.01847446896135807, -0.16123074293136597, -0.04035801813006401, -0.030307788401842117, -0.07157109677791595, 0.036614105105400085, 0.044635191559791565, -0.09851383417844772, 0.13917680084705353, 0.04481767117977142, -0.04115595668554306, -0.09918766468763351, 0.027348116040229797, -0.0006757351220585406, -0.09142038971185684, 0.11787494271993637, -0.048259176313877106, 0.017947817221283913, -0.15127690136432648, 0.02410687878727913, -0.017837505787611008, 0.07847283780574799, -0.03659573569893837, 0.019605666399002075, 0.022819871082901955, 0.02705373242497444, 0.07945732772350311, -0.04169720783829689, -0.004394411109387875, 0.11070086807012558, 0.04602964222431183, -0.04118458181619644, -0.025685567408800125, 0.059863220900297165, 0.015387588180601597, 0.059114739298820496, -0.008644212037324905, -0.04354773461818695, -0.07264290004968643, -0.10644134879112244, 0.12154165655374527, 0.06603142619132996, 0.07913028448820114, -0.02968672662973404, -0.07487618923187256, -0.049344491213560104, -0.09921194612979889, 0.011558928526937962, -0.07081285119056702, -0.05467803031206131, -0.062022220343351364, -0.022213038057088852, 0.2789653539657593, -0.10493551939725876, 0.06829719245433807, 0.04907171055674553, -0.009766452945768833, 0.00032950102468021214, -0.2936602532863617, -0.04088582843542099, 0.028864793479442596, 0.00843770895153284, -0.06595392525196075, 0.11002680659294128, 0.11332976073026657, 0.0564771443605423, 0.009104429744184017, 0.08148498833179474, -0.09793910384178162, -0.1146753579378128, -0.011154997162520885, -0.026135090738534927, -0.005714863538742065, -0.003077525645494461, 0.05493337661027908, -0.08817625045776367, 0.04887096956372261, 0.030942490324378014, 0.10133839398622513, 0.0041761239990592, -0.014900504611432552, -0.08411984145641327, -0.11528672277927399, 0.009196130558848381, 0.026961784809827805, -0.0006399514968506992, 0.15323230624198914, 0.015224494971334934, 0.010193203575909138, 0.015533426776528358, 0.15729059278964996, 0.026924025267362595, -0.14640046656131744, -0.11742671579122543, 0.14390617609024048, -0.031555287539958954, 0.008144805207848549, 0.04098648205399513, -0.1161559596657753, 0.06773656606674194, 0.1347053349018097, 0.11818807572126389, 0.005796114448457956, 0.03886612132191658, -0.06315886229276657, 0.03301118314266205, 0.038118939846754074, 0.04459617659449577, -0.021916206926107407, 0.2319474220275879, -0.1357652097940445, 0.14617186784744263, -0.05096510425209999, -0.0060195582918822765, -0.15303416550159454, 0.09503690153360367, -0.12016758322715759, -0.023487253114581108, -0.02620401419699192, 0.12675787508487701, -0.06854426115751266, -0.1036531999707222, -0.0017349469708278775, -0.027741743251681328, -0.09515392035245895, 0.04621145874261856, -0.18468782305717468, 0.01709670014679432, 0.10306256264448166, 0.02915491908788681, -0.06361912190914154, 0.16455300152301788, 0.07630345225334167, -0.09781799465417862, -0.06981685757637024, 0.16413219273090363, 0.06340280175209045, 0.1708822399377823, 0.02558596432209015, 0.09780203551054001, 0.10065876692533493, -0.05918266251683235, -0.15050584077835083, 0.09692981094121933, 0.03923864662647247, 0.011785931885242462, 0.001835089409723878, 0.12521034479141235, 0.043362654745578766, 0.06757688522338867, 0.06743054836988449, -0.0545102134346962, 0.09673146158456802, 0.03728972375392914, 0.043333444744348526, -0.05416812747716904, 0.15881036221981049, -0.04395286738872528, 0.13941067457199097, -0.0007752873934805393, -0.07293492555618286, -0.0025184026453644037, -0.054238613694906235, 0.07099570333957672, -0.004821837414056063, 0.04779583588242531, -0.06273921579122543, -0.12176823616027832, 0.043629273772239685, -0.07880573719739914, 0.073827363550663, -0.1443142294883728, -0.029534198343753815, 0.015514075756072998, -0.026736775413155556, -0.06828630715608597, 0.14207947254180908, 0.035126544535160065, -0.008062736131250858, -0.008152111433446407, -0.0462801493704319, 0.011538871563971043, 0.08747900277376175, -0.11333750933408737, 0.001738343620672822 ]
null
null
transformers
hello
{}
fill-mask
baicuya/bert_cn
[ "transformers", "pytorch", "bert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
hello
[]
[ "TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 36 ]
[ "passage: TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.06357412785291672, 0.00690077617764473, -0.008467365056276321, 0.020235946401953697, 0.12968459725379944, 0.03302915394306183, 0.09807441383600235, 0.07729126513004303, 0.10806342214345932, -0.009440856985747814, 0.15823203325271606, 0.20325462520122528, -0.03393663093447685, 0.18361465632915497, -0.065280981361866, -0.2617916762828827, 0.06820162385702133, 0.06229938939213753, -0.06546879559755325, 0.11254725605249405, 0.05687131732702255, -0.0869792252779007, 0.07119062542915344, -0.02728140354156494, -0.10962541401386261, 0.04230697825551033, 0.05219662934541702, -0.10162917524576187, 0.12035926431417465, 0.021910926327109337, 0.2123224288225174, 0.016079774126410484, -0.07168376445770264, -0.09227655827999115, 0.046415410935878754, -0.0007212258642539382, -0.07019324600696564, 0.04336128383874893, 0.01872076466679573, -0.07103253155946732, -0.03834306448698044, 0.05254431441426277, 0.027913549914956093, 0.0400172658264637, -0.146876260638237, -0.1195456013083458, -0.012633946724236012, 0.03010692074894905, 0.04268079623579979, 0.060200124979019165, 0.019220897927880287, 0.21056215465068817, -0.12365733087062836, 0.10502377897500992, 0.15344665944576263, -0.3129054605960846, 0.002867099829018116, 0.06838289648294449, 0.07103738188743591, -0.04424971342086792, -0.023489616811275482, 0.05856989696621895, 0.01071459986269474, 0.022148624062538147, 0.044038139283657074, -0.07715853303670883, -0.044328734278678894, 0.01152737345546484, -0.0813736543059349, -0.059741389006376266, 0.16159793734550476, -0.04301191866397858, 0.04482201114296913, 0.012044340372085571, -0.13646052777767181, -0.04112602770328522, -0.0220506452023983, -0.0066766394302248955, -0.034124407917261124, 0.043702688068151474, -0.030891025438904762, -0.01450628973543644, -0.11146465688943863, 0.02620311640202999, -0.2388714849948883, 0.25444263219833374, 0.025913868099451065, 0.06962989270687103, -0.19036757946014404, 0.04825044423341751, -0.032656311988830566, -0.12202991545200348, 0.05392675846815109, -0.09348990768194199, 0.023280160501599312, -0.004289025440812111, -0.06674729287624359, -0.024281397461891174, 0.07810400426387787, 0.19070357084274292, 0.07175330817699432, 0.038729287683963776, 0.022616418078541756, 0.10197576135396957, 0.015252627432346344, 0.0927167534828186, 0.02304348163306713, -0.03693375736474991, 0.058425456285476685, -0.11234977096319199, 0.02645397186279297, -0.06413000077009201, -0.13045111298561096, -0.03655298054218292, 0.026817962527275085, 0.07910523563623428, 0.039121128618717194, 0.05921967700123787, -0.09775126725435257, -0.00039136706618592143, 0.10266957432031631, -0.07596733421087265, 0.011552278883755207, -0.012669426389038563, 0.05071375519037247, 0.10554905235767365, 0.019699934870004654, -0.013951314613223076, -0.02595921792089939, 0.12606576085090637, -0.07414942234754562, -0.0338914580643177, -0.057671189308166504, -0.0717587098479271, 0.04045988991856575, -0.12276265770196915, 0.03537328913807869, -0.18295785784721375, -0.12786880135536194, 0.05937539413571358, 0.05743318796157837, 0.0075002796947956085, -0.02198064886033535, 0.027785688638687134, 0.0016502209473401308, 0.014095489867031574, -0.051991820335388184, -0.05198881775140762, -0.03936923295259476, 0.10392222553491592, 0.01174288708716631, 0.12164439260959625, -0.12003052234649658, 0.04834338277578354, -0.08534543961286545, 0.014317753724753857, -0.15386703610420227, -0.04163011908531189, -0.028065448626875877, 0.1477208137512207, -0.0017313070129603148, -0.044967345893383026, -0.1107422336935997, 0.03536098450422287, -0.008266955614089966, 0.174587219953537, -0.0640043392777443, -0.13445088267326355, 0.238468199968338, -0.10142715275287628, -0.15124346315860748, 0.08353633433580399, 0.00263609504327178, -0.00937309768050909, 0.05675121024250984, 0.109283447265625, 0.03876260668039322, -0.14184610545635223, 0.0926179513335228, 0.11292947828769684, -0.13638317584991455, -0.12760622799396515, 0.022435644641518593, -0.00732642924413085, -0.12322323024272919, 0.04600539803504944, 0.07860185950994492, 0.1112794280052185, -0.07251705229282379, -0.04695776477456093, -0.01390511728823185, -0.03809646517038345, 0.1488271951675415, 0.03689313679933548, 0.09978006780147552, -0.07845763862133026, -0.02166028693318367, -0.028704503551125526, -0.008114258758723736, 0.06035853177309036, 0.038866739720106125, -0.08729325234889984, 0.1360790729522705, -0.0566742941737175, 0.010620499961078167, -0.180439755320549, -0.12009736150503159, -0.0016330704092979431, 0.05382363870739937, -0.027322817593812943, 0.12601551413536072, 0.11395162343978882, -0.03539265692234039, -0.007137839682400227, -0.03099343180656433, 0.09945479035377502, 0.025088751688599586, -0.03798593953251839, -0.0885278731584549, 0.007986658252775669, -0.08452948927879333, -0.014333197847008705, 0.01457307767122984, 0.002566321985796094, 0.00016168280853889883, 0.13817834854125977, -0.0010485704988241196, 0.03795786574482918, -0.05177360028028488, 0.04081299155950546, -0.034957047551870346, 0.01450793631374836, 0.09004251658916473, -0.000576441758312285, -0.06362977623939514, 0.15637962520122528, -0.14581918716430664, 0.35973721742630005, 0.19387078285217285, -0.3088320195674896, -0.016066158190369606, 0.01958218589425087, -0.01481733750551939, -0.0028315566014498472, 0.056414127349853516, -0.015269504860043526, 0.04143389314413071, 0.014644528739154339, 0.15166911482810974, -0.015120322816073895, -0.02077334001660347, 0.027502890676259995, -0.0772947371006012, -0.04431246966123581, 0.03279697522521019, 0.09859511256217957, -0.13104848563671112, 0.17962734401226044, 0.2618531882762909, 0.004645867273211479, 0.13293692469596863, 0.01004520058631897, -0.0017370838904753327, 0.012384308502078056, -0.03448771312832832, -0.02204137109220028, 0.036397550255060196, -0.19078975915908813, -0.037138842046260834, 0.07815047353506088, -0.030133357271552086, 0.05545393377542496, -0.11835511028766632, -0.03323771432042122, 0.029111113399267197, 0.05119411274790764, -0.07707978785037994, 0.12659992277622223, 0.04097466543316841, 0.0710253193974495, 0.0037192106246948242, -0.07951492071151733, 0.11071927845478058, 0.007798798382282257, -0.038606591522693634, 0.15219268202781677, -0.13388566672801971, -0.3540363311767578, -0.1352192908525467, -0.186979740858078, 0.010174541734158993, 0.04617423936724663, 0.07225015014410019, -0.08286191523075104, -0.05899273604154587, 0.09581182152032852, -0.003480511251837015, -0.02892324887216091, 0.06940968334674835, -0.06169416382908821, 0.011217288672924042, -0.027349013835191727, -0.06347832828760147, -0.07560451328754425, -0.028934668749570847, -0.02698061801493168, 0.15005719661712646, -0.09269136935472488, 0.08664495497941971, 0.13057461380958557, 0.0057759047485888, 0.07016542553901672, -0.0002483248827047646, 0.18727800250053406, -0.06556744873523712, -0.005412220023572445, 0.18072476983070374, -0.05880381539463997, 0.1026553139090538, 0.1556575745344162, 0.020712751895189285, -0.05158966779708862, 0.00875561498105526, -0.05700365826487541, -0.11636948585510254, -0.1564129889011383, -0.11075278371572495, -0.13123051822185516, -0.011434734798967838, 0.05559059977531433, 0.04917698726058006, 0.13644592463970184, 0.08514466881752014, 0.03654884546995163, -0.018586870282888412, -0.06805557757616043, 0.0498523935675621, 0.17366138100624084, -0.030056441202759743, 0.1334504783153534, -0.036830224096775055, -0.14371523261070251, 0.059510327875614166, 0.0252390094101429, 0.12022719532251358, 0.10808205604553223, -0.004712763242423534, 0.03895212337374687, 0.16281089186668396, 0.1563887745141983, 0.16660696268081665, 0.025009524077177048, -0.057338543236255646, -0.004954719450324774, -0.009356440976262093, -0.058457158505916595, 0.02018333598971367, 0.15226905047893524, -0.1055486798286438, -0.051534514874219894, -0.145093634724617, 0.05207017809152603, 0.09619975835084915, 0.06738487631082535, -0.22444024682044983, 0.012990519404411316, 0.06385935842990875, 0.007989094592630863, -0.06883342564105988, 0.03757710009813309, -0.02228686586022377, -0.13463854789733887, 0.06749572604894638, -0.05030853673815727, 0.09488040208816528, 0.03667333722114563, 0.07960424572229385, -0.03426273167133331, -0.06298200786113739, 0.04128245636820793, 0.0669965147972107, -0.2517971694469452, 0.2858309745788574, -0.008294520899653435, -0.051533956080675125, -0.08108772337436676, -0.009787647053599358, 0.04465258866548538, 0.12031106650829315, 0.0992002934217453, 0.032960955053567886, -0.021231580525636673, -0.15835201740264893, -0.012746589258313179, 0.028594577684998512, 0.10843918472528458, -0.02854795753955841, -0.016072293743491173, -0.02141297422349453, -0.054353177547454834, -0.007548002991825342, 0.09288700670003891, 0.00021381601982284337, -0.13055965304374695, 0.0781245231628418, 0.056197553873062134, 0.0030072317458689213, -0.010090996511280537, -0.05736343935132027, -0.11168934404850006, 0.18835410475730896, -0.02566578984260559, -0.054508499801158905, -0.10566588491201401, -0.11198879778385162, 0.09742310643196106, -0.10951992124319077, 0.1106313019990921, -0.09603893011808395, 0.004723524209111929, -0.09463068842887878, -0.18368598818778992, 0.1582668572664261, -0.1269671618938446, -0.006225429475307465, -0.07936962693929672, 0.15473303198814392, -0.0639534443616867, 0.02866891399025917, 0.003773587988689542, 0.028899380937218666, -0.10591752827167511, -0.05296826362609863, 0.030782422050833702, -0.05678727477788925, 0.04187817499041557, 0.044521696865558624, -0.06555546075105667, -0.01695936545729637, 0.019335398450493813, 0.04292288422584534, 0.23622342944145203, 0.2353804111480713, -0.052708715200424194, 0.1417168378829956, 0.1806049793958664, -0.028383145108819008, -0.3410240709781647, -0.11411335319280624, -0.13666872680187225, -0.003915437962859869, 0.007809142116457224, -0.1327342689037323, 0.09345895051956177, -0.032195452600717545, -0.04637759178876877, 0.12031539529561996, -0.15053622424602509, -0.09246959537267685, 0.2436363250017166, 0.008315314538776875, 0.4863871932029724, -0.09246446192264557, -0.06652036309242249, -0.03995967283844948, -0.14584210515022278, 0.05183078721165657, 0.024809755384922028, 0.08875752240419388, -0.015901152044534683, 0.08785346150398254, 0.03374331444501877, -0.09186475723981857, 0.09677482396364212, -0.03436388820409775, 0.01234909426420927, -0.10329624265432358, -0.09800854325294495, 0.06808411329984665, -0.01401363592594862, -0.01322801224887371, 0.015540778636932373, 0.007425607182085514, -0.04579975828528404, -0.020523425191640854, -0.10680554807186127, 0.10987795889377594, 0.03320621699094772, -0.062224309891462326, 0.03879779577255249, -0.017917169257998466, -0.009515928104519844, 0.0034782900474965572, 0.1910327970981598, -0.008325624279677868, 0.17571797966957092, 0.08782124519348145, 0.0300945732742548, -0.16413554549217224, -0.0698731392621994, -0.050175994634628296, -0.0846821516752243, 0.08663877099752426, 0.008863678202033043, 0.05756894871592522, 0.11674199998378754, -0.021469993516802788, 0.040903765708208084, 0.11679863929748535, 0.013281558640301228, -0.03635825589299202, 0.15106870234012604, -0.2260168492794037, 0.040877439081668854, -0.024700431153178215, -0.002281648339703679, 0.06495176255702972, 0.0602131113409996, 0.08886897563934326, 0.04362958297133446, -0.03604341670870781, -0.0080631198361516, -0.011103777214884758, -0.059563565999269485, 0.05411487817764282, 0.060502372682094574, 0.05677267909049988, -0.13078919053077698, 0.0061960369348526, -0.020739618688821793, -0.2086004763841629, -0.016145547851920128, 0.07876262068748474, -0.12113361060619354, -0.10942773520946503, 0.0038382872007787228, 0.09838655591011047, -0.08085829019546509, -0.03981052711606026, -0.06243035942316055, -0.11349830776453018, 0.05747007206082344, 0.2176428735256195, 0.1169067993760109, 0.0780315026640892, -0.01989174261689186, -0.01007353700697422, -0.002601395593956113, -0.015962328761816025, 0.02512223646044731, 0.033555783331394196, -0.08247660100460052, 0.01702079549431801, -0.008670814335346222, 0.16094514727592468, -0.11036427319049835, -0.05973701551556587, -0.1687975972890854, 0.04017099365592003, -0.06963387876749039, -0.10318976640701294, -0.09188957512378693, -0.07771022617816925, 0.01973199099302292, -0.07843679189682007, -0.04138858988881111, -0.03797203674912453, -0.1261909008026123, 0.025888055562973022, 0.036669645458459854, -0.015996644273400307, -0.06865283101797104, -0.044388484209775925, 0.13997533917427063, -0.050470441579818726, 0.06897341459989548, 0.14721760153770447, -0.08223868906497955, 0.08987827599048615, -0.11864562332630157, -0.14169776439666748, 0.09844960272312164, 0.024490095674991608, 0.09209379553794861, 0.06073470786213875, 0.01991713047027588, 0.054184310138225555, 0.03840716555714607, 0.039452992379665375, 0.08403609693050385, -0.11287132650613785, 0.06809459626674652, 0.011329096741974354, -0.1869479864835739, -0.02397647127509117, -0.09611000120639801, 0.07828033715486526, 0.0018079385627061129, 0.11844782531261444, -0.0382930189371109, 0.10906048864126205, -0.0436384454369545, 0.014289634302258492, -0.02247670851647854, -0.16372942924499512, -0.004627579357475042, -0.048289380967617035, 0.012862684205174446, -0.013447915203869343, 0.23876222968101501, -0.024661000818014145, 0.024913061410188675, 0.03820062428712845, 0.0719211995601654, -0.003087579505518079, 0.0022083136718720198, 0.15241484344005585, 0.09013786166906357, -0.05284610390663147, -0.0749572142958641, 0.09104806929826736, 0.019679788500070572, -0.05150250345468521, 0.13582676649093628, 0.06253648549318314, 0.04935529828071594, 0.09676174819469452, 0.00193702126853168, 0.04410434886813164, -0.13451460003852844, -0.2456214725971222, -0.04142381623387337, 0.06802476942539215, 0.022965481504797935, 0.02864265814423561, 0.12449731677770615, -0.011933309026062489, 0.057093679904937744, -0.02881103754043579, -0.022149965167045593, -0.1927638053894043, -0.12258896976709366, -0.08218653500080109, -0.07139991223812103, 0.023771436884999275, -0.02313394285738468, -0.020754177123308182, 0.09821733087301254, 0.034732282161712646, -0.026418423280119896, 0.15178021788597107, -0.003468479262664914, -0.011058829724788666, 0.016801699995994568, -0.01001247763633728, 0.0172751322388649, 0.032349079847335815, -0.03294634073972702, -0.16857078671455383, 0.004473234061151743, -0.05259554460644722, 0.0047274017706513405, -0.08785852044820786, 0.02359730750322342, -0.09015554189682007, -0.13330627977848053, -0.07091958820819855, 0.0264219231903553, -0.04996372386813164, 0.09263461828231812, -0.013066912069916725, 0.05031539872288704, 0.0013845227658748627, 0.1200626865029335, -0.07606708258390427, -0.09816689789295197, -0.04547613114118576, 0.1901932656764984, 0.041288163512945175, 0.0920717865228653, -0.015353480353951454, 0.030952494591474533, -0.11943032592535019, 0.34167152643203735, 0.314802885055542, -0.049354273825883865, 0.0750916451215744, 0.054602526128292084, 0.03442682698369026, 0.07451198995113373, 0.1279372125864029, 0.0763775110244751, 0.2879911959171295, -0.09316780418157578, -0.04345858469605446, -0.044293951243162155, -0.03673816844820976, -0.1208759993314743, 0.01128399558365345, 0.03953966125845909, -0.03837299346923828, -0.0634862631559372, 0.07261399179697037, -0.17381651699543, 0.12662146985530853, 0.057949863374233246, -0.21046149730682373, -0.04841303452849388, -0.027771536260843277, 0.17428803443908691, 0.017816947773098946, 0.1136963814496994, -0.03833884000778198, -0.08398560434579849, 0.062350668013095856, 0.022619010880589485, -0.20338550209999084, -0.06756751984357834, 0.10970646142959595, -0.012227135710418224, 0.05940033122897148, -0.017002668231725693, 0.031783878803253174, 0.0780811533331871, 0.07013798505067825, -0.014899644069373608, 0.02075999788939953, 0.023412270471453667, -0.10955478996038437, -0.07060349732637405, 0.01478694099932909, -0.0013840675819665194, -0.11833599954843521, 0.02185012586414814, -0.16461415588855743, 0.04151973873376846, -0.09669603407382965, -0.027114197611808777, -0.0026749002281576395, 0.05793723464012146, -0.04355005547404289, 0.04500356316566467, 0.06464733183383942, 0.018565697595477104, -0.0383153110742569, -0.05022261664271355, -0.011393008753657341, 0.0629846602678299, -0.11954975128173828, -0.17594216763973236, -0.08240210264921188, -0.07172682136297226, 0.04485165327787399, -0.010793168097734451, -0.13988232612609863, -0.04391428083181381, -0.10527841746807098, 0.032555706799030304, -0.15290100872516632, 0.04201599210500717, 0.04696520045399666, 0.04337937757372856, 0.017507996410131454, -0.04434172064065933, 0.04486740753054619, 0.049446675926446915, -0.155558243393898, -0.09162718802690506 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Sinai Voice Arabic Speech Recognition Model # ู†ู…ูˆุฐุฌ **ุตูˆุช ุณูŠู†ุงุก** ู„ู„ุชุนุฑู ุนู„ู‰ ุงู„ุฃุตูˆุงุช ุงู„ุนุฑุจูŠุฉ ุงู„ูุตุญู‰ ูˆ ุชุญูˆูŠู„ู‡ุง ุฅู„ู‰ ู†ุตูˆุต This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - AR dataset. It achieves the following results on the evaluation set: - Loss: 0.2141 - Wer: 0.1808 It achieves the following results on the evaluation set: - eval_loss = 0.2141 - eval_samples = 10388 - eval_wer = 0.181 - eval_cer = 0.049 #### Evaluation Commands 1. To evaluate on `mozilla-foundation/common_voice_8_0` with split `test` ```bash python eval.py --model_id bakrianoo/sinai-voice-ar-stt --dataset mozilla-foundation/common_voice_8_0 --config ar --split test ``` ### Inference Without LM ```python from transformers import (Wav2Vec2Processor, Wav2Vec2ForCTC) import torchaudio import torch def speech_file_to_array_fn(voice_path, resampling_to=16000): speech_array, sampling_rate = torchaudio.load(voice_path) resampler = torchaudio.transforms.Resample(sampling_rate, resampling_to) return resampler(speech_array)[0].numpy(), sampling_rate # load the model cp = "bakrianoo/sinai-voice-ar-stt" processor = Wav2Vec2Processor.from_pretrained(cp) model = Wav2Vec2ForCTC.from_pretrained(cp) # recognize the text in a sample sound file sound_path = './my_voice.mp3' sample, sr = speech_file_to_array_fn(sound_path) inputs = processor([sample], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values,).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) ``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 32 - eval_batch_size: 10 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - total_train_batch_size: 256 - total_eval_batch_size: 80 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 1.354 | 0.64 | 1000 | 0.4109 | 0.4493 | | 0.5886 | 1.28 | 2000 | 0.2798 | 0.3099 | | 0.4977 | 1.92 | 3000 | 0.2387 | 0.2673 | | 0.4253 | 2.56 | 4000 | 0.2266 | 0.2523 | | 0.3942 | 3.2 | 5000 | 0.2171 | 0.2437 | | 0.3619 | 3.84 | 6000 | 0.2076 | 0.2253 | | 0.3245 | 4.48 | 7000 | 0.2088 | 0.2186 | | 0.308 | 5.12 | 8000 | 0.2086 | 0.2206 | | 0.2881 | 5.76 | 9000 | 0.2089 | 0.2105 | | 0.2557 | 6.4 | 10000 | 0.2015 | 0.2004 | | 0.248 | 7.04 | 11000 | 0.2044 | 0.1953 | | 0.2251 | 7.68 | 12000 | 0.2058 | 0.1932 | | 0.2052 | 8.32 | 13000 | 0.2117 | 0.1878 | | 0.1976 | 8.96 | 14000 | 0.2104 | 0.1825 | | 0.1845 | 9.6 | 15000 | 0.2156 | 0.1821 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.2+cu113 - Datasets 1.18.3 - Tokenizers 0.11.0
{"language": ["ar"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "hf-asr-leaderboard", "robust-speech-event"], "datasets": ["mozilla-foundation/common_voice_8_0"], "metrics": ["wer", "cer"], "widget": [{"example_title": "Example 1", "src": "https://huggingface.co/bakrianoo/sinai-voice-ar-stt/raw/main/examples/common_voice_ar_19077324.mp3"}, {"example_title": "Example 2", "src": "https://huggingface.co/bakrianoo/sinai-voice-ar-stt/raw/main/examples/common_voice_ar_19205138.mp3"}, {"example_title": "Example 3", "src": "https://huggingface.co/bakrianoo/sinai-voice-ar-stt/raw/main/examples/common_voice_ar_19331711.mp3"}], "model-index": [{"name": "Sinai Voice Arabic Speech Recognition Model", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice ar", "type": "mozilla-foundation/common_voice_8_0", "args": "ar"}, "metrics": [{"type": "wer", "value": 0.181, "name": "Test WER"}, {"type": "cer", "value": 0.049, "name": "Test CER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Dev Data", "type": "speech-recognition-community-v2/dev_data", "args": "ar"}, "metrics": [{"type": "wer", "value": 93.03, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Test Data", "type": "speech-recognition-community-v2/eval_data", "args": "ar"}, "metrics": [{"type": "wer", "value": 90.79, "name": "Test WER"}]}]}]}
automatic-speech-recognition
bakrianoo/sinai-voice-ar-stt
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "hf-asr-leaderboard", "robust-speech-event", "ar", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "model-index", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ar" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #hf-asr-leaderboard #robust-speech-event #ar #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us
Sinai Voice Arabic Speech Recognition Model =========================================== ู†ู…ูˆุฐุฌ ุตูˆุช ุณูŠู†ุงุก ู„ู„ุชุนุฑู ุนู„ู‰ ุงู„ุฃุตูˆุงุช ุงู„ุนุฑุจูŠุฉ ุงู„ูุตุญู‰ ูˆ ุชุญูˆูŠู„ู‡ุง ุฅู„ู‰ ู†ุตูˆุต ==================================================================== This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - AR dataset. It achieves the following results on the evaluation set: * Loss: 0.2141 * Wer: 0.1808 It achieves the following results on the evaluation set: * eval\_loss = 0.2141 * eval\_samples = 10388 * eval\_wer = 0.181 * eval\_cer = 0.049 #### Evaluation Commands 1. To evaluate on 'mozilla-foundation/common\_voice\_8\_0' with split 'test' ### Inference Without LM ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0002 * train\_batch\_size: 32 * eval\_batch\_size: 10 * seed: 42 * distributed\_type: multi-GPU * num\_devices: 8 * total\_train\_batch\_size: 256 * total\_eval\_batch\_size: 80 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1000 * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.2+cu113 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "#### Evaluation Commands\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_8\\_0' with split 'test'", "### Inference Without LM", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 10\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 8\n* total\\_train\\_batch\\_size: 256\n* total\\_eval\\_batch\\_size: 80\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #hf-asr-leaderboard #robust-speech-event #ar #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n", "#### Evaluation Commands\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_8\\_0' with split 'test'", "### Inference Without LM", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 10\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 8\n* total\\_train\\_batch\\_size: 256\n* total\\_eval\\_batch\\_size: 80\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ 92, 36, 8, 179, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #hf-asr-leaderboard #robust-speech-event #ar #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n#### Evaluation Commands\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_8\\_0' with split 'test'### Inference Without LM### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 10\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 8\n* total\\_train\\_batch\\_size: 256\n* total\\_eval\\_batch\\_size: 80\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.2+cu113\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ -0.07657581567764282, 0.09032603353261948, -0.0042309509590268135, 0.05508272349834442, 0.10443264991044998, 0.03557080402970314, 0.10543101280927658, 0.15268485248088837, -0.09164413809776306, 0.10575651377439499, 0.061715368181467056, 0.06479708105325699, 0.06960375607013702, 0.09450079500675201, 0.0012167611857876182, -0.24638894200325012, 0.00905015878379345, -0.04616115614771843, -0.11928088217973709, 0.10216759145259857, 0.07889760285615921, -0.11078083515167236, 0.04279434680938721, -0.014176810160279274, -0.08645614981651306, 0.009946408681571484, -0.05849927291274071, -0.01845879666507244, 0.09696632623672485, 0.045798324048519135, 0.03600461408495903, 0.011093701235949993, 0.11185608059167862, -0.31972143054008484, -0.0038079002406448126, 0.08670640736818314, 0.03042152337729931, 0.032149337232112885, 0.09824543446302414, 0.01391951646655798, 0.13214199244976044, -0.10055360943078995, 0.05635514110326767, 0.03671970218420029, -0.1016610786318779, -0.24294231832027435, -0.07242432236671448, 0.002898666076362133, 0.1051449179649353, 0.08846265077590942, -0.036911047995090485, 0.07213367521762848, -0.10320831835269928, 0.08766398578882217, 0.22678892314434052, -0.2233080267906189, -0.06958600133657455, 0.01612573303282261, 0.028885526582598686, 0.031860072165727615, -0.10812357813119888, -0.00647739926353097, 0.024184243753552437, 0.017709527164697647, 0.08230189979076385, 0.00882295798510313, -0.020042626187205315, 0.023357747122645378, -0.12985624372959137, -0.05083763971924782, 0.11455445736646652, 0.06921770423650742, 0.013995383866131306, -0.10294576734304428, -0.025470353662967682, -0.16375292837619781, -0.057808998972177505, 0.019146708771586418, 0.04532060772180557, -0.042088523507118225, -0.06227777898311615, 0.06362740695476532, -0.03833821788430214, -0.07708162069320679, 0.04385584220290184, 0.09658727049827576, 0.0535646490752697, -0.036582157015800476, -0.0028542897198349237, 0.08196567744016647, 0.02876381389796734, -0.1626141220331192, -0.008972841314971447, 0.042324770241975784, -0.12813998758792877, -0.01769522950053215, -0.01796582154929638, 0.029712192714214325, 0.07423362135887146, 0.1286213994026184, -0.02267143875360489, 0.09677302837371826, 0.024518992751836777, -0.0025155721232295036, -0.06329155713319778, 0.1515936702489853, -0.08642809092998505, -0.10572181642055511, -0.0663510337471962, 0.11160394549369812, -0.004490857943892479, -0.008888459764420986, -0.05067490041255951, 0.06350363045930862, 0.07569635659456253, 0.05862833186984062, -0.0019044347573071718, 0.05917923152446747, -0.04470477253198624, -0.0062356432899832726, -0.04729953035712242, -0.13454359769821167, 0.062281254678964615, 0.06417901813983917, -0.060443904250860214, -0.03541925549507141, -0.019341301172971725, 0.0019515680614858866, -0.032326098531484604, 0.10421953350305557, -0.055123813450336456, 0.006286412477493286, -0.09313364326953888, -0.11237255483865738, 0.03567671775817871, 0.004403820261359215, -0.007417249958962202, -0.04897259175777435, -0.05949905142188072, -0.060244787484407425, 0.07357298582792282, -0.06952749937772751, -0.03774808347225189, -0.08458385616540909, -0.08603139966726303, 0.04549981653690338, 0.0024349570740014315, 0.17801886796951294, -0.0661689043045044, 0.07480760663747787, 0.025905344635248184, 0.05416853353381157, 0.12130053341388702, 0.0625448152422905, -0.04513410106301308, 0.05627184361219406, -0.12715579569339752, 0.08253172039985657, -0.10472901910543442, 0.06357765197753906, -0.12557968497276306, -0.102846659719944, -0.012167664244771004, 0.014658532105386257, 0.0989704579114914, 0.1193402037024498, -0.22177813947200775, -0.06493774056434631, 0.17727293074131012, -0.04515921697020531, -0.09669356793165207, 0.13062213361263275, -0.02207808755338192, -0.05028857663273811, 0.023866774514317513, 0.18192575871944427, 0.10061625391244888, -0.10253770649433136, -0.0036379604134708643, -0.050046972930431366, 0.119105763733387, 0.07747037708759308, 0.06464824080467224, -0.039484407752752304, 0.06110584735870361, 0.009376692585647106, -0.0434114933013916, 0.03499862551689148, -0.07476949691772461, -0.08988011628389359, -0.0285054761916399, -0.07894778996706009, -0.01837373524904251, 0.06321872770786285, -0.000572698307223618, -0.09409870952367783, -0.13216833770275116, 0.014641638845205307, 0.08734065294265747, -0.10138846933841705, 0.023333260789513588, -0.07310591638088226, 0.03572910279035568, 0.019039448350667953, 0.025716714560985565, -0.14622923731803894, -0.037319328635931015, 0.03273959457874298, -0.07474265992641449, 0.019184188917279243, 0.003993045073002577, 0.058786243200302124, 0.04241405799984932, -0.04437188059091568, -0.02632896974682808, -0.014521462842822075, 0.007362987846136093, -0.05142584815621376, -0.2571045160293579, -0.062274474650621414, -0.011660205200314522, 0.13078303635120392, -0.19650505483150482, 0.0030905830208212137, 0.03577008098363876, 0.10681106895208359, -0.009622056037187576, -0.041643138974905014, -0.01299602072685957, 0.06131836399435997, -0.023383434861898422, -0.0786890760064125, 0.018365463241934776, -0.007745895069092512, -0.0891970694065094, -0.018824927508831024, -0.15786303579807281, 0.057231515645980835, 0.09733035415410995, -0.00563231622800231, -0.08284346759319305, 0.011653493158519268, -0.04949689283967018, -0.05544543266296387, -0.03819706663489342, -0.011328313499689102, 0.1620776355266571, 0.00906751211732626, 0.08064993470907211, -0.06904830783605576, -0.06501010060310364, 0.015949472784996033, 0.015062564052641392, -0.011452023871243, 0.1604960560798645, 0.09932894259691238, -0.0625065490603447, 0.0942259132862091, 0.03979262337088585, -0.06656281650066376, 0.13505297899246216, -0.06802387535572052, -0.06525076180696487, -0.05186258256435394, 0.023385295644402504, 0.019879110157489777, 0.11718352138996124, -0.15911367535591125, -0.014568797312676907, 0.016152549535036087, 0.00925968773663044, 0.020764857530593872, -0.16592884063720703, -0.0021281708031892776, 0.04400818794965744, -0.0823122188448906, -0.01973581872880459, -0.02260306105017662, 0.0007020302582532167, 0.0876639187335968, -0.0036111713852733374, -0.05243226885795593, -0.04291301593184471, -0.048975735902786255, -0.09453525394201279, 0.17600342631340027, -0.09726009517908096, -0.11299572139978409, -0.07410936057567596, -0.01770712248980999, -0.027237452566623688, -0.028108024969697, 0.04380379989743233, -0.11998434364795685, -0.060127850621938705, -0.08661525696516037, 0.006055304780602455, -0.026005819439888, 0.009533017873764038, 0.03401387855410576, 0.018880223855376244, 0.06933887302875519, -0.111849844455719, 0.0036248902324587107, -0.005041956435889006, -0.022070063278079033, 0.006065316032618284, 0.04418681189417839, 0.11126048117876053, 0.16250717639923096, 0.0443250797688961, 0.038116440176963806, -0.002689301036298275, 0.1821487993001938, -0.11212331801652908, 0.001330857165157795, 0.05973144993185997, 0.009510698728263378, 0.04063180088996887, 0.18357855081558228, 0.05626852065324783, -0.09486686438322067, 0.014318562112748623, 0.06943750381469727, -0.027602674439549446, -0.2077612578868866, -0.039030496031045914, -0.07360143214464188, -0.03476259484887123, 0.09540173411369324, 0.020453032106161118, -0.042866695672273636, -0.00877696368843317, -0.011938835494220257, 0.002296007238328457, 0.037530165165662766, 0.03777075558900833, 0.06757395714521408, 0.045816123485565186, 0.10759273171424866, -0.0252122450619936, -0.04014614596962929, 0.021256232634186745, -0.03313227370381355, 0.23126263916492462, -0.010662619024515152, 0.1574566513299942, 0.06565346568822861, 0.1310717612504959, -0.022321907803416252, 0.03799699246883392, 0.0031226102728396654, -0.005187176633626223, 0.030736293643712997, -0.06405724585056305, -0.0028623477555811405, 0.030731836333870888, 0.09369818866252899, 0.004420063458383083, -0.07626163959503174, 0.024667782709002495, 0.07184601575136185, 0.3030412197113037, 0.06443647295236588, -0.25585296750068665, -0.04438209533691406, 0.02824913151562214, -0.0579596571624279, -0.015027256682515144, 0.022274326533079147, 0.12918081879615784, -0.0634031891822815, 0.08448775112628937, -0.03877278417348862, 0.08006654679775238, -0.05804282799363136, 0.03815807029604912, 0.11812260746955872, 0.09912710636854172, 0.013316606171429157, 0.07256904244422913, -0.27768373489379883, 0.25425195693969727, -0.006507040932774544, 0.07281335443258286, -0.06910941749811172, 0.04518909752368927, 0.018972473219037056, -0.040712758898735046, 0.06815483421087265, -0.005068522412329912, -0.15186889469623566, -0.1671096384525299, -0.053103383630514145, 0.026589857414364815, 0.12899935245513916, -0.03652513772249222, 0.13429220020771027, -0.03958156704902649, -0.04473857954144478, 0.044479843229055405, -0.02697846293449402, -0.11968458443880081, -0.12267335504293442, 0.03759638965129852, -0.022414270788431168, 0.007219810038805008, -0.07715179026126862, -0.07454296946525574, -0.11722952872514725, 0.187474325299263, -0.05753326416015625, -0.01442353893071413, -0.11310924589633942, 0.06537085026502609, 0.16348755359649658, -0.05700508505105972, 0.025505492463707924, 0.027437016367912292, 0.10236711800098419, 0.02700837142765522, -0.031295254826545715, 0.1085369661450386, -0.08427136391401291, -0.17971792817115784, -0.04609197378158569, 0.13964618742465973, 0.03101714700460434, 0.07029113173484802, -0.019222324714064598, 0.030723674222826958, -0.001408178941346705, -0.09846996515989304, 0.05547778308391571, 0.05555843561887741, 0.015567511320114136, 0.017963005229830742, -0.0665302574634552, -0.0005566817708313465, -0.07057849317789078, -0.05191197246313095, 0.09606144577264786, 0.2487606555223465, -0.09254445880651474, 0.03719232976436615, 0.0075993286445736885, -0.0837925523519516, -0.1674000322818756, -0.011372637934982777, 0.08173248916864395, 0.020722320303320885, 0.029860105365514755, -0.17270633578300476, 0.03644279018044472, 0.05931837111711502, -0.006206832360476255, 0.0733267068862915, -0.3492928147315979, -0.1352686733007431, 0.07267671078443527, 0.05196914076805115, -0.0958239808678627, -0.17255444824695587, -0.05934348702430725, -0.0406835563480854, -0.0804869681596756, 0.052215758711099625, -0.024592431262135506, 0.12070602923631668, 0.008606556802988052, 0.011428844183683395, 0.0303990226238966, -0.05794304236769676, 0.14315654337406158, 0.013874056749045849, 0.06549437344074249, -0.03513815253973007, -0.0037572337314486504, 0.02577170915901661, -0.0643041804432869, 0.015379193238914013, -0.10748542845249176, 0.021913249045610428, -0.10114699602127075, -0.01804167591035366, -0.07717286795377731, -0.0020626664627343416, -0.06702125817537308, -0.01982205919921398, -0.043399207293987274, 0.054713934659957886, 0.11562270671129227, 0.0006838043336756527, 0.07266970723867416, -0.0490235909819603, 0.13007807731628418, 0.10562587529420853, 0.1008814349770546, 0.016805674880743027, -0.08078955113887787, 0.021503334864974022, 0.027826225385069847, 0.02412397228181362, -0.11979518830776215, 0.05791228637099266, 0.14620450139045715, 0.04698693007230759, 0.1280544102191925, 0.04880320653319359, -0.0775841772556305, -0.02481212094426155, 0.06281224638223648, -0.07995137572288513, -0.1444299817085266, 0.004085242282599211, -0.019572392106056213, -0.12312386184930801, -0.03640138357877731, 0.08414742350578308, -0.032446522265672684, 0.012190721929073334, 0.014797372743487358, 0.07484034448862076, -0.042320940643548965, 0.2273721694946289, 0.0098348930478096, 0.09575264900922775, -0.09341619163751602, 0.07364865392446518, 0.05382050946354866, -0.0830540657043457, 0.04933929070830345, 0.09117956459522247, -0.04793192818760872, -0.027142589911818504, 0.08524295687675476, 0.10211998224258423, 0.07585502415895462, -0.040505893528461456, -0.1272127628326416, -0.1628359854221344, 0.08628345280885696, 0.05801261588931084, 0.026224683970212936, 0.029269736260175705, -0.017725510522723198, 0.029488088563084602, -0.09527899324893951, 0.12371276319026947, 0.08797446638345718, 0.04801823943853378, -0.09310534596443176, 0.11544780433177948, 0.009282062761485577, -0.007487835828214884, 0.006216499954462051, 0.032912932336330414, -0.14091715216636658, 0.027080431580543518, -0.10656078904867172, -0.026875630021095276, -0.04610422998666763, 0.0050836605951189995, 0.012648341245949268, -0.030381830409169197, -0.06068504601716995, 0.01232460793107748, -0.09932460635900497, -0.06683018058538437, -0.034075621515512466, 0.08331625163555145, -0.11846327036619186, -0.014126433059573174, 0.03429326415061951, -0.12829099595546722, 0.06877287477254868, 0.016085680574178696, 0.020590441301465034, 0.016283897683024406, -0.09634114801883698, -0.01558087207376957, 0.012245794758200645, 0.024299373850226402, 0.035994697362184525, -0.19783873856067657, -0.016135238111019135, -0.02558901347219944, 0.022040091454982758, 0.0035158235114067793, 0.00436070142313838, -0.116175077855587, 0.03676241636276245, -0.047698717564344406, -0.0689326822757721, -0.02431745082139969, 0.04876415431499481, 0.07835349440574646, -0.00088979781139642, 0.1444806605577469, -0.07728474587202072, 0.059764474630355835, -0.23640799522399902, -0.012976786121726036, 0.008058315142989159, -0.05599905923008919, -0.040248773992061615, -0.018448282033205032, 0.11231228709220886, -0.07284950464963913, 0.13368916511535645, -0.009051400236785412, 0.03558525815606117, 0.032790590077638626, -0.0679808184504509, 0.01320799719542265, 0.06764554232358932, 0.1554892361164093, 0.013101465068757534, -0.020860999822616577, 0.05958360433578491, -0.024512356147170067, 0.05086086690425873, 0.10926096141338348, 0.17784146964550018, 0.15044431388378143, 0.07813188433647156, 0.0626552626490593, 0.0859089121222496, -0.1660057157278061, -0.09435054659843445, 0.18517497181892395, -0.07917356491088867, 0.15717075765132904, -0.052067190408706665, 0.1637234091758728, 0.08849511295557022, -0.2145218700170517, 0.08552998304367065, -0.07745300978422165, -0.09885230660438538, -0.09280449151992798, -0.07464408129453659, -0.0869988203048706, -0.1623050421476364, 0.021530024707317352, -0.11196425557136536, 0.08345857262611389, 0.07904820889234543, 0.043596189469099045, 0.03611737862229347, 0.12484053522348404, 0.038329750299453735, 0.01872207038104534, 0.08477562665939331, 0.01743837259709835, -0.04076812043786049, -0.03602585569024086, -0.06319321691989899, 0.03262753039598465, -0.024779507890343666, 0.05675657466053963, -0.021406779065728188, -0.09750577062368393, 0.06537340581417084, 0.007813150994479656, -0.07946223765611649, 0.023816687986254692, -0.044748708605766296, 0.05682768672704697, 0.07300166040658951, 0.0463794507086277, -0.014227897860109806, -0.0159011110663414, 0.1837586909532547, -0.09275011718273163, -0.10168072581291199, -0.12363812327384949, 0.17980994284152985, -0.0009220996871590614, 0.011221825145184994, 0.01936241239309311, -0.07120997458696365, -0.016324274241924286, 0.1815592497587204, 0.18293032050132751, -0.03975851461291313, -0.014801868237555027, 0.013696329668164253, -0.006523041985929012, -0.027394283562898636, 0.07310264557600021, 0.11810895055532455, 0.048049814999103546, -0.04160446301102638, -0.023226387798786163, -0.0270302165299654, -0.0733174979686737, -0.032806605100631714, 0.11217530071735382, 0.04683631286025047, 0.006018387619405985, -0.009746859781444073, 0.10155071318149567, -0.07330720871686935, -0.15102073550224304, 0.04634859412908554, -0.19135083258152008, -0.17589859664440155, -0.017399998381733894, 0.08984849601984024, 0.013981365598738194, 0.05423649773001671, -0.007153186947107315, -0.03783582150936127, 0.109415203332901, 0.004978798795491457, -0.052666764706373215, -0.11827678233385086, 0.05739504098892212, -0.1031879410147667, 0.1797551065683365, -0.03442363813519478, 0.03790980577468872, 0.12641505897045135, 0.03404548764228821, -0.07761593163013458, 0.050456538796424866, 0.09215115755796432, -0.1469467580318451, 0.04018501937389374, 0.14597617089748383, -0.0490664467215538, 0.13576127588748932, 0.05969597026705742, -0.057490430772304535, 0.03668128699064255, -0.03844418376684189, -0.02865198627114296, -0.07480023801326752, -0.01080635841935873, -0.052478671073913574, 0.14397652447223663, 0.2323364019393921, -0.06262334436178207, 0.004561317153275013, -0.044621456414461136, 0.02738707698881626, -0.011146831326186657, 0.14973869919776917, -0.05893935263156891, -0.2831801772117615, 0.045525964349508286, -0.021630611270666122, 0.038727208971977234, -0.17466777563095093, -0.09907108545303345, 0.06649232655763626, -0.055825140327215195, -0.061618249863386154, 0.12612690031528473, 0.11706379055976868, 0.04279889911413193, -0.0696672722697258, -0.16974620521068573, -0.012941638939082623, 0.18100802600383759, -0.1640734076499939, -0.06253158301115036 ]
null
null
transformers
## Arabic T5 Base Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for `google/mt5-base` model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 ``` T5 is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. The T5 model was presented in Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer by Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu. ``` [Read More](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html)
{"language": "Arabic", "license": "apache-2.0", "datasets": ["mc4"]}
text2text-generation
bakrianoo/t5-arabic-base
[ "transformers", "pytorch", "t5", "text2text-generation", "dataset:mc4", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "Arabic" ]
TAGS #transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## Arabic T5 Base Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-base' model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 Read More
[ "## Arabic T5 Base Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-base' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Arabic T5 Base Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-base' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ 63, 54, 7 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Arabic T5 Base Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-base' model, as it's much smaller and only targets Arabic and English based tasks.### About T5\n\n\n\nRead More" ]
[ -0.011279740370810032, -0.0841737762093544, -0.001882073120214045, 0.08252676576375961, 0.08490898460149765, 0.05583036318421364, 0.2091376632452011, 0.09918879717588425, -0.05184024199843407, -0.11732640862464905, 0.12092927098274231, 0.06653232127428055, 0.0290500707924366, 0.16440564393997192, 0.007034891285002232, -0.2456861138343811, -0.027312979102134705, -0.01697174273431301, 0.05092895030975342, 0.13478289544582367, 0.17964215576648712, 0.036950353533029556, 0.13509266078472137, -0.04006176441907883, -0.06584754586219788, 0.09449739009141922, 0.025554824620485306, -0.15146677196025848, 0.09017748385667801, 0.15549218654632568, 0.03617245331406593, 0.08664398640394211, 0.003505234606564045, -0.06381615996360779, 0.04051463305950165, 0.02700125239789486, -0.1335991621017456, -0.002542640781030059, 0.02887718752026558, -0.02834269590675831, 0.2482786327600479, -0.04194650053977966, -0.02614537626504898, 0.02037177048623562, -0.04904146119952202, 0.028623636811971664, 0.06922350823879242, 0.13772012293338776, 0.06839507073163986, 0.07942678034305573, 0.011131754145026207, 0.10990538448095322, -0.13611651957035065, 0.17265081405639648, 0.1496959626674652, -0.33293190598487854, -0.08532551676034927, 0.10172339528799057, -0.004357258323580027, 0.16074125468730927, 0.08134429901838303, 0.08065705746412277, -0.016187049448490143, -0.000924939988180995, 0.0239527877420187, -0.04964565113186836, -0.14504362642765045, 0.03183477371931076, -0.06546960026025772, -0.09840169548988342, 0.370536744594574, 0.019587840884923935, 0.009493605233728886, -0.05410878360271454, -0.1099313497543335, 0.04605572670698166, -0.030363138765096664, -0.07136545330286026, -0.01509412843734026, 0.059095483273267746, 0.0495050810277462, -0.09506616741418839, -0.11812463402748108, -0.05356551706790924, -0.23135149478912354, 0.04207098111510277, 0.04032502695918083, 0.050685521215200424, -0.13064569234848022, 0.11703819036483765, 0.03849095478653908, -0.06922096759080887, 0.02639729529619217, -0.11361520737409592, 0.02724558673799038, -0.031620632857084274, -0.024026907980442047, -0.12187749147415161, 0.060722969472408295, 0.025446828454732895, 0.11499571800231934, 0.07694705575704575, -0.008487858809530735, 0.026469551026821136, -0.0346870981156826, -0.000918508623726666, -0.08863528072834015, -0.06778830289840698, 0.07327058166265488, -0.028480341657996178, -0.05080507695674896, 0.029810471460223198, -0.11271366477012634, -0.08053690940141678, 0.037965014576911926, 0.06051440164446831, 0.01667449250817299, 0.18162491917610168, 0.02873397246003151, -0.06569571793079376, 0.016584008932113647, -0.12600156664848328, -0.1066652163863182, -0.025462035089731216, -0.06631889194250107, -0.009683891199529171, 0.1334926187992096, -0.010225190781056881, -0.21901217103004456, -0.113002248108387, -0.01443235669285059, -0.05523927882313728, 0.017355170100927353, -0.015699690207839012, 0.001257237046957016, -0.012407081201672554, 0.010721856728196144, -0.22258655726909637, -0.22694090008735657, 0.07458215206861496, 0.08660164475440979, -0.01328213233500719, -0.08771203458309174, -0.0473017543554306, 0.020501891151070595, 0.01639450713992119, -0.0918886587023735, 0.15587694942951202, -0.07184100896120071, 0.07456273585557938, 0.008190592750906944, 0.0778363049030304, -0.10707631707191467, 0.02553216554224491, -0.0413641482591629, -0.09055274724960327, -0.06887892633676529, 0.1590956151485443, 0.01776408776640892, 0.0480671264231205, -0.13606679439544678, 0.011708314530551434, -0.09199808537960052, 0.02213691920042038, -0.027753563597798347, 0.1851729303598404, -0.0958828404545784, -0.0725393071770668, 0.14362558722496033, -0.03657408803701401, -0.2729325294494629, 0.09510409832000732, 0.019411398097872734, 0.14807748794555664, 0.14062002301216125, 0.08352074027061462, 0.003611744847148657, -0.05724991858005524, 0.07238193601369858, 0.14847871661186218, -0.09660397469997406, -0.049953363835811615, 0.034064389765262604, 0.06771966069936752, -0.17235057055950165, 0.04950176179409027, -0.03095787949860096, 0.1017833799123764, -0.02089446410536766, -0.06123539060354233, -0.0501088947057724, -0.07797501236200333, 0.05574817955493927, 0.05296723544597626, 0.12884362041950226, -0.038995321840047836, -0.09446027874946594, -0.02404973655939102, -0.013303135521709919, 0.017537837848067284, 0.03706720843911171, -0.10453379154205322, 0.015956701710820198, -0.009029818698763847, 0.06957267969846725, -0.1471664011478424, -0.08080978691577911, -0.07943716645240784, 0.15366987884044647, 0.08699265867471695, 0.004028343595564365, 0.06371872872114182, -0.06651084125041962, -0.13558444380760193, 0.027711186558008194, 0.08672846853733063, 0.03531251847743988, -0.06475762277841568, -0.12641577422618866, 0.05657607689499855, 0.03187141567468643, -0.038839444518089294, 0.0339343436062336, 0.026273077353835106, -0.05220935493707657, 0.15115886926651, -0.016774868592619896, 0.12833529710769653, 0.017277877777814865, -0.019585931673645973, 0.01199563778936863, -0.021554382517933846, 0.06983395665884018, -0.01302889920771122, -0.10386278480291367, 0.13984963297843933, -0.10119441151618958, 0.30124109983444214, 0.1948740929365158, -0.08305475115776062, -0.020747704431414604, 0.032874416559934616, -0.032763972878456116, -0.01493983343243599, 0.0766388401389122, 0.015793487429618835, 0.10259401798248291, -0.059012558311223984, 0.16552482545375824, -0.15184402465820312, -0.014115417376160622, 0.006663486361503601, 0.005429105833172798, -0.02382463961839676, 0.042504969984292984, 0.06537861377000809, -0.2727132737636566, 0.04396692290902138, 0.17077237367630005, 0.0707794800400734, 0.16559244692325592, 0.023995524272322655, -0.0417700856924057, 0.07656482607126236, 0.016195235773921013, -0.05700457841157913, -0.03958753123879433, -0.1964525729417801, -0.013804838992655277, 0.04309382289648056, 0.016288893297314644, 0.06252805143594742, 0.003079604357481003, 0.001296272617764771, -0.00488964281976223, -0.06463406980037689, -0.0776224359869957, 0.08633311837911606, -0.014423203654587269, 0.15588730573654175, 0.007808658294379711, -0.015915973111987114, 0.05361909046769142, 0.013760121539235115, -0.13688893616199493, 0.1719190925359726, -0.02616027370095253, -0.2255045473575592, 0.04531301185488701, -0.09451980888843536, -0.07907453924417496, 0.009859182871878147, 0.12053143978118896, -0.06162967160344124, -0.015584632754325867, -0.1204768717288971, 0.019377177581191063, -0.0010121666127815843, 0.06998643279075623, -0.04482690989971161, 0.013728736899793148, -0.012039502151310444, -0.12671330571174622, -0.015952106565237045, 0.014375879429280758, -0.08809196203947067, 0.049352265894412994, -0.2861323356628418, -0.007318598683923483, 0.050126731395721436, -0.046922728419303894, 0.1046794131398201, -0.10037644952535629, 0.22453650832176208, -0.04946758970618248, 0.10736020654439926, 0.13504073023796082, 0.07182471454143524, -0.0059289271011948586, 0.1708841770887375, -0.029690558090806007, -0.061364445835351944, 0.09637078642845154, 0.012429301626980305, -0.04968548193573952, -0.2498077154159546, -0.05890802666544914, -0.10697708278894424, -0.0029859142377972603, 0.033182334154844284, 0.07795616239309311, 0.11240316927433014, 0.05753704532980919, -0.07554841786623001, 0.12862657010555267, 0.07962954789400101, 0.025739142671227455, 0.18594171106815338, 0.000549464370124042, 0.06763992458581924, -0.14298351109027863, -0.03505910560488701, 0.12693104147911072, 0.04050407558679581, 0.0822548195719719, -0.022969428449869156, 0.04776041954755783, 0.05102244019508362, 0.047986164689064026, 0.0681900605559349, 0.15774479508399963, 0.04452883452177048, -0.0002521454880479723, -0.0269954614341259, -0.04480640962719917, -0.05246490612626076, 0.01828528381884098, 0.0006904929177835584, -0.12753348052501678, -0.05564044415950775, 0.044759828597307205, 0.09543439745903015, 0.1509070247411728, -0.055665768682956696, -0.22733750939369202, -0.013252639211714268, 0.0006727337022311985, -0.09881635755300522, -0.10664113610982895, 0.0418589748442173, -0.06334653496742249, -0.12612496316432953, 0.12034804373979568, 0.05767183005809784, 0.13124310970306396, 0.08047956973314285, -0.0010697576217353344, 0.0013751190854236484, -0.1247919499874115, -0.010850762948393822, 0.09900282323360443, -0.42961058020591736, 0.24408550560474396, 0.05551448464393616, 0.03511003777384758, -0.11619412899017334, -0.021377429366111755, 0.05832155793905258, 0.18078821897506714, 0.16242845356464386, 0.0052633811719715595, 0.04428588226437569, 0.09517037123441696, -0.07008830457925797, 0.10167320817708969, 0.040015578269958496, -0.021176308393478394, -0.041974976658821106, -0.043040961027145386, -0.004044918343424797, 0.022989386692643166, 0.16399292647838593, -0.23092138767242432, -0.10399206727743149, 0.031031399965286255, 0.019553761929273605, 0.005590556189417839, -0.031372714787721634, -0.0672987625002861, -0.013450099155306816, 0.09454470872879028, -0.0457427054643631, -0.1724327802658081, -0.12266401201486588, -0.006496292073279619, 0.017092004418373108, -0.07015641778707504, 0.026952549815177917, -0.047851815819740295, -0.092616505920887, 0.060456570237874985, -0.22750252485275269, 0.10939661413431168, -0.060437142848968506, 0.012541545554995537, -0.009802219457924366, 0.10490556806325912, -0.1049964427947998, 0.019093461334705353, 0.007543528918176889, -0.10099303722381592, -0.050981633365154266, -0.02864799275994301, -0.08074863255023956, -0.09160604327917099, -0.000617935904301703, 0.02116849273443222, -0.054515741765499115, -0.12275870889425278, 0.01386331208050251, -0.07599730789661407, 0.12185356020927429, -0.08454909175634384, 0.029597297310829163, 0.13470418751239777, 0.16661065816879272, -0.0237587857991457, -0.2155478298664093, -0.2056579887866974, 0.01419626921415329, 0.004937086720019579, -0.061633504927158356, -0.1015874370932579, 0.07181445509195328, -0.0809120312333107, 0.02457759529352188, -0.07841179519891739, -0.2544960379600525, -0.13762414455413818, 0.13882088661193848, 0.05279134586453438, 0.2510119378566742, -0.11854290217161179, -0.0074575962498784065, -0.04526703432202339, -0.14588670432567596, 0.05097723752260208, -0.1991158127784729, 0.09558060020208359, -0.013896038755774498, -0.008160068653523922, -0.024101831018924713, 0.039273958653211594, -0.007269109133630991, 0.0033010125625878572, -0.020131589844822884, -0.17238479852676392, 0.025154806673526764, 0.10082748532295227, -0.0019170970190316439, 0.10239268839359283, -0.11376851797103882, 0.09618410468101501, -0.12841184437274933, -0.11159872263669968, -0.06970587372779846, -0.013158737681806087, -0.00007678650581510738, -0.06629430502653122, 0.03373180329799652, -0.07103921473026276, 0.10290345549583435, -0.023297296836972237, 0.04310102388262749, -0.06872764229774475, 0.013560215011239052, 0.08176182955503464, 0.2723994255065918, -0.225944384932518, 0.06782165169715881, -0.038589734584093094, -0.028881946578621864, 0.05117475241422653, -0.29091304540634155, 0.006362374871969223, 0.05691090226173401, -0.08319342881441116, 0.014630129560828209, 0.05837110057473183, 0.04163747280836105, -0.05806400999426842, 0.1512989103794098, -0.11634907871484756, -0.08372052758932114, -0.10725852847099304, -0.11366786807775497, -0.03449277579784393, -0.016479529440402985, 0.08896811306476593, -0.03826497867703438, -0.005313552916049957, -0.028053078800439835, -0.06891829520463943, -0.06861342489719391, 0.06801124662160873, 0.04803993180394173, -0.0031561616342514753, -0.08147823065519333, 0.14121344685554504, 0.07113310694694519, -0.08300714939832687, 0.010081212967634201, 0.18140168488025665, -0.18724533915519714, -0.11729328334331512, 0.0638485997915268, 0.07001930475234985, -0.06164645776152611, -0.07063630223274231, -0.0589679516851902, -0.10223285853862762, 0.06703508645296097, 0.17842017114162445, 0.0191471166908741, 0.051592204719781876, -0.026059038937091827, -0.04736024886369705, 0.08093684911727905, 0.10188673436641693, -0.01084146648645401, -0.0342111811041832, -0.0786418542265892, 0.01871773786842823, -0.011735735461115837, 0.14311324059963226, -0.08396437019109726, 0.02191193401813507, -0.08260861039161682, 0.052467137575149536, -0.16101272404193878, -0.009716627188026905, -0.0442974679172039, 0.005977270193397999, -0.027823815122246742, -0.0646202564239502, -0.08310718089342117, 0.018756354227662086, -0.0427958108484745, 0.02009117417037487, 0.02092522569000721, 0.1086656004190445, -0.03316228836774826, 0.035028308629989624, 0.005716521758586168, -0.0034796760883182287, 0.15654875338077545, 0.07618547230958939, -0.1250695437192917, 0.12169504910707474, -0.1569618284702301, -0.019976967945694923, -0.037371233105659485, 0.054754871875047684, -0.0057352278381586075, 0.0362386554479599, 0.0882883369922638, 0.1083870679140091, 0.042339298874139786, 0.03579520434141159, 0.04094478487968445, -0.04501597210764885, -0.010337477549910545, -0.06166723370552063, 0.006896688137203455, -0.04183468222618103, 0.03179300203919411, 0.06254833936691284, 0.01734534278512001, 0.04383916035294533, -0.06134793907403946, -0.09476430714130402, -0.1261630654335022, 0.03194862976670265, 0.01748098060488701, -0.07375702261924744, -0.09564246237277985, -0.13179057836532593, 0.03163621202111244, -0.04753345251083374, 0.20142875611782074, 0.06021791696548462, 0.003544887062162161, 0.06414905190467834, 0.08933359384536743, 0.09552568942308426, -0.03000827133655548, 0.26657265424728394, -0.023120790719985962, 0.00335136242210865, -0.038939908146858215, -0.023887882009148598, -0.015880119055509567, 0.03430059552192688, 0.1141551285982132, 0.06883986294269562, -0.02326042391359806, 0.08679328113794327, -0.04588527977466583, 0.03878745809197426, -0.004160616546869278, -0.13870705664157867, 0.021610373631119728, 0.04653049632906914, -0.07555048167705536, -0.16110242903232574, 0.2547989785671234, 0.0024952527601271868, -0.003946607932448387, 0.00843628030270338, -0.032719437032938004, -0.11357580870389938, -0.14303778111934662, -0.07315358519554138, -0.13870695233345032, -0.004293263424187899, -0.11069285869598389, -0.03162560239434242, 0.059070903807878494, 0.09342964738607407, -0.051899105310440063, 0.1510789692401886, 0.011395955458283424, -0.1213088408112526, 0.08406117558479309, -0.05057273805141449, 0.034535910934209824, -0.010464799590408802, 0.021736642345786095, -0.06567069888114929, -0.02953995391726494, -0.06810873001813889, 0.014046444557607174, -0.07682443410158157, 0.04010128602385521, -0.06851367652416229, -0.008445467799901962, -0.0753321424126625, 0.02868032269179821, -0.011094380170106888, 0.18701595067977905, 0.022771870717406273, -0.08555369079113007, 0.04838884621858597, 0.19897952675819397, 0.02188664861023426, -0.2505197525024414, -0.13868016004562378, 0.06595776975154877, 0.06032866984605789, -0.0031690967734903097, 0.04365231469273567, 0.010346473194658756, -0.05426613241434097, 0.2805064022541046, 0.20443052053451538, -0.04171840846538544, 0.008754394948482513, 0.011400248855352402, 0.0331055223941803, 0.0705580934882164, 0.17210714519023895, 0.07468283176422119, 0.1732313185930252, -0.04669888690114021, 0.020628880709409714, 0.014707966707646847, 0.025991544127464294, -0.07411763072013855, 0.14774107933044434, 0.06829655170440674, -0.09816563129425049, 0.014385721646249294, 0.10484669357538223, -0.11586452275514603, 0.15770989656448364, -0.09151743352413177, -0.015168051235377789, -0.04289352521300316, -0.055257927626371384, 0.03204479441046715, 0.08836760371923447, -0.03645322844386101, -0.0006751185865141451, 0.019713720306754112, 0.003382898634299636, 0.05095536634325981, -0.23629577457904816, -0.06117987260222435, 0.12098101526498795, -0.02867109701037407, 0.09907914698123932, 0.010264407843351364, 0.11002473533153534, 0.04224000871181488, 0.007312259636819363, -0.061750248074531555, 0.1306876838207245, 0.027101831510663033, 0.10782993584871292, 0.11007601767778397, -0.061906322836875916, -0.003391172969713807, -0.06679008156061172, 0.015844188630580902, 0.08513961732387543, 0.05002057924866676, 0.11725210398435593, 0.020763207226991653, -0.06339160352945328, 0.11616992205381393, -0.09582357853651047, 0.06803972274065018, 0.10943561792373657, -0.030147617682814598, 0.029805578291416168, -0.11603843420743942, 0.01579844206571579, 0.034826286137104034, -0.11685089021921158, -0.05127272754907608, -0.01799451932311058, -0.04593181982636452, -0.050208814442157745, -0.0014047378208488226, -0.12268681824207306, 0.0021356649231165648, -0.10921042412519455, -0.018322007730603218, -0.14022071659564972, 0.1126849427819252, 0.18588176369667053, 0.0014660839224234223, 0.024653084576129913, -0.09610879421234131, 0.010344542562961578, 0.0829080119729042, -0.15168194472789764, -0.12630705535411835 ]
null
null
transformers
## Arabic T5 Large Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for `google/mt5-large` model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 ``` T5 is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. The T5 model was presented in Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer by Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu. ``` [Read More](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html)
{"language": "Arabic", "license": "apache-2.0", "datasets": ["mc4"]}
text2text-generation
bakrianoo/t5-arabic-large
[ "transformers", "pytorch", "t5", "text2text-generation", "dataset:mc4", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "Arabic" ]
TAGS #transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## Arabic T5 Large Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-large' model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 Read More
[ "## Arabic T5 Large Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-large' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Arabic T5 Large Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-large' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ 63, 55, 7 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Arabic T5 Large Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-large' model, as it's much smaller and only targets Arabic and English based tasks.### About T5\n\n\n\nRead More" ]
[ -0.016571247950196266, -0.10427235066890717, -0.0018624559743329883, 0.09119723737239838, 0.0916118249297142, 0.05487321689724922, 0.19583851099014282, 0.09539815038442612, -0.043594200164079666, -0.09470641613006592, 0.12559238076210022, 0.060031723231077194, 0.02532513625919819, 0.16865073144435883, 0.014568627811968327, -0.25764137506484985, -0.03976505994796753, -0.022655755281448364, 0.01865382120013237, 0.1276499330997467, 0.17682893574237823, 0.014135772362351418, 0.14831644296646118, -0.03837936744093895, -0.07574829459190369, 0.07232680916786194, 0.017792342230677605, -0.15288423001766205, 0.0841435119509697, 0.1539250910282135, 0.020557589828968048, 0.08010245114564896, -0.0008064936846494675, -0.03646570444107056, 0.03764007240533829, 0.015116386115550995, -0.11355780065059662, 0.000513062346726656, 0.053509604185819626, 0.020141160115599632, 0.25319305062294006, -0.07819817960262299, -0.03159060701727867, 0.026706404983997345, -0.02308592200279236, 0.0019238179083913565, 0.060799241065979004, 0.13163641095161438, 0.04150219261646271, 0.07687948644161224, 0.021197356283664703, 0.12625691294670105, -0.137799933552742, 0.1602812260389328, 0.14729472994804382, -0.32966917753219604, -0.08015987277030945, 0.11451461166143417, 0.0031562147196382284, 0.16064047813415527, 0.07875250279903412, 0.07721012085676193, -0.008006935939192772, 0.004175079055130482, 0.04638931155204773, -0.04874751344323158, -0.15405544638633728, 0.019218308851122856, -0.06583143025636673, -0.09663940966129303, 0.3639489710330963, 0.02749723754823208, 0.02081811986863613, -0.09480459243059158, -0.10717214643955231, 0.059318721294403076, -0.027786022052168846, -0.06406586617231369, 0.0006764684803783894, 0.034620754420757294, 0.03661181032657623, -0.09920211881399155, -0.12822090089321136, -0.045813389122486115, -0.26585525274276733, 0.02150665409862995, 0.03287389874458313, 0.05423644185066223, -0.15547962486743927, 0.08338619023561478, 0.012827586382627487, -0.060168441385030746, 0.026098433881998062, -0.11349594593048096, 0.02508704364299774, -0.034327998757362366, -0.014388165436685085, -0.12775912880897522, 0.07925983518362045, -0.024733740836381912, 0.09903909265995026, 0.046825673431158066, 0.010789105668663979, 0.030574098229408264, -0.0282808318734169, 0.019331704825162888, -0.11679064482450485, -0.058810509741306305, 0.06343746185302734, -0.024276692420244217, -0.05084136873483658, 0.03274647891521454, -0.11593650281429291, -0.09304226189851761, 0.014751134440302849, 0.06124958023428917, 0.012343022972345352, 0.1742481291294098, 0.04572710022330284, -0.06397563219070435, 0.02937839925289154, -0.11519134044647217, -0.09686553478240967, -0.020149607211351395, -0.06675749272108078, 0.027784250676631927, 0.10216395556926727, -0.031555112451314926, -0.20375922322273254, -0.09035506844520569, -0.022308163344860077, -0.053633518517017365, 0.023429160937666893, -0.021579254418611526, 0.01780831627547741, 0.0025799633003771305, 0.012149718590080738, -0.23110587894916534, -0.1977834701538086, 0.0857929140329361, 0.05504428595304489, -0.010199636220932007, -0.09513673186302185, -0.017068255692720413, 0.01562659814953804, 0.009044941514730453, -0.09393613040447235, 0.17469343543052673, -0.06353069841861725, 0.07940652966499329, 0.004174795001745224, 0.09559986740350723, -0.11144942790269852, 0.017898177728056908, -0.04234058037400246, -0.0947207361459732, -0.04745502769947052, 0.14946241676807404, 0.042637068778276443, 0.04985157027840614, -0.13337863981723785, -0.0106264129281044, -0.08611659705638885, 0.026605332270264626, -0.04219596087932587, 0.16469113528728485, -0.11694026738405228, -0.0729578360915184, 0.18744100630283356, -0.018191449344158173, -0.2951696813106537, 0.1200331449508667, 0.03420589119195938, 0.13223591446876526, 0.14419996738433838, 0.06828756630420685, -0.0040273517370224, -0.04606962203979492, 0.03928020969033241, 0.15840522944927216, -0.10501063615083694, -0.07201886177062988, 0.05524305999279022, 0.06086784228682518, -0.1443510204553604, 0.044020891189575195, -0.006435643415898085, 0.10418249666690826, -0.02478129416704178, -0.0588054433465004, -0.05226950719952583, -0.09082692861557007, 0.044665224850177765, 0.04699272662401199, 0.13430282473564148, -0.05421175807714462, -0.07878060638904572, -0.03247614949941635, -0.0037660901434719563, 0.026554638519883156, 0.025089869275689125, -0.10922091454267502, 0.014354219660162926, -0.023623473942279816, 0.07261508703231812, -0.14037218689918518, -0.08063511550426483, -0.06970790773630142, 0.12094864994287491, 0.07738693803548813, 0.04599924013018608, 0.06852768361568451, -0.0681157112121582, -0.1272004097700119, 0.04872212931513786, 0.09262312948703766, 0.017658306285738945, -0.06505464762449265, -0.096336230635643, 0.06278204172849655, 0.03411658853292465, -0.05098223686218262, 0.018561555072665215, 0.01644258387386799, -0.03171868249773979, 0.11408339440822601, -0.005851098336279392, 0.1360359936952591, 0.017997225746512413, -0.027083884924650192, 0.006329535506665707, -0.014176566153764725, 0.06322219967842102, -0.024501938372850418, -0.12168554961681366, 0.1419852077960968, -0.11141658574342728, 0.32226505875587463, 0.1982252150774002, -0.05720900371670723, -0.011675284244120121, 0.027300337329506874, -0.033114243298769, -0.008101016283035278, 0.06242635101079941, 0.005246460437774658, 0.06660349667072296, -0.053267329931259155, 0.17394545674324036, -0.15610072016716003, -0.011647550389170647, 0.021726422011852264, 0.016978399828076363, -0.01852267049252987, 0.04904913157224655, 0.07677683979272842, -0.27225261926651, 0.040502145886421204, 0.1934185028076172, 0.05938588082790375, 0.18904924392700195, 0.02772417850792408, -0.0559290274977684, 0.08119550347328186, 0.008976416662335396, -0.049478862434625626, -0.05659711733460426, -0.17225879430770874, 0.0002627770882099867, 0.0651385635137558, 0.02373451180756092, 0.04552335664629936, 0.015329512767493725, -0.003392571583390236, -0.0002360544167459011, -0.05782410129904747, -0.0781022310256958, 0.10292757302522659, 0.011886727064847946, 0.16348913311958313, 0.006925015710294247, -0.013038082979619503, 0.05465874820947647, 0.014318717643618584, -0.12258365005254745, 0.1690017431974411, -0.0025779325515031815, -0.2534443140029907, 0.03737817704677582, -0.07445012032985687, -0.08492384105920792, -0.003614382352679968, 0.11516532301902771, -0.0701681450009346, -0.009158993139863014, -0.12115101516246796, 0.048624321818351746, 0.0033132792450487614, 0.07996848970651627, -0.04109141603112221, 0.023972120136022568, -0.01356789842247963, -0.12360307574272156, -0.013601699844002724, 0.015232045203447342, -0.10829385370016098, 0.052659302949905396, -0.2594371736049652, 0.006867832504212856, 0.05144600570201874, -0.05113615095615387, 0.09051212668418884, -0.09704969823360443, 0.1984615921974182, -0.05603049695491791, 0.09635146707296371, 0.17236535251140594, 0.09497814625501633, -0.0064519247971475124, 0.15948060154914856, -0.017421500757336617, -0.055177804082632065, 0.08580166101455688, 0.004227897152304649, -0.055107079446315765, -0.24899986386299133, -0.07153704017400742, -0.11605407297611237, 0.0305634792894125, 0.040166765451431274, 0.06559629738330841, 0.08063215762376785, 0.060500986874103546, -0.074571892619133, 0.12573270499706268, 0.05592600628733635, 0.018052471801638603, 0.18491588532924652, 0.008211636915802956, 0.06524582207202911, -0.15616969764232635, -0.04542393237352371, 0.15083912014961243, 0.02309579774737358, 0.08006562292575836, -0.035615649074316025, 0.08233576267957687, 0.04616711661219597, 0.014069495722651482, 0.06314339488744736, 0.17101600766181946, 0.03265382722020149, -0.016999205574393272, -0.033480726182460785, -0.05151180922985077, -0.040062785148620605, 0.010662347078323364, 0.034820154309272766, -0.10638737678527832, -0.07553442567586899, 0.06918686628341675, 0.09819664806127548, 0.1518770158290863, -0.047539472579956055, -0.24755774438381195, -0.017413988709449768, -0.005606822669506073, -0.09830791503190994, -0.10243106633424759, 0.04382053762674332, -0.004587023518979549, -0.13356877863407135, 0.12393933534622192, 0.05655202269554138, 0.12321791052818298, 0.08075343817472458, 0.014930603094398975, -0.001349115278571844, -0.10035715997219086, 0.0023875355254858732, 0.10851440578699112, -0.420401006937027, 0.23613929748535156, 0.05616152286529541, 0.013284965418279171, -0.14350268244743347, -0.024788036942481995, 0.06386784464120865, 0.15727946162223816, 0.15940217673778534, 0.0029889834113419056, 0.08198267221450806, 0.09433137625455856, -0.08275769650936127, 0.09452316910028458, 0.03246288374066353, -0.0187448188662529, -0.04322538897395134, -0.05040226876735687, -0.006508077494800091, 0.008396966382861137, 0.1726144254207611, -0.1703965812921524, -0.11886405944824219, 0.026589132845401764, 0.0501648485660553, -0.040677718818187714, -0.02607153356075287, -0.06838149577379227, -0.03486667573451996, 0.08966238796710968, -0.03733529895544052, -0.16490884125232697, -0.1090945303440094, 0.000974251888692379, 0.03491734340786934, -0.0817577913403511, 0.018496539443731308, -0.0550866536796093, -0.06878548115491867, 0.04949183017015457, -0.1991252899169922, 0.11101481318473816, -0.06804034113883972, -0.0019926601089537144, -0.0029022928792983294, 0.10020405799150467, -0.13624367117881775, 0.03519933670759201, 0.0035016732290387154, -0.10783268511295319, -0.05489255487918854, -0.036959052085876465, -0.08951026201248169, -0.06149744242429733, 0.010560737922787666, 0.038358382880687714, -0.04173973202705383, -0.11889224499464035, 0.03728499636054039, -0.07707816362380981, 0.15798208117485046, -0.05758656561374664, 0.015607042238116264, 0.12041343748569489, 0.15037524700164795, 0.005359051749110222, -0.22565029561519623, -0.20306424796581268, 0.006430498324334621, 0.0056969644501805305, -0.03595290333032608, -0.07262276113033295, 0.0886496901512146, -0.048368655145168304, 0.03313227742910385, -0.07170798629522324, -0.28610777854919434, -0.13040630519390106, 0.10520370304584503, 0.02347513847053051, 0.25252240896224976, -0.11865987628698349, -0.00023681111633777618, -0.05648228898644447, -0.10863452404737473, 0.05973055213689804, -0.19294844567775726, 0.1015247106552124, -0.024273144081234932, 0.009148422628641129, -0.01277117058634758, 0.02562977746129036, 0.0009408178739249706, -0.016392245888710022, -0.02822449617087841, -0.17314250767230988, 0.02547486498951912, 0.10691846907138824, -0.00517189409583807, 0.12416323274374008, -0.14243046939373016, 0.08425338566303253, -0.1150394082069397, -0.11952260881662369, -0.08775753527879715, -0.012670686468482018, 0.0007303992751985788, -0.06503934413194656, 0.022769132629036903, -0.06469851732254028, 0.10004549473524094, -0.020596692338585854, 0.018682993948459625, -0.058550260961055756, -0.01786726713180542, 0.0525948591530323, 0.2957804501056671, -0.23675574362277985, 0.02894793637096882, -0.04684901610016823, -0.022913580760359764, 0.05623181164264679, -0.28312575817108154, 0.017786841839551926, 0.06390636414289474, -0.08045411109924316, -0.008154889568686485, 0.0608949139714241, 0.049819957464933395, -0.050285350531339645, 0.15174426138401031, -0.09383444488048553, -0.11264821887016296, -0.0930573120713234, -0.05280061066150665, -0.056006234139204025, -0.02752218022942543, 0.07691127806901932, -0.031315430998802185, -0.0035126516595482826, -0.025301679968833923, -0.05995837599039078, -0.04795437678694725, 0.07571437954902649, 0.03800085932016373, 0.006229020655155182, -0.08850076794624329, 0.11487068980932236, 0.07352665811777115, -0.10885778069496155, -0.0013864723732694983, 0.190243661403656, -0.17647796869277954, -0.10957816243171692, 0.0820271223783493, 0.03730010241270065, -0.06043335422873497, -0.0681823194026947, -0.03557947278022766, -0.10388772934675217, 0.06841260194778442, 0.1511266827583313, 0.029509812593460083, 0.03979799151420593, -0.023038823157548904, -0.05571926757693291, 0.05476319044828415, 0.10700546950101852, -0.03158188611268997, -0.021040964871644974, -0.09696604311466217, 0.03997325152158737, -0.02498188056051731, 0.1627577245235443, -0.07617050409317017, 0.043207358568906784, -0.06891924887895584, 0.03600723296403885, -0.1633528769016266, -0.003996447194367647, -0.04272545501589775, 0.0018691113218665123, -0.0352630577981472, -0.038404934108257294, -0.0797145664691925, 0.0017288601957261562, -0.04868362843990326, 0.013969761319458485, 0.01909796893596649, 0.12882737815380096, -0.02147953025996685, 0.03111201338469982, 0.0011271892581135035, 0.0008442804682999849, 0.17188182473182678, 0.04523833468556404, -0.12173406779766083, 0.10175406187772751, -0.12517715990543365, -0.04680626466870308, -0.027204403653740883, 0.07292920351028442, 0.012603022158145905, 0.051838457584381104, 0.09469954669475555, 0.11148618161678314, 0.05662021040916443, 0.04348533973097801, 0.010398264043033123, -0.04640117287635803, 0.011812854558229446, -0.061174504458904266, 0.008495105430483818, -0.04304351285099983, 0.01684608682990074, 0.03891802206635475, 0.008802497759461403, 0.057839103043079376, -0.05032883957028389, -0.08757288008928299, -0.13857820630073547, 0.03477800637483597, -0.008503432385623455, -0.08178551495075226, -0.0870116800069809, -0.11018878221511841, 0.04578316956758499, -0.0476580485701561, 0.21564216911792755, 0.06738082319498062, -0.01916300505399704, 0.06572912633419037, 0.09219318628311157, 0.09918300807476044, -0.044700853526592255, 0.26871544122695923, -0.017697099596261978, 0.0011595888063311577, -0.021268852055072784, -0.02506042830646038, -0.0018383350688964128, 0.08600360155105591, 0.15554088354110718, 0.08587879687547684, -0.032899774610996246, 0.07888132333755493, -0.028479792177677155, 0.03173893317580223, -0.02039317786693573, -0.1642986536026001, 0.015562855638563633, 0.05384515970945358, -0.074895940721035, -0.16945555806159973, 0.25117698311805725, 0.0028964232187718153, -0.008251333609223366, -0.01208037044852972, -0.01721927523612976, -0.1074594110250473, -0.1220509260892868, -0.08773881196975708, -0.1098615899682045, -0.007729560602456331, -0.10777245461940765, -0.033642545342445374, 0.07353280484676361, 0.09574545174837112, -0.0608515590429306, 0.15206953883171082, -0.029635010287165642, -0.11182419210672379, 0.08048421889543533, -0.05077800527215004, 0.033096976578235626, 0.008351746015250683, 0.014631175436079502, -0.07476827502250671, -0.024249421432614326, -0.056608181446790695, 0.016107913106679916, -0.08253056555986404, 0.027280151844024658, -0.0810982882976532, -0.01058740820735693, -0.08953768014907837, 0.018487975001335144, 0.007364092860370874, 0.15837977826595306, 0.024628490209579468, -0.09223507344722748, 0.03601473569869995, 0.17924942076206207, 0.012504211626946926, -0.2542932331562042, -0.12246623635292053, 0.051720231771469116, 0.046217549592256546, -0.00391010008752346, 0.03523476421833038, 0.011594525538384914, -0.07998430728912354, 0.2879999279975891, 0.22813932597637177, -0.07098105549812317, 0.00856140349060297, 0.0036403476260602474, 0.03394482284784317, 0.04717728868126869, 0.18559162318706512, 0.08299695700407028, 0.2120201289653778, -0.040563128888607025, 0.03201352432370186, 0.02210698276758194, 0.04826951399445534, -0.08430695533752441, 0.1481267809867859, 0.06690198928117752, -0.08314298093318939, 0.027842536568641663, 0.09577258676290512, -0.09691278636455536, 0.17229197919368744, -0.10096560418605804, -0.03388713672757149, -0.05851970613002777, -0.0394325815141201, 0.03523727133870125, 0.09174706786870956, -0.02363574504852295, -0.009779749438166618, 0.017708923667669296, -0.03560552000999451, 0.040488436818122864, -0.2545657157897949, -0.044646281749010086, 0.12730762362480164, -0.013865849934518337, 0.1174241155385971, 0.014636709354817867, 0.12073647975921631, 0.05337867885828018, -0.01756996661424637, -0.07000009715557098, 0.12067621201276779, 0.03072447143495083, 0.11557935178279877, 0.11543702334165573, -0.05226045101881027, -0.014179810881614685, -0.08346209675073624, 0.030610810965299606, 0.055919189006090164, 0.023390892893075943, 0.1363278478384018, 0.03723253309726715, -0.07808561623096466, 0.12399927526712418, -0.08655139803886414, 0.06903620064258575, 0.13214384019374847, -0.03674895316362381, 0.01539516169577837, -0.10794678330421448, 0.02741534262895584, 0.03882104530930519, -0.10371556878089905, -0.07382391393184662, -0.01576019451022148, -0.05805116519331932, -0.09157773107290268, 0.002901799976825714, -0.13769172132015228, 0.0038317807484418154, -0.10822220146656036, -0.014746425673365593, -0.1432350128889084, 0.09530976414680481, 0.18255984783172607, -0.0013900100020691752, 0.02910681441426277, -0.07311820983886719, -0.0060699633322656155, 0.0698511078953743, -0.1661369949579239, -0.12720909714698792 ]
null
null
transformers
## Arabic T5 Small Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for `google/mt5-small` model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 ``` T5 is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. The T5 model was presented in Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer by Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu. ``` [Read More](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html)
{"language": "Arabic", "license": "apache-2.0", "datasets": ["mc4"]}
text2text-generation
bakrianoo/t5-arabic-small
[ "transformers", "pytorch", "t5", "text2text-generation", "dataset:mc4", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "Arabic" ]
TAGS #transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## Arabic T5 Small Model A customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-small' model, as it's much smaller and only targets Arabic and English based tasks. ### About T5 Read More
[ "## Arabic T5 Small Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-small' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Arabic T5 Small Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-small' model, as it's much smaller and only targets Arabic and English based tasks.", "### About T5\n\n\n\nRead More" ]
[ 63, 55, 7 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #dataset-mc4 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Arabic T5 Small Model\n\nA customized T5 Model for Arabic and English Task. It could be used as an alternative for 'google/mt5-small' model, as it's much smaller and only targets Arabic and English based tasks.### About T5\n\n\n\nRead More" ]
[ -0.010144056752324104, -0.12374555319547653, -0.002215723041445017, 0.09161084145307541, 0.08470426499843597, 0.041951734572649, 0.22133511304855347, 0.11206141859292984, -0.0556921623647213, -0.10254856199026108, 0.11652994155883789, 0.06226419284939766, 0.03536907583475113, 0.15238969027996063, 0.0068307253532111645, -0.25131112337112427, -0.03969539329409599, -0.013507157564163208, 0.057580772787332535, 0.12617242336273193, 0.18993277847766876, 0.007314556743949652, 0.1412498503923416, -0.019390862435102463, -0.060649171471595764, 0.0683576911687851, 0.03047346882522106, -0.1548633575439453, 0.07651761919260025, 0.14861388504505157, 0.024730173870921135, 0.08602172136306763, -0.00391812901943922, -0.038619235157966614, 0.036564018577337265, 0.014162503182888031, -0.10407629609107971, -0.009580723941326141, 0.04865208640694618, 0.0011154193198308349, 0.2536054849624634, -0.06757504492998123, -0.035831183195114136, 0.023717204108834267, -0.020222146064043045, 0.001644000643864274, 0.0576966218650341, 0.1307695060968399, 0.04560526832938194, 0.07766959071159363, 0.02279781922698021, 0.15661421418190002, -0.17117372155189514, 0.15458279848098755, 0.1685900092124939, -0.31023335456848145, -0.08079558610916138, 0.0989578440785408, 0.0025649196468293667, 0.15555289387702942, 0.06721866875886917, 0.07311951369047165, -0.02905527874827385, 0.009027115069329739, 0.030869394540786743, -0.040228985249996185, -0.15785041451454163, 0.0218032356351614, -0.06142408028244972, -0.1047745943069458, 0.36673131585121155, 0.009546760469675064, 0.01276243757456541, -0.09501731395721436, -0.08762063086032867, 0.043532662093639374, -0.051489848643541336, -0.08067788183689117, -0.0037096692249178886, 0.04021131992340088, 0.052026864141225815, -0.08701822906732559, -0.14043980836868286, -0.04920462146401405, -0.23503190279006958, 0.028095407411456108, 0.03254609555006027, 0.04721202328801155, -0.1718178540468216, 0.07090752571821213, 0.03101051039993763, -0.05047783628106117, 0.016844617202878, -0.11509386450052261, 0.08392325043678284, -0.03451966866850853, -0.027119187638163567, -0.12497487664222717, 0.0785309299826622, 0.0020806826651096344, 0.10759159922599792, 0.0512130968272686, 0.02782854251563549, 0.028882816433906555, -0.03983819857239723, 0.0005483989370986819, -0.0683971717953682, -0.051204681396484375, 0.07933646440505981, -0.014521547593176365, -0.03511563688516617, 0.039471134543418884, -0.1199156641960144, -0.07764838635921478, 0.031194988638162613, 0.07017973065376282, 0.027377966791391373, 0.16408869624137878, 0.03253742307424545, -0.059088632464408875, 0.029362943023443222, -0.12326379120349884, -0.09082014858722687, -0.01032175775617361, -0.06081634759902954, 0.03024214319884777, 0.1001824364066124, -0.03312072530388832, -0.22127665579319, -0.10475117713212967, -0.012290777638554573, -0.049092985689640045, 0.03556884825229645, -0.031124960631132126, 0.008476121351122856, -0.000691770575940609, 0.016559205949306488, -0.24605144560337067, -0.19402378797531128, 0.08894166350364685, 0.054971642792224884, -0.012306003831326962, -0.09412238746881485, -0.02822064235806465, -0.0019657739903777838, 0.025622298941016197, -0.07733623683452606, 0.13662219047546387, -0.07188442349433899, 0.07566122710704803, 0.00379849411547184, 0.09334485232830048, -0.09300081431865692, 0.02057657390832901, -0.04541216045618057, -0.1093655675649643, -0.0443418063223362, 0.1432752162218094, 0.028064578771591187, 0.03773424029350281, -0.13072605431079865, -0.0013173427432775497, -0.08857673406600952, 0.028628041967749596, -0.038003429770469666, 0.1818944364786148, -0.11382584273815155, -0.0863356962800026, 0.1645606905221939, -0.036788664758205414, -0.26690831780433655, 0.10508112609386444, 0.041903238743543625, 0.15289165079593658, 0.14871418476104736, 0.08798041194677353, 0.029408805072307587, -0.039662569761276245, 0.04012782499194145, 0.16940656304359436, -0.10035958141088486, -0.07575662434101105, 0.04442470148205757, 0.07196067273616791, -0.14099639654159546, 0.041691649705171585, -0.02483908273279667, 0.09801749140024185, -0.022772908210754395, -0.062180254608392715, -0.04054023697972298, -0.08821678906679153, 0.0636102631688118, 0.06269855797290802, 0.14507780969142914, -0.056065790355205536, -0.08302091062068939, -0.013900203630328178, -0.001610761508345604, 0.029534541070461273, 0.0181222353130579, -0.1305360198020935, 0.02299690805375576, -0.017656821757555008, 0.06306412816047668, -0.1252632886171341, -0.06377415359020233, -0.06830628216266632, 0.12914906442165375, 0.061240844428539276, 0.039218731224536896, 0.07814257591962814, -0.080108642578125, -0.11981547623872757, 0.05713438242673874, 0.08401922881603241, 0.03244512155652046, -0.07207639515399933, -0.10877949744462967, 0.05700835585594177, 0.036912426352500916, -0.037284303456544876, 0.020887382328510284, 0.008983494713902473, -0.023002183064818382, 0.11739621311426163, -0.0046528177335858345, 0.13818566501140594, 0.01481974869966507, -0.025521295145154, 0.014550733380019665, -0.01806524023413658, 0.08098335564136505, -0.005386132746934891, -0.08132980763912201, 0.14288531243801117, -0.1022549495100975, 0.26925772428512573, 0.1976855993270874, -0.0667826384305954, -0.010747473686933517, 0.04158828407526016, -0.031357329338788986, -0.005789669696241617, 0.06739991903305054, 0.0026978810783475637, 0.06397269666194916, -0.06597078591585159, 0.16843785345554352, -0.16160373389720917, -0.024899285286664963, 0.013227219693362713, 0.005916891153901815, -0.013617472723126411, 0.0633731335401535, 0.07132337242364883, -0.24819102883338928, 0.04349561035633087, 0.1716727316379547, 0.0654614120721817, 0.16955417394638062, 0.01945353113114834, -0.04262332618236542, 0.0939895510673523, 0.01666099764406681, -0.029013045132160187, -0.06118057668209076, -0.16284337639808655, -0.00901181809604168, 0.06433143466711044, 0.013755455613136292, 0.051636554300785065, 0.005991234444081783, 0.007604117505252361, 0.013274384662508965, -0.06639374047517776, -0.052771762013435364, 0.09227236360311508, 0.001221629441715777, 0.1609082669019699, 0.011934270150959492, -0.012859242968261242, 0.05774764344096184, 0.02476336993277073, -0.12730973958969116, 0.1661621630191803, -0.009138379245996475, -0.25402259826660156, 0.04121185466647148, -0.051705047488212585, -0.05953274667263031, 0.011054563336074352, 0.13297279179096222, -0.07614681869745255, -0.014752975665032864, -0.1222677081823349, 0.026383288204669952, 0.005251335445791483, 0.06898597627878189, -0.04420832172036171, 0.03286447748541832, -0.01592332124710083, -0.13463984429836273, -0.013779478147625923, 0.02186625823378563, -0.09310296922922134, 0.059016332030296326, -0.26475054025650024, -0.02789095602929592, 0.05343851447105408, -0.046878036111593246, 0.08972787857055664, -0.10210335999727249, 0.21565976738929749, -0.03388851881027222, 0.1152731254696846, 0.1803574413061142, 0.10026101768016815, -0.01128946803510189, 0.14177240431308746, -0.026756390929222107, -0.056892987340688705, 0.08082704246044159, 0.004063406027853489, -0.0518750324845314, -0.24236519634723663, -0.07025729864835739, -0.10686684399843216, 0.029796242713928223, 0.037745531648397446, 0.07856468856334686, 0.09445159137248993, 0.06059661880135536, -0.08337628096342087, 0.14152580499649048, 0.05452941730618477, 0.022602912038564682, 0.1996871829032898, 0.023215608671307564, 0.07026784867048264, -0.15735802054405212, -0.02855980396270752, 0.14461632072925568, -0.0066701751202344894, 0.05393247306346893, -0.02756066992878914, 0.08524707704782486, 0.05652491748332977, 0.013935199938714504, 0.08808084577322006, 0.1600826233625412, 0.03851981461048126, -0.011017149314284325, -0.029610611498355865, -0.059939004480838776, -0.06604123115539551, 0.003494459670037031, 0.010411903262138367, -0.11391374468803406, -0.07902567833662033, 0.08213024586439133, 0.08387012034654617, 0.1586458534002304, -0.04074225574731827, -0.23470593988895416, -0.04950665682554245, -0.012307099997997284, -0.10712173581123352, -0.09041517227888107, 0.04074670001864433, -0.01838558539748192, -0.12856468558311462, 0.08822251111268997, 0.06699997186660767, 0.11906633526086807, 0.0699777826666832, 0.013400088995695114, -0.012728720903396606, -0.1058514416217804, 0.01809619553387165, 0.09675092995166779, -0.4076746106147766, 0.2157047539949417, 0.05677133798599243, 0.04484822228550911, -0.13992728292942047, -0.022563299164175987, 0.06665191054344177, 0.13889917731285095, 0.14760807156562805, 0.012089650146663189, 0.055339790880680084, 0.08223312348127365, -0.08070849627256393, 0.0967416912317276, 0.04339452460408211, -0.015584636479616165, -0.04645088315010071, -0.07397963851690292, -0.0014897090150043368, 0.01320936344563961, 0.1471778303384781, -0.19814081490039825, -0.11369436979293823, 0.03713858872652054, 0.06026003882288933, -0.04903492331504822, -0.031266920268535614, -0.06870798766613007, -0.023296285420656204, 0.0784367024898529, -0.05338709056377411, -0.17354774475097656, -0.10525823384523392, -0.019591763615608215, 0.009787810035049915, -0.0784505307674408, 0.040261849761009216, -0.04596274346113205, -0.06980276852846146, 0.06209464371204376, -0.21252471208572388, 0.0921318382024765, -0.08104126900434494, -0.01174075435847044, 0.009177906438708305, 0.09814700484275818, -0.1114659309387207, 0.03889576718211174, 0.025063417851924896, -0.10263238847255707, -0.048011261969804764, -0.026859141886234283, -0.09440290927886963, -0.04474595561623573, -0.005864864215254784, 0.02044767141342163, -0.04460901767015457, -0.1191735789179802, 0.036429569125175476, -0.08892792463302612, 0.16056382656097412, -0.03806855529546738, 0.008205791935324669, 0.142361119389534, 0.1627560704946518, -0.0024931677617132664, -0.2441348433494568, -0.21170344948768616, 0.00924055464565754, -0.007342174649238586, -0.0375901535153389, -0.06934160739183426, 0.09607673436403275, -0.07215746492147446, 0.027496594935655594, -0.08123189955949783, -0.30503085255622864, -0.12188053876161575, 0.1113579124212265, 0.03951913118362427, 0.24625778198242188, -0.12902498245239258, -0.013316800817847252, -0.0481618195772171, -0.14210447669029236, 0.06973717361688614, -0.1799076646566391, 0.10414160043001175, -0.03020510822534561, -0.008549928665161133, -0.0121718505397439, 0.040120773017406464, 0.0030383183620870113, -0.028273221105337143, -0.04082583263516426, -0.1791260838508606, 0.0073084281757473946, 0.09985596686601639, -0.0019094364251941442, 0.09701486676931381, -0.13458238542079926, 0.08109336346387863, -0.12886318564414978, -0.11061619222164154, -0.08686019480228424, -0.0022333755623549223, 0.0006446726620197296, -0.06612511724233627, 0.00564115634188056, -0.07327787578105927, 0.10045322775840759, -0.014962686225771904, 0.004371653776615858, -0.06842345744371414, -0.0025022528134286404, 0.05234115943312645, 0.29883724451065063, -0.24324963986873627, 0.045158058404922485, -0.04713975265622139, -0.023108892142772675, 0.052919499576091766, -0.2696514129638672, 0.019985755905508995, 0.04880889505147934, -0.07604338228702545, -0.01479797251522541, 0.05607622116804123, 0.05266954004764557, -0.05535763502120972, 0.14483118057250977, -0.08965414017438889, -0.10347064584493637, -0.09645140916109085, -0.047198887914419174, -0.045626092702150345, -0.03125627338886261, 0.07890259474515915, -0.04229385405778885, -0.011283526197075844, -0.035380251705646515, -0.05642405152320862, -0.04593750089406967, 0.06642930209636688, 0.044312167912721634, 0.010123591870069504, -0.08160970360040665, 0.12128522992134094, 0.08054470270872116, -0.11164120584726334, 0.015840154141187668, 0.19561073184013367, -0.18097476661205292, -0.12390484660863876, 0.10944192856550217, 0.03659611567854881, -0.05275776609778404, -0.07204890996217728, -0.04622809961438179, -0.10843297839164734, 0.06832730770111084, 0.14691752195358276, 0.02242944948375225, 0.0465363971889019, -0.016601262614130974, -0.04154156148433685, 0.05424277111887932, 0.09718883782625198, -0.022151123732328415, -0.033106494694948196, -0.09264538437128067, 0.045055922120809555, -0.030747924000024796, 0.14876216650009155, -0.07524311542510986, 0.041929297149181366, -0.08911512792110443, 0.042364299297332764, -0.15798471868038177, 0.0415738970041275, -0.04888395592570305, 0.0005565468454733491, -0.023173077031970024, -0.029814019799232483, -0.0917954370379448, 0.014436542987823486, -0.05552567541599274, 0.017727365717291832, 0.016131192445755005, 0.13201747834682465, -0.004234015010297298, 0.04479411244392395, 0.00044098077341914177, 0.0007593867485411465, 0.16071321070194244, 0.0509738065302372, -0.1263405829668045, 0.10758709907531738, -0.1307990700006485, -0.04471616819500923, -0.03626412898302078, 0.06221754103899002, 0.0025018337182700634, 0.04732459783554077, 0.09414905309677124, 0.12399694323539734, 0.04874320700764656, 0.04368790239095688, 0.05333460494875908, -0.04143068939447403, 0.016893118619918823, -0.0644073337316513, 0.010863413102924824, -0.047066401690244675, 0.016908902674913406, 0.05963510274887085, -0.007173455320298672, 0.06204816326498985, -0.06937198340892792, -0.09417564421892166, -0.1222994476556778, 0.033630430698394775, -0.01169496588408947, -0.08223564922809601, -0.09232743829488754, -0.10885889828205109, 0.04103971645236015, -0.03836531564593315, 0.17744304239749908, 0.07955462485551834, -0.012011376209557056, 0.06890439987182617, 0.06957778334617615, 0.0888267457485199, -0.040900878608226776, 0.27196311950683594, -0.006454644724726677, -0.005456349812448025, -0.02629346027970314, -0.03532589599490166, -0.0021374342031776905, 0.06577400118112564, 0.17286603152751923, 0.10523633658885956, -0.04394130781292915, 0.10307378321886063, -0.056762028485536575, 0.028633063659071922, -0.03001227043569088, -0.1577243059873581, 0.025283515453338623, 0.05340106412768364, -0.08961648494005203, -0.157767653465271, 0.2638855576515198, 0.011050494387745857, -0.0003669133875519037, -0.04144907742738724, -0.015597330406308174, -0.1202862486243248, -0.14774517714977264, -0.08677752315998077, -0.11882302165031433, -0.025900620967149734, -0.09014833718538284, -0.04163045436143875, 0.06691112369298935, 0.09239786118268967, -0.060058098286390305, 0.14197252690792084, -0.04333092272281647, -0.10918912291526794, 0.07717499881982803, -0.06571084260940552, 0.026461191475391388, -0.019074423238635063, 0.009839634411036968, -0.0728207379579544, -0.021201901137828827, -0.0562574528157711, 0.02003180980682373, -0.0836159959435463, 0.020180149003863335, -0.08354221284389496, -0.020451432093977928, -0.07209879904985428, 0.01332785189151764, -0.0028270042967051268, 0.15676765143871307, 0.021137308329343796, -0.09216602146625519, 0.04375127702951431, 0.1887921392917633, 0.021478503942489624, -0.2876771092414856, -0.11302199214696884, 0.059153519570827484, 0.03817514330148697, -0.0035575288347899914, 0.03902176767587662, 0.0053211539052426815, -0.07501370459794998, 0.30939745903015137, 0.21973027288913727, -0.03835910186171532, 0.020620811730623245, 0.0034744522999972105, 0.02976032719016075, 0.027458377182483673, 0.19053269922733307, 0.0657840222120285, 0.2022174447774887, -0.04391831159591675, 0.0302133671939373, 0.01489434763789177, 0.046752821654081345, -0.10815948992967606, 0.14823485910892487, 0.06384976953268051, -0.07995199412107468, 0.026309093460440636, 0.08295897394418716, -0.09308728575706482, 0.17409348487854004, -0.08179035782814026, -0.018238769844174385, -0.042792052030563354, -0.040719158947467804, 0.035969074815511703, 0.08818861842155457, -0.023370694369077682, -0.002367977052927017, 0.009687250480055809, -0.04731537774205208, 0.037216659635305405, -0.2509971261024475, -0.035256240516901016, 0.13550010323524475, -0.03187497332692146, 0.12678757309913635, 0.01182536967098713, 0.11095133423805237, 0.061610251665115356, -0.0013455911539494991, -0.08866538852453232, 0.1236717626452446, 0.03727157041430473, 0.10031434893608093, 0.13154742121696472, -0.04326979070901871, -0.016905946657061577, -0.07069453597068787, 0.01857462339103222, 0.056819941848516464, 0.03683100640773773, 0.16048605740070343, 0.04816862940788269, -0.06119856983423233, 0.11082231998443604, -0.09169897437095642, 0.06471070647239685, 0.10136169195175171, -0.03100559115409851, 0.004341051913797855, -0.10569014400243759, 0.04320846125483513, 0.0422881543636322, -0.1089264303445816, -0.0702488124370575, -0.029211221262812614, -0.05146029591560364, -0.08406341820955276, 0.019895363599061966, -0.12537701427936554, 0.007694680709391832, -0.11341951787471771, -0.018104739487171173, -0.15696634352207184, 0.09523811936378479, 0.19223107397556305, -0.0059552546590566635, 0.021402660757303238, -0.07668089121580124, -0.011995384469628334, 0.0864790827035904, -0.15446631610393524, -0.12268785387277603 ]
null
null
null
The main card for Saturdayโ€™s Manny Pacquiao vs Yordenis Ugas fight gets underway at T-Mobile Arena in Las Vegas at 9 p.m. ET and the main event is expected to start sometime around 11:30 p.m. This is going to air on FOX Sports PPV and YouTube PPV. The card will cost https://web.sites.google.com/view/ppv-livemanny-pacquiao-vs-yord/home https://web.sites.google.com/view/freevasyl-manny-pacquiao-vs-yo/home https://web.sites.google.com/view/ppvlivestreammannypacquiaovsyo/home https://web.sites.google.com/view/watchtv-manny-pacquiao-vs-yord/home https://web.sites.google.com/view/heresmannypacquiaovsyordenisug/home https://web.sites.google.com/view/mannypacquiaovsyordenislive/home https://web.sites.google.com/view/free-2021-manny-pacquiao-vs-yo/home LIVE::Watch Full Fight Live Here LIVE::Watch Full Fight Live Here https://goodavail.com/boxing/ The most intriguing storyline for this fight is the belt itself that is on the line. Pacquiao won the Super version of the WBAโ€™s welterweight title in July 2019 when he beat Keith Thurman in a split decision. The WBA stripped Pacquiao of the title this past January due to inactivity. The organizing body then promoted Ugas into the Super belt. Ugas won the WBAโ€™s Regular title in September 2020 when he beat Abel Ramos in a split decision. That means Ugas won a title last held by Pacquiao without having to beat Pacquiao. Pacquiao-Spence would have been a significantly bigger fight for welterweight supremacy, but this is still interesting. Pacquiao was an underdog against Spence, but comes into this fight as a -360 favorite at DraftKings Sportsbook. How to watch Manny Pacquiao vs. Yordenis Ugas TV channel: FOX Sports PPV
{}
null
balalsahabi/fdgdfg
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
The main card for Saturdayโ€™s Manny Pacquiao vs Yordenis Ugas fight gets underway at T-Mobile Arena in Las Vegas at 9 p.m. ET and the main event is expected to start sometime around 11:30 p.m. This is going to air on FOX Sports PPV and YouTube PPV. The card will cost URL URL URL URL URL URL URL LIVE::Watch Full Fight Live Here LIVE::Watch Full Fight Live Here URL The most intriguing storyline for this fight is the belt itself that is on the line. Pacquiao won the Super version of the WBAโ€™s welterweight title in July 2019 when he beat Keith Thurman in a split decision. The WBA stripped Pacquiao of the title this past January due to inactivity. The organizing body then promoted Ugas into the Super belt. Ugas won the WBAโ€™s Regular title in September 2020 when he beat Abel Ramos in a split decision. That means Ugas won a title last held by Pacquiao without having to beat Pacquiao. Pacquiao-Spence would have been a significantly bigger fight for welterweight supremacy, but this is still interesting. Pacquiao was an underdog against Spence, but comes into this fight as a -360 favorite at DraftKings Sportsbook. How to watch Manny Pacquiao vs. Yordenis Ugas TV channel: FOX Sports PPV
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
transformers
# Named Entity Recognition using Transformers This is a Fine-tuned version of BERT using HuggingFace transformers to perform Named Entity Recognition on Text data. BERT is a state-of-the-art model with attention mechanism as underlying architecture trained with masked-language-modeling and next-sentence-prediction objectives, used for various tasks including Question answering systems, Text Summarization, etc... which can also perform token classification tasks such as NER with great performance. # Dataset **CoNLL-2003** : The shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations, and names of miscellaneous entities that do not belong to the previous three groups.<br><br> **Link** : https://huggingface.co/datasets/conll2003 # Using this fine-tuned version From python, download the whole pipeline and use it instantly using the following code : ``` from transformers import pipeline # Loading the pipeline from hub # Pipeline handles the preprocessing and post processing steps model_checkpoint = "balamurugan1603/bert-finetuned-ner" namedEntityRecogniser = pipeline( "token-classification", model=model_checkpoint, aggregation_strategy="simple" ) ``` Reference for using this pipeline to find NER tags can be found in this <a href="https://github.com/balamurugan1603/Named-Entity-Recognition-using-Tranformers/blob/main/named-entity-recognition-using-transfer-learning.ipynb">notebook</a>.
{}
token-classification
balamurugan1603/bert-finetuned-ner
[ "transformers", "pytorch", "tf", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tf #bert #token-classification #autotrain_compatible #endpoints_compatible #has_space #region-us
# Named Entity Recognition using Transformers This is a Fine-tuned version of BERT using HuggingFace transformers to perform Named Entity Recognition on Text data. BERT is a state-of-the-art model with attention mechanism as underlying architecture trained with masked-language-modeling and next-sentence-prediction objectives, used for various tasks including Question answering systems, Text Summarization, etc... which can also perform token classification tasks such as NER with great performance. # Dataset CoNLL-2003 : The shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations, and names of miscellaneous entities that do not belong to the previous three groups.<br><br> Link : URL # Using this fine-tuned version From python, download the whole pipeline and use it instantly using the following code : Reference for using this pipeline to find NER tags can be found in this <a href="URL
[ "# Named Entity Recognition using Transformers\nThis is a Fine-tuned version of BERT using HuggingFace transformers to perform Named Entity Recognition on Text data. BERT is a state-of-the-art model with attention mechanism as underlying architecture trained with masked-language-modeling and next-sentence-prediction objectives, used for various tasks including Question answering systems, Text Summarization, etc... which can also perform token classification tasks such as NER with great performance.", "# Dataset\nCoNLL-2003 :\nThe shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations, and names of miscellaneous entities that do not belong to the previous three groups.<br><br>\nLink : URL", "# Using this fine-tuned version\n\nFrom python, download the whole pipeline and use it instantly using the following code :\n\n\nReference for using this pipeline to find NER tags can be found in this <a href=\"URL" ]
[ "TAGS\n#transformers #pytorch #tf #bert #token-classification #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Named Entity Recognition using Transformers\nThis is a Fine-tuned version of BERT using HuggingFace transformers to perform Named Entity Recognition on Text data. BERT is a state-of-the-art model with attention mechanism as underlying architecture trained with masked-language-modeling and next-sentence-prediction objectives, used for various tasks including Question answering systems, Text Summarization, etc... which can also perform token classification tasks such as NER with great performance.", "# Dataset\nCoNLL-2003 :\nThe shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations, and names of miscellaneous entities that do not belong to the previous three groups.<br><br>\nLink : URL", "# Using this fine-tuned version\n\nFrom python, download the whole pipeline and use it instantly using the following code :\n\n\nReference for using this pipeline to find NER tags can be found in this <a href=\"URL" ]
[ 44, 122, 79, 50 ]
[ "passage: TAGS\n#transformers #pytorch #tf #bert #token-classification #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Named Entity Recognition using Transformers\nThis is a Fine-tuned version of BERT using HuggingFace transformers to perform Named Entity Recognition on Text data. BERT is a state-of-the-art model with attention mechanism as underlying architecture trained with masked-language-modeling and next-sentence-prediction objectives, used for various tasks including Question answering systems, Text Summarization, etc... which can also perform token classification tasks such as NER with great performance.# Dataset\nCoNLL-2003 :\nThe shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations, and names of miscellaneous entities that do not belong to the previous three groups.<br><br>\nLink : URL# Using this fine-tuned version\n\nFrom python, download the whole pipeline and use it instantly using the following code :\n\n\nReference for using this pipeline to find NER tags can be found in this <a href=\"URL" ]
[ 0.007528549060225487, 0.12225406616926193, -0.005482603795826435, 0.07834136486053467, 0.16907888650894165, -0.006162520498037338, 0.025740865617990494, 0.035751648247241974, 0.01377036515623331, 0.06663728505373001, -0.049724236130714417, 0.03469553589820862, -0.008502813056111336, 0.010965624824166298, 0.0228731669485569, -0.17571376264095306, 0.013792689889669418, -0.018684446811676025, 0.01417973730713129, 0.07645481824874878, 0.05029646307229996, -0.046168841421604156, 0.09513140469789505, 0.07975509762763977, -0.08415573090314865, 0.06610231846570969, -0.022023197263479233, -0.04557172209024429, 0.10452590882778168, 0.052429165691137314, 0.15616446733474731, -0.04430682957172394, -0.008222244679927826, -0.1166985034942627, 0.015569187700748444, 0.0999191477894783, -0.031979694962501526, 0.02845636010169983, 0.09893042594194412, -0.13079528510570526, -0.018156468868255615, -0.06067188084125519, 0.019962361082434654, 0.04796477407217026, -0.07989838719367981, -0.0708577111363411, -0.0027477373369038105, 0.0931185632944107, -0.01845436543226242, 0.08625508844852448, 0.01674116961658001, 0.19178789854049683, -0.1273261159658432, 0.11549551039934158, 0.2567889988422394, -0.20586948096752167, -0.034742433577775955, -0.025675715878605843, -0.031076418235898018, 0.0011836992343887687, -0.02176542952656746, 0.011781525798141956, -0.056636352092027664, 0.05854285880923271, 0.08624453097581863, -0.055741772055625916, -0.2435453087091446, 0.017150726169347763, -0.14749127626419067, 0.011875377967953682, 0.038131214678287506, -0.008099626749753952, -0.017492936924099922, -0.05911659449338913, -0.11911209672689438, -0.004714626353234053, -0.03921918943524361, 0.003631198313087225, 0.015147014521062374, -0.0568079873919487, 0.0011188517091795802, -0.03913149610161781, -0.04931601881980896, -0.01647445745766163, -0.0735938772559166, 0.16676881909370422, 0.0016491520218551159, 0.0398126095533371, -0.11107420176267624, 0.08906181156635284, 0.004619103390723467, -0.00850161723792553, -0.04589742794632912, -0.050219688564538956, -0.20872776210308075, 0.008805463090538979, -0.07144483923912048, 0.024235352873802185, 0.03320371359586716, 0.17186813056468964, 0.05574410781264305, 0.056834593415260315, -0.06466758996248245, 0.031921885907649994, -0.00880408100783825, 0.14262712001800537, -0.061806682497262955, 0.03066432662308216, 0.06582997739315033, -0.08054088801145554, 0.021980594843626022, -0.01494657527655363, -0.1008189395070076, 0.015377414412796497, 0.01873454637825489, 0.0632517859339714, -0.023996036499738693, 0.16189314424991608, -0.051280632615089417, -0.0348912738263607, 0.07501714676618576, -0.10633185505867004, 0.04697591811418533, 0.07598205655813217, -0.03079054318368435, -0.006084990222007036, 0.08182619512081146, -0.024950725957751274, -0.09933055937290192, 0.050864581018686295, -0.09353524446487427, -0.01061373483389616, -0.018955154344439507, -0.11634696274995804, 0.05350701883435249, -0.06182946264743805, -0.04688260704278946, -0.23085248470306396, -0.116592176258564, 0.04641709104180336, 0.059124357998371124, -0.02167496271431446, 0.029095496982336044, 0.00509471632540226, -0.026145005598664284, -0.006825603544712067, 0.008677193894982338, -0.09975142776966095, -0.038552287966012955, -0.005008756648749113, -0.07431547343730927, 0.055140964686870575, -0.0366334430873394, 0.024592431262135506, -0.11250126361846924, -0.002831262070685625, -0.10528066009283066, 0.05033581703901291, -0.11840037256479263, 0.02191789820790291, -0.10670246928930283, -0.019583098590373993, 0.016234654933214188, 0.03325715288519859, 0.019137682393193245, 0.03729182109236717, 0.003349733306095004, -0.09351709485054016, 0.08794499188661575, -0.03303469344973564, 0.006372281815856695, 0.09561990946531296, 0.002537296386435628, -0.008347442373633385, 0.041700661182403564, 0.18704724311828613, 0.12114100903272629, -0.09694544970989227, 0.026206357404589653, 0.037645261734724045, -0.12268376350402832, 0.012260775081813335, -0.0261988565325737, 0.03653653711080551, -0.016092849895358086, 0.03627632558345795, -0.0664743185043335, 0.06544248014688492, 0.004597943276166916, -0.03189895674586296, 0.030517539009451866, 0.007858154363930225, 0.007505638524889946, -0.054855141788721085, 0.07747026532888412, 0.0614258237183094, -0.0859527662396431, 0.14518927037715912, 0.0004219585098326206, -0.04960870370268822, 0.07427213340997696, -0.029369143769145012, 0.0690884217619896, 0.06382710486650467, 0.011993790976703167, -0.2371787428855896, -0.20476241409778595, 0.054484810680150986, -0.11927218735218048, 0.05727986618876457, 0.0037755670491605997, 0.013396731577813625, 0.04110291600227356, 0.05874529108405113, 0.017827073112130165, 0.02680782973766327, 0.007999387569725513, -0.02674286440014839, -0.04410442337393761, -0.04635302349925041, -0.0321216881275177, 0.11080633103847504, -0.08929337561130524, 0.04294728860259056, 0.0451342836022377, 0.14145155251026154, 0.03848038241267204, -0.00413353368639946, -0.019938740879297256, 0.024860167875885963, -0.03369009494781494, -0.006547586992383003, 0.07188739627599716, 0.010700319893658161, -0.0544079951941967, 0.07871149480342865, -0.05967026576399803, -0.03578779473900795, 0.029612034559249878, -0.002788483165204525, -0.07484087347984314, 0.009406107477843761, -0.05298344045877457, -0.03156745061278343, 0.03170338273048401, -0.053018927574157715, 0.12117224931716919, 0.06665357947349548, 0.14326083660125732, -0.04787827655673027, -0.08463054150342941, -0.010012264363467693, 0.002290785312652588, -0.011913163587450981, 0.11291185021400452, -0.030001049861311913, -0.17955519258975983, 0.0947900265455246, -0.04417363554239273, -0.12200003862380981, 0.0946345403790474, 0.02874640002846718, -0.01686173491179943, 0.058231040835380554, 0.0060033309273421764, 0.041682321578264236, -0.06156741455197334, -0.019791387021541595, -0.029236197471618652, 0.054601408541202545, 0.024657076224684715, 0.001531794317997992, -0.04418300837278366, 0.059070885181427, -0.01093826349824667, 0.010751510970294476, -0.014386318624019623, 0.058447252959012985, 0.024856066331267357, 0.06694179028272629, 0.05801599472761154, -0.07131721079349518, 0.043866124004125595, -0.031441882252693176, -0.11036687344312668, 0.15450257062911987, -0.10640324652194977, -0.31663888692855835, -0.17474868893623352, -0.10251780599355698, -0.13045915961265564, -0.021095385774970055, 0.09187258034944534, 0.004853078164160252, -0.04040161892771721, -0.053989242762327194, 0.02300950512290001, 0.014831802807748318, -0.0438138023018837, -0.06616150587797165, -0.0486990362405777, 0.03172772750258446, -0.15105532109737396, 0.025074835866689682, -0.03762895613908768, -0.048296332359313965, 0.08100838959217072, -0.043722592294216156, 0.12244302779436111, 0.06922124326229095, 0.000178752132342197, 0.003300340147688985, -0.02719198539853096, 0.15692922472953796, -0.033759817481040955, 0.03952539339661598, 0.03377293795347214, -0.1000346839427948, 0.10793492943048477, 0.052186623215675354, 0.027308199554681778, 0.006949512287974358, 0.013280520215630531, 0.11147014796733856, -0.055946338921785355, -0.11900018155574799, -0.11921951174736023, -0.09929061681032181, 0.014770982787013054, 0.05269020050764084, 0.07918308675289154, 0.0007363133481703699, -0.08150121569633484, -0.015111168846487999, 0.022913653403520584, 0.05733899027109146, 0.133086159825325, 0.11750178784132004, 0.02861732244491577, 0.07850917428731918, -0.006726662628352642, -0.09874781221151352, 0.039448510855436325, 0.05125320702791214, 0.26325663924217224, -0.06100137159228325, 0.11573874205350876, 0.06838897615671158, -0.028443709015846252, 0.09820979833602905, 0.08670695871114731, -0.07231596857309341, 0.05189136415719986, -0.01893848553299904, -0.07170017063617706, 0.07923094183206558, 0.01782308891415596, -0.04622502252459526, -0.08494523167610168, 0.025331344455480576, -0.12516731023788452, 0.07645899057388306, 0.31199243664741516, -0.0027568102814257145, -0.19188787043094635, 0.001988853793591261, -0.029187718406319618, -0.04145548865199089, -0.04417731240391731, -0.02253812551498413, 0.016632555052638054, -0.08659819513559341, 0.058694470673799515, -0.06812687963247299, 0.07961954176425934, 0.019848283380270004, 0.03179961442947388, 0.029139410704374313, 0.051775600761175156, 0.004584327805787325, 0.09070379287004471, -0.03970819339156151, 0.15963706374168396, 0.012713945470750332, -0.019156455993652344, -0.05649633705615997, -0.050274305045604706, 0.00388504471629858, 0.035882219672203064, 0.09573185443878174, 0.03432620316743851, 0.06372645497322083, -0.14692144095897675, -0.031178681179881096, 0.036935530602931976, 0.07361043244600296, -0.013885204680263996, 0.06578613072633743, -0.02055586874485016, -0.06511886417865753, 0.012563490308821201, 0.044135887175798416, -0.055773764848709106, -0.146074116230011, 0.009205779060721397, 0.07511065900325775, -0.03506916016340256, -0.021174782887101173, -0.02209971286356449, 0.00344155658967793, 0.17279678583145142, -0.1496015042066574, -0.01847018301486969, -0.11932064592838287, -0.06844697147607803, 0.015757832676172256, -0.08926480263471603, 0.07231657207012177, -0.06472791731357574, 0.08954556286334991, 0.004917249083518982, -0.2249099761247635, 0.08004652708768845, -0.06014963611960411, -0.05170386657118797, -0.014514853246510029, 0.04281896725296974, 0.10477959364652634, -0.013203270733356476, -0.05532493069767952, 0.02256680279970169, -0.0007964743417687714, -0.0572342649102211, -0.05807977542281151, 0.2011169195175171, 0.05828240513801575, 0.1592150777578354, -0.15796369314193726, -0.11162178963422775, -0.009678740985691547, 0.08993425220251083, 0.1624414175748825, 0.10804605484008789, -0.06803890317678452, 0.1394357681274414, 0.19262458384037018, -0.024184687063097954, -0.30093061923980713, -0.01629030890762806, 0.08250518143177032, -0.021826207637786865, -0.006314060650765896, -0.12333434075117111, 0.1401858776807785, 0.12192831188440323, -0.009074673987925053, 0.037388697266578674, -0.1308896392583847, -0.07099995762109756, 0.07261845469474792, 0.0021266653202474117, 0.05627337470650673, -0.04581744968891144, -0.01373043842613697, -0.04144321009516716, -0.14620454609394073, 0.20532922446727753, -0.16551440954208374, 0.09666625410318375, -0.03552229329943657, 0.04510483890771866, -0.0017311564879491925, -0.031950436532497406, -0.018867624923586845, 0.07009860128164291, 0.05480695515871048, 0.002547154203057289, -0.0746462419629097, 0.0730796605348587, -0.06803783029317856, 0.07906073331832886, 0.0022209968883544207, 0.022823642939329147, 0.006728145759552717, -0.00675116665661335, -0.12181207537651062, 0.1582196056842804, 0.027466604486107826, -0.023414041846990585, -0.053320709615945816, 0.039589714258909225, 0.05366329103708267, -0.007687225937843323, 0.1564885675907135, -0.04300830513238907, 0.027926310896873474, 0.1789221167564392, 0.03552437573671341, 0.01638863794505596, -0.08911625295877457, -0.04139852896332741, -0.03743959963321686, 0.10752720385789871, -0.010347667150199413, 0.08104335516691208, 0.08754952251911163, 0.04357913136482239, 0.06735939532518387, 0.07631662487983704, -0.09393317997455597, -0.07066429406404495, 0.05165300890803337, -0.12102581560611725, -0.02754862792789936, -0.06585313379764557, -0.08473406732082367, -0.037287626415491104, 0.09780854731798172, 0.14433997869491577, 0.007150072604417801, -0.02796545997262001, 0.06007035821676254, -0.009827633388340473, 0.0009434791281819344, 0.0072874161414802074, -0.01839950494468212, 0.04057087004184723, -0.06491623818874359, 0.0700593963265419, 0.06512999534606934, 0.026660049334168434, -0.01694480888545513, 0.11943064630031586, -0.1729116588830948, -0.0876353308558464, -0.044134922325611115, 0.16964514553546906, -0.09262805432081223, 0.00951513834297657, -0.008437174372375011, -0.014061222784221172, 0.036126092076301575, 0.2175002545118332, 0.008766197599470615, 0.03113110363483429, -0.08090975135564804, 0.015085157006978989, -0.01272633671760559, 0.0434197373688221, -0.029457686468958855, 0.0835447907447815, -0.11109551787376404, 0.001112875179387629, 0.003963523544371128, 0.09228016436100006, -0.042111337184906006, -0.07539061456918716, -0.21409520506858826, 0.020274568349123, -0.13478194177150726, 0.019855037331581116, 0.005049507133662701, 0.0190481785684824, 0.012033571489155293, 0.025717107579112053, -0.015742681920528412, 0.04449060186743736, -0.04036808758974075, -0.02895941771566868, -0.02627171017229557, 0.05114791542291641, -0.137689009308815, -0.043882761150598526, 0.009021780453622341, -0.07109451293945312, 0.10011611878871918, 0.07285381108522415, -0.05184081569314003, 0.0756513699889183, -0.11627110838890076, -0.09868133813142776, 0.048202648758888245, 0.036486439406871796, 0.08960118144750595, 0.0320558063685894, -0.007539490703493357, -0.038561537861824036, 0.020349552854895592, -0.0019192374311387539, 0.04495123028755188, -0.09701628983020782, 0.015640148892998695, -0.060605429112911224, -0.06398545205593109, -0.09674399346113205, -0.02802390418946743, 0.04474398121237755, 0.11583689600229263, 0.10150883346796036, -0.043144941329956055, 0.03104584291577339, -0.029523678123950958, -0.05136996507644653, 0.05410061031579971, -0.11798983812332153, -0.028348851948976517, -0.07778258621692657, 0.014053789898753166, -0.03966878354549408, 0.14188657701015472, 0.02310808189213276, -0.011185159906744957, 0.013563019223511219, -0.011963427998125553, -0.06590631604194641, -0.0034458988811820745, -0.09049626439809799, 0.02180691994726658, -0.04579438641667366, 0.02337988093495369, -0.021617397665977478, -0.01777769811451435, -0.0006165424711070955, 0.1574171930551529, 0.09446651488542557, -0.010116377845406532, 0.022055553272366524, 0.09216108173131943, -0.0931442603468895, -0.13213184475898743, -0.15424714982509613, -0.03989216312766075, 0.13055342435836792, -0.06186142563819885, 0.015185177326202393, 0.08365604281425476, -0.022909287363290787, 0.06270594149827957, -0.029851313680410385, -0.04906554892659187, -0.06369198113679886, -0.15779083967208862, -0.03953764587640762, -0.031475018709897995, -0.036851488053798676, -0.030420590192079544, -0.01819382794201374, 0.026310548186302185, 0.024526793509721756, 0.042096253484487534, 0.10201093554496765, -0.04102851450443268, -0.06032437086105347, 0.02301444672048092, -0.0046924506314098835, -0.006063580513000488, 0.07524658739566803, 0.03548477590084076, 0.0504482239484787, 0.06319965422153473, 0.026988334953784943, 0.03728800266981125, 0.1083715558052063, 0.033010270446538925, -0.11476446688175201, -0.06634347885847092, 0.001739110448397696, 0.05697738379240036, -0.08538524806499481, -0.0018556482391431928, 0.07653828710317612, -0.07210464030504227, 0.011768141761422157, 0.1469677835702896, -0.07161097973585129, -0.03447828069329262, -0.06826569139957428, 0.43943750858306885, 0.011293591000139713, 0.03118187189102173, -0.04183332994580269, -0.11730010062456131, -0.09073043614625931, 0.24402908980846405, 0.1765201985836029, -0.03294721990823746, 0.0403815433382988, -0.003148841205984354, 0.005034603178501129, 0.034341294318437576, 0.10425983369350433, 0.08131813257932663, 0.32085826992988586, 0.004321419633924961, 0.0130739975720644, 0.02883758954703808, 0.003954386804252863, -0.14843939244747162, 0.12129614502191544, 0.03329913690686226, -0.05303168669342995, -0.09377173334360123, -0.026433972641825676, -0.16452188789844513, -0.18953102827072144, 0.0469587966799736, 0.0017970285844057798, -0.10727667063474655, -0.031625423580408096, -0.041152168065309525, -0.0339234359562397, 0.0869092047214508, -0.025589272379875183, -0.06231439486145973, 0.061381638050079346, 0.006527516525238752, -0.07689716666936874, -0.059732891619205475, 0.06162500008940697, -0.07466757297515869, 0.1028650626540184, -0.00538240559399128, 0.051485348492860794, 0.07267442345619202, 0.043157316744327545, -0.05203089490532875, 0.025677340105175972, -0.01952969655394554, -0.13672590255737305, -0.03963203355669975, 0.051255963742733, -0.049987271428108215, 0.021951880306005478, -0.009895832277834415, -0.1278521567583084, 0.02201090194284916, 0.13160505890846252, 0.031555820256471634, -0.09747973829507828, -0.034750502556562424, -0.04109690338373184, 0.0807982012629509, 0.15625806152820587, 0.015536467544734478, -0.0010649659670889378, -0.062257129698991776, -0.057036805897951126, 0.020904097706079483, -0.04627259448170662, -0.01459499541670084, -0.11158247292041779, 0.007958609610795975, -0.02330971509218216, 0.07975999265909195, -0.21415093541145325, -0.013388508930802345, 0.029234658926725388, 0.011919749900698662, -0.022527212277054787, 0.05283833667635918, 0.0830390527844429, 0.01732832007110119, 0.007599837612360716, -0.15979567170143127, 0.03752562031149864, 0.06539442390203476, -0.07740014046430588, -0.05005139112472534 ]
null
null
transformers
# Test Bot DialoGTP Model
{"tags": ["conversational"]}
text-generation
balta/DialoGPT-small-TestBot
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Test Bot DialoGTP Model
[ "# Test Bot DialoGTP Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Test Bot DialoGTP Model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Test Bot DialoGTP Model" ]
[ -0.04349280893802643, 0.012079959735274315, -0.0047833239659667015, -0.003339349292218685, 0.08934434503316879, -0.016524765640497208, 0.15666301548480988, 0.1434221863746643, 0.055868543684482574, -0.030258648097515106, 0.13060198724269867, 0.174707293510437, -0.0024456295650452375, 0.14732307195663452, -0.08187790215015411, -0.1959865838289261, 0.07326769083738327, 0.05250118300318718, 0.05510367825627327, 0.16432325541973114, 0.09980912506580353, -0.04636353254318237, 0.06787983328104019, 0.009996764361858368, -0.17382067441940308, 0.014652514830231667, 0.012141114100813866, -0.12404689192771912, 0.101505808532238, 0.022127186879515648, 0.04576393961906433, 0.029055356979370117, -0.054233256727457047, -0.0914701446890831, 0.027803633362054825, -0.022212188690900803, -0.0005706567899324, 0.03782236948609352, -0.0013932103756815195, -0.08667516708374023, 0.14267480373382568, 0.1617392599582672, 0.028684554621577263, 0.04819684103131294, -0.09443525969982147, -0.011103861965239048, -0.007740641012787819, 0.020771313458681107, 0.07591584324836731, 0.14995551109313965, -0.07293687015771866, 0.25861406326293945, -0.15195918083190918, 0.1252053827047348, 0.07495872676372528, -0.31259870529174805, -0.026777146384119987, 0.0946706086397171, 0.0834943875670433, 0.033498216420412064, -0.018667293712496758, 0.039627011865377426, 0.053527042269706726, 0.051515959203243256, -0.1151074767112732, -0.07549036294221878, -0.12273851782083511, -0.011600526981055737, -0.11273282766342163, -0.05916866287589073, 0.27634185552597046, -0.03350494056940079, 0.010777023620903492, -0.0799441933631897, -0.04616814851760864, -0.03295524790883064, -0.013984978199005127, -0.03508565202355385, -0.09869056195020676, 0.029076356440782547, -0.03965786471962929, -0.06368953734636307, -0.11747509986162186, -0.09802275896072388, -0.17086264491081238, 0.1580704152584076, 0.062321972101926804, 0.036546170711517334, -0.22663122415542603, 0.11366543918848038, 0.07447946816682816, -0.09847048670053482, -0.023437002673745155, -0.0902329683303833, -0.02247735485434532, -0.026430008932948112, -0.05278381332755089, -0.12842752039432526, 0.1127433031797409, 0.17148694396018982, -0.035567428916692734, 0.04830288514494896, -0.0728769525885582, 0.032600030303001404, 0.036074116826057434, 0.1190914437174797, 0.045931506901979446, 0.008900911547243595, 0.024125881493091583, -0.04894696921110153, -0.029646530747413635, -0.009930419735610485, -0.11873113363981247, -0.010667018592357635, 0.10761056840419769, 0.08063140511512756, 0.06092090159654617, 0.0956803560256958, -0.03087906911969185, -0.030173014849424362, 0.014618941582739353, -0.06728722900152206, -0.012995092198252678, 0.03166671842336655, -0.04997754469513893, 0.04395969212055206, 0.011787334457039833, 0.0077535114251077175, -0.14980539679527283, -0.049980174750089645, -0.03351153805851936, 0.02140883356332779, -0.027198784053325653, -0.05608033016324043, 0.029020125046372414, 0.013277031481266022, -0.018927138298749924, -0.1874474138021469, -0.16278646886348724, -0.039648547768592834, -0.025790834799408913, -0.044876810163259506, -0.06004469096660614, -0.09699922800064087, -0.01983291283249855, 0.021110035479068756, -0.056403499096632004, -0.011957438662648201, -0.05794522538781166, 0.09276821464300156, 0.0389375314116478, 0.09342093765735626, -0.008558034896850586, 0.06181175261735916, -0.14895260334014893, -0.047626793384552, -0.12893594801425934, 0.15345875918865204, -0.07127343863248825, 0.11191632598638535, -0.065483458340168, -0.025287805125117302, -0.1016865074634552, 0.04347057640552521, -0.01570189744234085, 0.2828312814235687, -0.10161040723323822, -0.09649679064750671, 0.3282998204231262, -0.07461509108543396, -0.14172358810901642, 0.16849501430988312, 0.004205547273159027, 0.08955448120832443, 0.16263239085674286, 0.1818506419658661, -0.0012848381884396076, -0.06956163048744202, 0.045141350477933884, 0.08351071923971176, -0.09179972112178802, 0.015585124492645264, 0.01718270778656006, -0.05950184166431427, -0.09182070195674896, 0.02753378450870514, 0.0854385495185852, 0.07801827788352966, -0.05957663431763649, -0.026512254029512405, -0.03168560937047005, -0.006043965928256512, 0.10895225405693054, 0.011973453685641289, 0.1328286975622177, -0.08207929879426956, -0.08660642802715302, -0.03974216431379318, 0.04469615966081619, -0.03995107486844063, 0.0007829946698620915, -0.15716466307640076, 0.10835591703653336, -0.010541719384491444, 0.06749524176120758, -0.15281334519386292, -0.04297579824924469, 0.0038947109133005142, 0.10623512417078018, 0.08787329494953156, 0.08225158601999283, 0.08060343563556671, -0.07254302501678467, 0.04319814592599869, 0.0068772099912166595, 0.15216031670570374, -0.029426302760839462, -0.09506487101316452, -0.08556603640317917, 0.10496469587087631, -0.05121556669473648, 0.13659445941448212, -0.06292685866355896, 0.012047159485518932, 0.005994581151753664, 0.12116459012031555, -0.0227517019957304, 0.028212256729602814, 0.05684419721364975, -0.008862345479428768, -0.044084496796131134, -0.01015773881226778, 0.04624349996447563, -0.022493718191981316, -0.06686893105506897, 0.19475822150707245, -0.08942436426877975, 0.05514654517173767, 0.1818540394306183, -0.15634270012378693, -0.05349823832511902, -0.017480935901403427, -0.040785983204841614, 0.022203752771019936, 0.016772981733083725, -0.030874598771333694, 0.24709388613700867, -0.017619719728827477, 0.14659127593040466, -0.06292649358510971, -0.04017825424671173, -0.03787187486886978, -0.08157094568014145, 0.03421003371477127, 0.11524492502212524, 0.05800969898700714, -0.17616240680217743, 0.11592582613229752, 0.033430445939302444, -0.024071112275123596, 0.18766100704669952, 0.004253431223332882, -0.011216049082577229, 0.049389977008104324, 0.0871143788099289, -0.0475972555577755, -0.054418012499809265, -0.3412359356880188, -0.026005955412983894, 0.06377820670604706, 0.04447663947939873, 0.11060469597578049, -0.09523160755634308, -0.00466920156031847, -0.028619201853871346, -0.045808739960193634, 0.03424997255206108, 0.10968838632106781, 0.05117304250597954, 0.12175296247005463, -0.028825365006923676, -0.10760381817817688, 0.05859785154461861, 0.017973385751247406, -0.1430927962064743, 0.2045094072818756, -0.11807607859373093, -0.35720953345298767, -0.06412989646196365, -0.08514988422393799, -0.03494494408369064, 0.03857208415865898, 0.11707484722137451, -0.14720278978347778, -0.004894827026873827, -0.0028757865075021982, 0.07220571488142014, 0.02048041671514511, 0.02117714285850525, -0.020663287490606308, -0.001442396896891296, -0.07185546308755875, -0.09373456239700317, -0.07265372574329376, -0.033262915909290314, -0.07349512726068497, 0.15754470229148865, -0.15133972465991974, 0.027250543236732483, 0.21629005670547485, 0.016517383977770805, 0.06069764122366905, -0.031734343618154526, 0.2165050059556961, -0.13316042721271515, -0.008085529319941998, 0.18706119060516357, 0.005243218038231134, -0.008825612254440784, 0.09284297376871109, -0.04621012136340141, -0.09310446679592133, 0.06243473291397095, -0.044836610555648804, -0.04124264419078827, -0.2664557993412018, -0.13339391350746155, -0.0713782086968422, 0.06526410579681396, 0.02611086703836918, 0.06138278916478157, 0.12545835971832275, 0.06568313390016556, -0.009795756079256535, -0.028348084539175034, 0.0503346212208271, 0.059956975281238556, 0.21209432184696198, -0.07861748337745667, 0.1490325927734375, -0.038246288895606995, -0.16132846474647522, 0.04804568737745285, 0.010615513660013676, 0.09089358896017075, 0.05016670376062393, 0.015837470069527626, 0.03372875973582268, 0.06332068145275116, 0.12231207638978958, 0.04026294872164726, 0.06462481617927551, -0.061483874917030334, 0.02063913643360138, -0.039221860468387604, -0.083299420773983, 0.020490551367402077, 0.10006803274154663, -0.14223475754261017, -0.029432039707899094, -0.03707689419388771, 0.08471505343914032, 0.07876329869031906, 0.12161802500486374, -0.2426495999097824, -0.06211472302675247, 0.005216607823967934, -0.0898486003279686, -0.09283584356307983, 0.0747111365199089, 0.01665058545768261, -0.18055883049964905, 0.0469791479408741, -0.01587214693427086, 0.10409104824066162, -0.03622756525874138, 0.05604642257094383, -0.12864908576011658, -0.09353920072317123, -0.0003462779277469963, 0.12135309725999832, -0.33387500047683716, 0.160807266831398, -0.024248940870165825, 0.0007313518435694277, -0.12866590917110443, 0.0026174508966505527, 0.015395166352391243, 0.06946815550327301, 0.09797365218400955, -0.027767373248934746, 0.03352072834968567, -0.02654348500072956, -0.0873248353600502, 0.05357125774025917, 0.04931865260004997, 0.05099877715110779, 0.022244790568947792, -0.016084497794508934, 0.0043797302059829235, -0.046284135431051254, -0.13545821607112885, -0.06793171167373657, -0.06204117462038994, 0.04660557582974434, 0.05822927877306938, 0.06066858023405075, 0.0010284795425832272, -0.046809084713459015, -0.012201334349811077, 0.25606000423431396, -0.0035962641704827547, -0.10912639647722244, -0.061884887516498566, 0.013689207844436169, 0.069427989423275, -0.07533331215381622, -0.025839339941740036, -0.05939546972513199, 0.014396429061889648, -0.02936345711350441, -0.14927923679351807, 0.08581043034791946, -0.10172121226787567, -0.05763314291834831, -0.0076633598655462265, 0.18911059200763702, 0.013351469300687313, 0.05249445140361786, 0.08478853851556778, -0.0299775879830122, -0.11076440662145615, -0.08283156156539917, -0.023298759013414383, -0.0008248116937465966, -0.06218565255403519, 0.09156572818756104, -0.003489841939881444, -0.1338844895362854, -0.09156559407711029, 0.016829604282975197, 0.3153288662433624, 0.09904734045267105, -0.016819046810269356, 0.14304517209529877, 0.08939626812934875, -0.03878980502486229, -0.27846112847328186, -0.05327949672937393, -0.010779958218336105, -0.005662440322339535, -0.11892632395029068, -0.13461865484714508, 0.09891077876091003, -0.04347408935427666, -0.03254693001508713, 0.1335865706205368, -0.3087243437767029, -0.09940013289451599, 0.24491316080093384, -0.049568288028240204, 0.2871873378753662, -0.09744606912136078, -0.07851483672857285, -0.02468744106590748, -0.11654505133628845, 0.11371085047721863, 0.0036127360071986914, 0.10889241844415665, -0.0192701518535614, 0.17781125009059906, 0.05037398636341095, -0.023119574412703514, 0.08863881230354309, 0.0092606907710433, -0.06782171130180359, -0.08823621273040771, 0.019971555098891258, -0.014518244192004204, 0.03448003903031349, 0.12389108538627625, -0.04536247253417969, 0.048815447837114334, -0.18603384494781494, -0.04799685254693031, -0.10623951256275177, 0.10021871328353882, 0.03935462236404419, -0.01314764004200697, -0.018128450959920883, -0.05237387493252754, -0.03974269703030586, 0.0230411384254694, 0.08715875446796417, -0.10816735774278641, 0.1804114282131195, 0.10610948503017426, 0.15796475112438202, -0.15238253772258759, -0.009983764961361885, -0.00754476897418499, -0.04539472609758377, 0.07817646116018295, -0.06803888827562332, 0.014795755967497826, 0.12196692079305649, -0.014400207437574863, 0.09986412525177002, 0.06022569537162781, -0.05894839018583298, 0.01670542173087597, 0.10006627440452576, -0.20746974647045135, -0.03557749092578888, -0.055991288274526596, -0.03430528566241264, 0.029524166136980057, 0.08073290437459946, 0.19846156239509583, -0.01339755393564701, -0.04127328097820282, -0.017516139894723892, 0.002747894264757633, -0.07382618635892868, 0.1917341947555542, -0.008630444295704365, 0.035629432648420334, -0.1499967873096466, 0.03211704269051552, -0.007598869502544403, -0.050900302827358246, 0.07543811202049255, 0.13864775002002716, -0.10707548260688782, -0.14140759408473969, -0.06825371831655502, 0.1180209293961525, -0.08516854792833328, -0.024197064340114594, -0.08694639801979065, -0.12834638357162476, 0.07622717320919037, 0.06915276497602463, 0.07618952542543411, 0.07598640024662018, -0.03568889573216438, -0.046537846326828, -0.029520776122808456, 0.029624080285429955, 0.0695299580693245, -0.02275967039167881, -0.036892231553792953, 0.11789774149656296, -0.01138516329228878, 0.12277190387248993, -0.10043113678693771, -0.1047685295343399, -0.1643647849559784, 0.07658392190933228, -0.1697278767824173, -0.07571276277303696, -0.07178860902786255, -0.04503663256764412, 0.02067515440285206, -0.037226781249046326, -0.03781126067042351, -0.03014143370091915, -0.12582515180110931, 0.02514008805155754, -0.03470728546380997, 0.02226846292614937, -0.10788451135158539, 0.06672830879688263, 0.06252827495336533, -0.04234783723950386, 0.14835256338119507, 0.18553096055984497, -0.12508566677570343, 0.09040992707014084, -0.13932958245277405, -0.06857980787754059, 0.11489235609769821, 0.017384661361575127, 0.05814831331372261, 0.005979788489639759, 0.05288344994187355, 0.11327435076236725, 0.09550216048955917, 0.11088427901268005, 0.06529658287763596, -0.08798087388277054, 0.013232238590717316, -0.04582437127828598, -0.10221666097640991, -0.05036855861544609, 0.005301277153193951, 0.0008170217042788863, 0.063886858522892, 0.12014741450548172, -0.08908510208129883, 0.06012488156557083, -0.0830940380692482, 0.005814727861434221, -0.00810791365802288, -0.10102443397045135, 0.009185681119561195, -0.029902521520853043, 0.07313664257526398, -0.003921641502529383, 0.1481061428785324, 0.055238768458366394, 0.009426149539649487, 0.02678941935300827, 0.07134320586919785, 0.07088438421487808, -0.009174374863505363, 0.09162581712007523, 0.08688797801733017, -0.03376520797610283, -0.07189349830150604, 0.033936455845832825, 0.02753598801791668, -0.07054173946380615, 0.13934326171875, 0.035272493958473206, -0.08387258648872375, 0.10339052975177765, 0.008677791804075241, 0.03475981205701828, -0.10318764299154282, -0.1222957894206047, -0.11221977323293686, 0.09935034066438675, -0.02116122469305992, 0.14923346042633057, 0.13568934798240662, -0.021721448749303818, -0.01024174876511097, -0.03554088622331619, -0.07640039175748825, -0.182083860039711, -0.1593005657196045, -0.09674050658941269, -0.17319829761981964, 0.04489561915397644, -0.09876654297113419, 0.009092873893678188, -0.022066211327910423, 0.10061046481132507, -0.09435199201107025, 0.14145097136497498, 0.024776339530944824, -0.11786425858736038, 0.09144100546836853, -0.06112753972411156, 0.04207291081547737, 0.004082012921571732, -0.0032957540825009346, -0.021008329465985298, 0.06682734191417694, 0.04752259701490402, 0.03157242015004158, -0.09760241955518723, 0.03904968872666359, -0.1343352198600769, -0.06514010578393936, -0.034302715212106705, 0.06993764638900757, 0.004969933535903692, 0.11293525993824005, 0.055162250995635986, -0.012725342065095901, 0.018783727660775185, 0.24718695878982544, -0.05683836713433266, -0.15606693923473358, -0.13797463476657867, 0.26412779092788696, -0.01366421114653349, 0.07660382986068726, -0.024699870496988297, -0.0023223243188112974, -0.08871755003929138, 0.3080032467842102, 0.2675771415233612, -0.0634090006351471, -0.00574599951505661, -0.018517928197979927, 0.024005595594644547, 0.05626029521226883, 0.12114665657281876, 0.11084870249032974, 0.3093368709087372, -0.04882589355111122, 0.008467958308756351, -0.019446369260549545, -0.061461880803108215, -0.08807370811700821, -0.03585177659988403, 0.050348054617643356, -0.04998618736863136, -0.019172517582774162, 0.13729216158390045, -0.22915881872177124, 0.07765939086675644, -0.1424342393875122, -0.13563965260982513, -0.10503886640071869, 0.011521115899085999, 0.07987253367900848, 0.014660240150988102, 0.08804410696029663, -0.012389793992042542, -0.04302006587386131, 0.036822155117988586, -0.03309537470340729, -0.14925630390644073, -0.06885119527578354, 0.09651131182909012, -0.11894886940717697, -0.014963089488446712, -0.01096082292497158, 0.17150996625423431, 0.08886699378490448, 0.031434256583452225, -0.039100468158721924, 0.086384117603302, 0.0030711241997778416, -0.05496649071574211, 0.08324380964040756, 0.07555553317070007, 0.05432247743010521, -0.02310565672814846, 0.09675729274749756, -0.09334156662225723, 0.023788634687662125, -0.11870630830526352, 0.04200660437345505, -0.12458760291337967, 0.06972932070493698, -0.07748540490865707, 0.06367151439189911, 0.14496402442455292, -0.03958473727107048, 0.045430123805999756, -0.038381390273571014, 0.02192980796098709, -0.030561568215489388, -0.1207846850156784, -0.10802249610424042, -0.24181050062179565, -0.07936836779117584, 0.04148741811513901, -0.015333188697695732, -0.21018750965595245, 0.025824831798672676, -0.12765921652317047, 0.02861269749701023, -0.0633501261472702, 0.12304191291332245, 0.07175665348768234, 0.05254248529672623, 0.0007014198927208781, -0.026110107079148293, 0.028272025287151337, 0.10157531499862671, -0.14245624840259552, -0.08619525283575058 ]
null
null
transformers
TRIGGER WARNING --------------- This model was created by training GPT2-medium on a custom dataset containing tens of thousands of blog posts about people's experiences living with mental illnesses. As such, the texts that this model generates may be triggering and/or NSFW. Please explore at your own discretion. The blog posts that were compiled were specifically about 6 different mental health conditions: depression, ptsd, cptsd, borderline personality disorder, bipolar (non-specific), and dissociation. These are very serious illnesses so please treat this with respect, and I encourage everyone to learn more about these conditions. Thank you, and enjoy!
{"language": "en", "widget": [{"text": "I feel "}, {"text": "I want "}, {"text": "I believe "}]}
text-generation
banalyst/wonder-egg
[ "transformers", "pytorch", "gpt2", "text-generation", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
TRIGGER WARNING --------------- This model was created by training GPT2-medium on a custom dataset containing tens of thousands of blog posts about people's experiences living with mental illnesses. As such, the texts that this model generates may be triggering and/or NSFW. Please explore at your own discretion. The blog posts that were compiled were specifically about 6 different mental health conditions: depression, ptsd, cptsd, borderline personality disorder, bipolar (non-specific), and dissociation. These are very serious illnesses so please treat this with respect, and I encourage everyone to learn more about these conditions. Thank you, and enjoy!
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 49 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.02475670352578163, 0.021775994449853897, -0.007059284020215273, 0.01180080883204937, 0.1715736985206604, 0.03246768191456795, 0.09108353406190872, 0.13304546475410461, -0.004883658140897751, -0.016561711207032204, 0.16769328713417053, 0.20489005744457245, -0.0030231480486691, 0.07370427250862122, -0.06891095638275146, -0.2764018177986145, 0.05247420817613602, 0.07061995565891266, -0.017013268545269966, 0.12806929647922516, 0.0739644467830658, -0.05563471466302872, 0.09292574971914291, -0.013980518095195293, -0.1778498888015747, 0.016624871641397476, 0.04981158301234245, -0.1223616898059845, 0.12442155182361603, 0.05598319321870804, 0.09582921117544174, 0.013509029522538185, -0.05756191164255142, -0.13026553392410278, 0.022256100550293922, 0.026271162554621696, -0.057974930852651596, 0.06850673258304596, 0.09956607967615128, -0.09310630708932877, 0.09703163057565689, 0.08761205524206161, -0.03042648918926716, 0.0530402846634388, -0.1622297763824463, -0.05651210993528366, -0.02659640647470951, 0.00454084062948823, 0.06345225870609283, 0.0922161191701889, -0.01569903828203678, 0.11206038296222687, -0.09322184324264526, 0.10000225901603699, 0.1554674357175827, -0.3145952522754669, 0.006835123989731073, 0.09325872361660004, 0.037305742502212524, 0.02924501523375511, -0.022390395402908325, 0.0634554848074913, 0.034756045788526535, 0.022287026047706604, 0.014734541065990925, -0.07226059585809708, -0.09685170650482178, 0.05057109147310257, -0.0958705022931099, -0.075267493724823, 0.22633932530879974, -0.07288973033428192, 0.05937040224671364, -0.02182759903371334, -0.11159462481737137, -0.05292976275086403, -0.008323820307850838, 0.00042787924758158624, -0.06399190425872803, 0.08731500059366226, 0.028679583221673965, -0.07817689329385757, -0.1369359791278839, -0.03771807253360748, -0.18853022158145905, 0.19493389129638672, 0.017528997734189034, 0.059727542102336884, -0.1896311193704605, 0.11831236630678177, 0.003938484471291304, -0.0835525169968605, 0.03002442978322506, -0.09622533619403839, 0.04652011767029762, -0.005215132143348455, -0.062180645763874054, -0.06716510653495789, 0.06803932785987854, 0.13551567494869232, -0.006375118158757687, 0.017361963167786598, -0.03438643366098404, 0.09149863570928574, 0.026479102671146393, 0.10359357297420502, -0.011089221574366093, -0.010387847200036049, 0.045451704412698746, -0.14821015298366547, -0.006320280954241753, -0.06752147525548935, -0.15842296183109283, -0.06090424954891205, 0.059932392090559006, 0.08445456624031067, 0.011295073665678501, 0.09186914563179016, -0.047634340822696686, -0.031574416905641556, 0.05388902127742767, -0.06612688302993774, -0.010856048204004765, -0.008164327591657639, 0.015889186412096024, 0.1321278214454651, -0.01060862559825182, 0.02455804869532585, -0.13197419047355652, 0.07503842562437057, -0.07748552411794662, -0.003167348448187113, -0.03643878549337387, -0.04592232033610344, 0.014235424809157848, -0.09847619384527206, 0.01782030612230301, -0.15215416252613068, -0.17248772084712982, 0.02140350453555584, 0.006103968247771263, -0.031316060572862625, -0.04375796392560005, -0.037711627781391144, -0.026556603610515594, 0.04127205163240433, -0.069469153881073, 0.012334401719272137, -0.05490204691886902, 0.10921807587146759, -0.035915594547986984, 0.06703599542379379, -0.1138170138001442, 0.08033181726932526, -0.1228807121515274, -0.009521805681288242, -0.089187853038311, 0.06765218824148178, 0.008832919411361217, 0.09810461103916168, -0.025562088936567307, -0.03585495054721832, -0.08012468367815018, 0.054918188601732254, -0.029163913801312447, 0.1952102780342102, -0.05747575685381889, -0.12971548736095428, 0.2555959224700928, -0.07224182039499283, -0.13138240575790405, 0.1036304160952568, 0.015917401760816574, 0.04643147438764572, 0.08614059537649155, 0.17128022015094757, 0.023423168808221817, 0.004989448934793472, 0.08318941295146942, 0.1024169847369194, -0.12458416819572449, -0.08533135801553726, 0.015541856177151203, -0.035738199949264526, -0.1394561529159546, 0.054521508514881134, 0.06450530141592026, 0.08003360778093338, -0.05199698358774185, -0.0236055925488472, -0.036505963653326035, 0.010384837165474892, 0.08067304641008377, 0.007073577959090471, 0.12774622440338135, -0.05131116136908531, -0.0351095013320446, -0.02387653850018978, -0.017410883679986, -0.022319525480270386, 0.038297779858112335, -0.024976154789328575, 0.144541397690773, -0.05493677034974098, 0.05069015175104141, -0.18243137001991272, -0.07108470052480698, -0.0027940222062170506, 0.12680847942829132, -0.011005111038684845, 0.09315073490142822, 0.055364515632390976, -0.02280275523662567, -0.011276355013251305, -0.012260403484106064, 0.1581266224384308, -0.026576092466711998, -0.07132825255393982, -0.051135655492544174, 0.06771866977214813, -0.053177133202552795, -0.00011576346878428012, -0.05089122802019119, 0.012663308531045914, 0.05297533795237541, 0.11152546107769012, -0.009890194982290268, 0.03739846125245094, -0.02669081836938858, 0.014205466024577618, -0.08552035689353943, 0.003078873036429286, 0.10067883878946304, -0.004922674503177404, -0.06420966237783432, 0.2001977264881134, -0.15881513059139252, 0.2220752239227295, 0.19801972806453705, -0.2996901273727417, 0.008226542733609676, -0.07750441879034042, -0.033868324011564255, 0.01390016544610262, 0.04717644676566124, -0.0341244712471962, 0.13397572934627533, -0.0001671792269917205, 0.18755541741847992, -0.05269179120659828, -0.05928400158882141, 0.005859028548002243, -0.055954623967409134, -0.005590829066932201, 0.06221532076597214, 0.12563452124595642, -0.13901729881763458, 0.19617488980293274, 0.19160354137420654, 0.036279309540987015, 0.1614362746477127, 0.01151591632515192, -0.029504457488656044, 0.07349423319101334, -0.009864090010523796, -0.03811991959810257, -0.06794720143079758, -0.18849626183509827, -0.019344767555594444, 0.08520065248012543, 0.05022380128502846, 0.10704094171524048, -0.13740862905979156, -0.039894770830869675, -0.019768957048654556, -0.00976532232016325, 0.015582730993628502, 0.09396158903837204, 0.051734983921051025, 0.1247156411409378, -0.0030131367966532707, -0.001536445808596909, 0.11545835435390472, 0.02593437395989895, -0.09734364598989487, 0.19949719309806824, -0.12101639807224274, -0.3628603518009186, -0.16813601553440094, -0.16428300738334656, -0.049804918467998505, 0.0642201155424118, 0.10288447141647339, -0.11876867711544037, -0.027449671179056168, 0.02447819896042347, 0.11552916467189789, -0.09929122775793076, 0.030133932828903198, -0.0793774202466011, 0.042624589055776596, -0.08525561541318893, -0.07367627322673798, -0.06255385279655457, -0.03066887892782688, -0.06544684618711472, 0.15920574963092804, -0.11005846410989761, 0.04295359551906586, 0.18713481724262238, 0.0312747061252594, 0.055917415767908096, -0.03407035022974014, 0.18848863244056702, -0.10323154926300049, -0.006914193741977215, 0.20017904043197632, -0.043500274419784546, 0.08120299875736237, 0.10290893167257309, -0.004199240356683731, -0.09134737402200699, 0.021489975973963737, -0.029924124479293823, -0.09809733182191849, -0.23382112383842468, -0.12726885080337524, -0.11668778210878372, 0.07133220136165619, 0.05980675667524338, 0.05644579231739044, 0.16627761721611023, 0.09485958516597748, -0.028067734092473984, 0.04733084887266159, -0.005399790592491627, 0.08163555711507797, 0.2116672694683075, -0.02895347774028778, 0.13845708966255188, -0.044490229338407516, -0.13365459442138672, 0.09720277041196823, 0.0647907629609108, 0.13387760519981384, 0.060887835919857025, 0.04465141519904137, 0.005137486383318901, 0.07425639033317566, 0.14636678993701935, 0.12661507725715637, 0.02179379016160965, -0.016216447576880455, -0.026202790439128876, -0.01032145507633686, -0.060605473816394806, 0.045667972415685654, 0.034318242222070694, -0.15080679953098297, -0.05024860426783562, -0.10842032730579376, 0.08019711077213287, 0.10085909068584442, 0.06820562481880188, -0.23021289706230164, 0.0014392047887668014, 0.07768179476261139, -0.02666773833334446, -0.12036287784576416, 0.08830516785383224, -0.022323021665215492, -0.15019826591014862, 0.045269425958395004, -0.06515369564294815, 0.11804332584142685, -0.0705902948975563, 0.08574675023555756, -0.04040885716676712, -0.06234096735715866, 0.021050335839390755, 0.1225423663854599, -0.3004131019115448, 0.21169844269752502, -0.0007325364276766777, -0.06442920863628387, -0.11483842134475708, 0.01329885981976986, 0.011260363273322582, 0.10479757934808731, 0.11196339875459671, 0.0034150953870266676, -0.04293595254421234, -0.11251498013734818, -0.02692062221467495, 0.02862722985446453, 0.13267509639263153, -0.06421144306659698, -0.003470148192718625, -0.051188983023166656, 0.0015975087881088257, -0.025829454883933067, -0.05000091344118118, 0.029489796608686447, -0.16357684135437012, 0.0867508128285408, 0.008885513059794903, 0.0931304544210434, 0.01738009974360466, -0.024236058816313744, -0.11006181687116623, 0.23592853546142578, -0.08272545039653778, -0.11042473465204239, -0.11883529275655746, -0.0364999957382679, 0.0651044026017189, -0.07150653749704361, 0.052636485546827316, -0.0810762345790863, 0.028166428208351135, -0.05023537203669548, -0.210233673453331, 0.12687622010707855, -0.08861299604177475, -0.057178013026714325, -0.03652799874544144, 0.19267110526561737, -0.0780320093035698, 0.0071064201183617115, 0.015245867893099785, 0.030895939096808434, -0.11375197023153305, -0.10291676968336105, 0.03511594235897064, -0.008997295051813126, 0.06394096463918686, 0.032449111342430115, -0.0656287744641304, 0.02061527781188488, -0.023794954642653465, -0.012383661232888699, 0.31593769788742065, 0.16413018107414246, -0.047328125685453415, 0.17381098866462708, 0.11517556011676788, -0.08653349429368973, -0.3156975209712982, -0.08819037675857544, -0.11019104719161987, -0.03208387643098831, -0.06435072422027588, -0.2094677835702896, 0.08416374027729034, 0.04296618700027466, -0.014879420399665833, 0.14672943949699402, -0.24430371820926666, -0.07942993938922882, 0.15273989737033844, -0.007498474791646004, 0.383838415145874, -0.11925133317708969, -0.10655616968870163, -0.052972014993429184, -0.14112725853919983, 0.14779356122016907, 0.0038995344657450914, 0.10564578324556351, -0.041487280279397964, 0.11028849333524704, 0.04698486253619194, -0.050963904708623886, 0.0871778354048729, 0.00901566818356514, -0.008102099411189556, -0.10615245997905731, -0.028431715443730354, 0.043248843401670456, 0.012928525917232037, 0.022058509290218353, -0.023581117391586304, 0.03146994486451149, -0.12029249221086502, -0.0490250326693058, -0.07671220600605011, 0.06609407812356949, 0.04178333282470703, -0.08317110687494278, -0.003475768491625786, -0.06620330363512039, -0.016071593388915062, 0.0016490468988195062, 0.18812362849712372, -0.03603435307741165, 0.14380168914794922, 0.0692576915025711, 0.10064544528722763, -0.1374540627002716, 0.009991574101150036, -0.07289420813322067, -0.05816662684082985, 0.08805005997419357, -0.10822049528360367, 0.060802605003118515, 0.12238024175167084, -0.04054837301373482, 0.06268671154975891, 0.11557701230049133, 0.013979239389300346, -0.0014341940404847264, 0.12171857059001923, -0.26797550916671753, 0.00804979633539915, -0.07441525906324387, -0.04654378443956375, 0.09235677123069763, 0.083685502409935, 0.1670708954334259, 0.023332294076681137, -0.045308325439691544, -0.004546184558421373, 0.010280659422278404, -0.04165256395936012, 0.06902558356523514, 0.011849727481603622, 0.016005773097276688, -0.14536504447460175, 0.07238991558551788, 0.012861641123890877, -0.12773451209068298, 0.016180578619241714, 0.1329706758260727, -0.1388436108827591, -0.11416247487068176, -0.036740742623806, 0.08184423297643661, -0.15637212991714478, -0.023823270574212074, -0.05030139535665512, -0.12589925527572632, 0.09382318705320358, 0.12539169192314148, 0.0776093527674675, 0.10342948138713837, -0.051991209387779236, -0.023240962997078896, -0.04503847658634186, -0.025356359779834747, -0.008649970404803753, 0.026877980679273605, -0.08182281255722046, 0.05256769806146622, -0.018435975536704063, 0.14170870184898376, -0.09257227182388306, -0.07500877231359482, -0.15419673919677734, 0.03408411145210266, -0.12401173263788223, -0.07909072935581207, -0.08558586984872818, -0.056456610560417175, -0.009236719459295273, -0.023542076349258423, -0.04220154881477356, -0.04511674493551254, -0.12027659267187119, 0.02458345703780651, -0.0497446283698082, 0.02151406742632389, -0.07035335153341293, 0.0027168807573616505, 0.09223131090402603, -0.03715268895030022, 0.14139708876609802, 0.13686618208885193, -0.08011692017316818, 0.11905833333730698, -0.12568727135658264, -0.08341789245605469, 0.1108071506023407, 0.016084618866443634, 0.04227926954627037, 0.07023494690656662, 0.03152864798903465, 0.06700090318918228, 0.020481519401073456, 0.05684690177440643, 0.008230681531131268, -0.12141445279121399, 0.04263344407081604, -0.03560594841837883, -0.14864064753055573, -0.058225907385349274, -0.047264229506254196, 0.03685939684510231, 0.023609746247529984, 0.10751762986183167, -0.037262506783008575, 0.10714813321828842, -0.05631037801504135, 0.023597221821546555, 0.00468487711623311, -0.18142090737819672, -0.04498005285859108, -0.07720267027616501, 0.026681145653128624, 0.02301940880715847, 0.27059924602508545, 0.03589962422847748, 0.020518187433481216, 0.02283250167965889, 0.09921247512102127, 0.046299465000629425, 0.012437593191862106, 0.20372003316879272, 0.11734872311353683, -0.05738286301493645, -0.10566538572311401, 0.08767566084861755, 0.023038573563098907, 0.0022363553289324045, 0.1302972137928009, 0.009786593727767467, -0.007485245820134878, 0.09078863263130188, -0.037452202290296555, 0.0049662054516375065, -0.12128829956054688, -0.12830637395381927, -0.023122357204556465, 0.05679672211408615, 0.0006949229864403605, 0.10587357729673386, 0.13887187838554382, -0.026676904410123825, 0.035694900900125504, -0.01451646164059639, -0.04455423355102539, -0.1763908863067627, -0.1639343500137329, -0.0775918960571289, -0.12919656932353973, 0.013535698875784874, -0.10388235002756119, 0.043417803943157196, 0.0889660120010376, 0.06274251639842987, -0.06086258590221405, 0.09773243963718414, 0.05885498598217964, -0.10778224468231201, 0.056807056069374084, -0.029356004670262337, 0.06282379478216171, -0.021361524239182472, -0.023060930892825127, -0.09434284269809723, 0.006500772666186094, 0.001862326287664473, 0.04623982682824135, -0.050197191536426544, 0.015488714911043644, -0.14935217797756195, -0.09849414229393005, -0.052288029342889786, 0.06924542039632797, -0.059045322239398956, 0.1067616194486618, -0.0014256195863708854, -0.011969448998570442, 0.0363389253616333, 0.20138400793075562, -0.0668824091553688, -0.04519839584827423, -0.040705081075429916, 0.22357724606990814, 0.04243519902229309, 0.10906081646680832, -0.011560790240764618, 0.0029203658923506737, -0.0764424279332161, 0.3651666045188904, 0.2845011055469513, -0.06503085047006607, 0.010859494097530842, 0.03784650191664696, 0.03098106011748314, 0.13891355693340302, 0.13461892306804657, 0.09901760518550873, 0.2655550539493561, -0.08578096330165863, -0.05793455243110657, -0.02230701595544815, -0.014269283041357994, -0.0794128030538559, 0.11161618679761887, 0.04048392176628113, -0.06677433848381042, -0.045808903872966766, 0.09253464639186859, -0.23108267784118652, 0.14928162097930908, -0.08250284194946289, -0.15658150613307953, -0.062460873275995255, 0.020124157890677452, 0.10419663041830063, 0.0010382836917415261, 0.08877632021903992, -0.01142740249633789, -0.10325799137353897, 0.05254621058702469, 0.029224438592791557, -0.236823171377182, 0.004280368331819773, 0.054358288645744324, -0.05244016647338867, 0.01895728148519993, -0.019771026447415352, 0.05561700835824013, 0.06523369252681732, 0.05272016301751137, -0.038820743560791016, 0.024426285177469254, -0.005227750167250633, -0.04925939813256264, 0.022136535495519638, 0.044338252395391464, 0.025861451402306557, -0.12093951553106308, 0.05876215919852257, -0.1296541690826416, 0.04744231328368187, -0.03151528909802437, -0.03816784918308258, 0.002454610075801611, -0.006519368849694729, -0.06504569947719574, 0.06310195475816727, 0.08436178416013718, 0.003248278284445405, -0.010247940197587013, -0.08446147292852402, -0.007913918234407902, -0.010410798713564873, -0.07269034534692764, -0.09439866989850998, -0.13774390518665314, -0.10698703676462173, 0.12328177690505981, -0.018790537491440773, -0.18962348997592926, 0.012347078882157803, -0.08767691254615784, 0.05564672872424126, -0.17583008110523224, 0.0842139944434166, 0.05508223921060562, 0.021255923435091972, -0.0026963416021317244, -0.017235424369573593, 0.038487758487463, 0.07905291020870209, -0.1065288633108139, -0.08977312594652176 ]
null
null
transformers
# Rick Sanchez DialoGPT Model
{"tags": ["conversational"]}
text-generation
banden/DialoGPT-medium-RickBot
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick Sanchez DialoGPT Model
[ "# Rick Sanchez DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick Sanchez DialoGPT Model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Rick Sanchez DialoGPT Model" ]
[ -0.05704520270228386, 0.1080707237124443, -0.005703833419829607, 0.024355918169021606, 0.1347416192293167, -0.009864812716841698, 0.13915762305259705, 0.13641619682312012, -0.014821183867752552, -0.025234131142497063, 0.13788719475269318, 0.23441068828105927, -0.0040086545050144196, 0.0579121895134449, -0.09891517460346222, -0.29657089710235596, 0.032161157578229904, 0.05994465947151184, -0.0033263780642300844, 0.11953802406787872, 0.0843273475766182, -0.043882302939891815, 0.08131924271583557, 0.0018995096907019615, -0.14373421669006348, 0.011346335522830486, 0.04636937007308006, -0.13702289760112762, 0.11601521074771881, 0.08168332278728485, 0.03479677438735962, 0.06183161959052086, -0.03211790323257446, -0.10245182365179062, 0.03838932886719704, -0.008999419398605824, -0.03427799046039581, 0.06022527068853378, 0.031745243817567825, -0.1152564138174057, 0.09468080848455429, 0.0923495814204216, -0.005728692281991243, 0.049891795963048935, -0.17913517355918884, -0.010704654268920422, -0.021677182987332344, 0.055929314345121384, 0.08334671705961227, 0.09012723714113235, -0.03841714933514595, 0.09080751240253448, -0.04324564337730408, 0.07660475373268127, 0.08530165255069733, -0.28937292098999023, -0.030906030908226967, 0.06582700461149216, 0.05632982775568962, 0.06222769245505333, -0.012348905205726624, 0.10370101034641266, 0.04712950810790062, -0.014536825940012932, -0.020318668335676193, -0.09295357018709183, -0.08654087781906128, 0.01945175603032112, -0.06887777894735336, -0.00602162629365921, 0.2560831606388092, -0.030627258121967316, 0.0731891319155693, -0.08859450370073318, -0.10249431431293488, 0.004030927084386349, -0.03497277945280075, -0.05301825329661369, -0.08995666354894638, 0.0700719803571701, -0.003147976705804467, -0.07389466464519501, -0.1290099322795868, -0.021585101261734962, -0.1774103045463562, 0.19141581654548645, 0.030293408781290054, 0.023750921711325645, -0.2208036631345749, 0.08902101963758469, 0.045917924493551254, -0.11613845080137253, 0.04777151718735695, -0.0842585563659668, 0.027763593941926956, 0.03184368088841438, -0.015087970532476902, -0.059821996837854385, 0.06819985806941986, 0.09782791137695312, 0.01923726685345173, 0.009750176221132278, -0.024553336203098297, 0.05652669072151184, 0.04215332120656967, 0.09357214719057083, -0.002894732868298888, -0.0361810103058815, 0.015707144513726234, -0.09473340213298798, 0.013285800814628601, -0.06974472850561142, -0.20285023748874664, -0.01343308761715889, 0.030690573155879974, 0.06404593586921692, 0.0605037696659565, 0.11637318879365921, -0.016192488372325897, -0.07014694064855576, 0.04394420608878136, -0.011219148524105549, -0.02391223981976509, 0.008600763976573944, 0.005595726426690817, 0.1813707947731018, 0.011592349968850613, 0.04672492668032646, -0.08541523665189743, 0.0021381767001003027, -0.04958593100309372, -0.03829700127243996, -0.02637419104576111, -0.043410710990428925, 0.004262345843017101, -0.01814153417944908, 0.01607448235154152, -0.16901227831840515, -0.1370115429162979, -0.022032571956515312, -0.023941930383443832, -0.05494401231408119, -0.10453005880117416, -0.1106305792927742, 0.010454395785927773, 0.0411594994366169, -0.07347700744867325, -0.006038041319698095, -0.05349244922399521, 0.10734197497367859, -0.0012300090165808797, 0.07703675329685211, -0.07842288166284561, 0.07365122437477112, -0.06573385000228882, -0.025447865948081017, -0.10093135386705399, 0.13386206328868866, 0.00915649812668562, 0.06603219360113144, -0.046941112726926804, -0.007621072698384523, -0.10842157155275345, 0.08102882653474808, -0.06603449583053589, 0.25178468227386475, -0.10469595342874527, -0.11071616411209106, 0.2805403769016266, -0.03361968696117401, -0.1264788806438446, 0.10199446976184845, -0.01158563606441021, 0.11561296135187149, 0.15402540564537048, 0.2052226960659027, 0.02928532473742962, -0.02268681675195694, 0.10423996299505234, 0.1035756841301918, -0.057625655084848404, -0.03801124170422554, 0.041060492396354675, -0.03943869471549988, -0.08562899380922318, 0.029806632548570633, 0.017141755670309067, 0.06328166276216507, -0.04716106131672859, -0.014570656232535839, 0.02421940304338932, 0.004092389717698097, 0.07496705651283264, -0.024279450997710228, 0.13514691591262817, -0.023218633607029915, -0.0656837597489357, -0.06177486851811409, 0.01964477449655533, -0.04003036767244339, 0.020440705120563507, -0.08908485621213913, 0.06243852525949478, -0.03572830185294151, 0.057841621339321136, -0.13513502478599548, -0.06440334767103195, -0.052385009825229645, 0.2330688089132309, 0.06854398548603058, 0.08484960347414017, 0.04045721888542175, -0.06823423504829407, -0.0003254515759181231, 0.0230836383998394, 0.19774475693702698, -0.012182417325675488, -0.07111652195453644, -0.1060645654797554, 0.09369703382253647, -0.06330309808254242, 0.08606080710887909, -0.059275124222040176, 0.007605451624840498, -0.026874825358390808, 0.06961184740066528, -0.0269512627273798, 0.040469661355018616, 0.0006163049256429076, -0.023647982627153397, -0.07465367019176483, -0.0149429552257061, 0.10460925847291946, 0.010055731050670147, -0.10779253393411636, 0.2434917390346527, -0.19769670069217682, 0.11526761204004288, 0.16096965968608856, -0.22217433154582977, -0.008169818669557571, -0.11098682880401611, -0.011975600384175777, 0.010464251041412354, 0.03273862600326538, -0.04017077758908272, 0.22867226600646973, -0.010608958080410957, 0.1853235810995102, -0.052233923226594925, -0.01996525749564171, -0.027332648634910583, -0.06422537565231323, 0.019287196919322014, 0.10971896350383759, 0.12856750190258026, -0.161560520529747, 0.1724557876586914, 0.06672785431146622, 0.08213970810174942, 0.16275310516357422, 0.023765239864587784, 0.018852105364203453, 0.055699001997709274, 0.014213238842785358, -0.01790653169155121, -0.06882723420858383, -0.18352845311164856, -0.02486453764140606, 0.06696174293756485, 0.03141540661454201, 0.11100433766841888, -0.11560750752687454, -0.04409283027052879, 0.01155218854546547, -0.007257997989654541, 0.04618756100535393, 0.12712539732456207, -0.003129301592707634, 0.12465701252222061, -0.008397440426051617, -0.08989791572093964, 0.06356243789196014, 0.025369787588715553, -0.09814108163118362, 0.182835653424263, -0.1034003496170044, -0.32152217626571655, -0.1001197025179863, -0.18328212201595306, -0.03920574113726616, 0.07533707469701767, 0.11137884855270386, -0.13466644287109375, 0.003127885051071644, 0.03510995954275131, 0.09001503139734268, -0.10122435539960861, -0.0034966380335390568, -0.012268266640603542, -0.018193284049630165, -0.13420377671718597, -0.08291704952716827, -0.06102906912565231, -0.04295424371957779, -0.03346853703260422, 0.10250663757324219, -0.17059698700904846, 0.0539877749979496, 0.26917001605033875, 0.09507580101490021, 0.05430034175515175, -0.04344318434596062, 0.1592034101486206, -0.10573985427618027, 0.012773060239851475, 0.22541004419326782, -0.028745699673891068, 0.05365750566124916, 0.0892547219991684, -0.01033081579953432, -0.0708671286702156, 0.020250815898180008, -0.02718094177544117, -0.0714477151632309, -0.2280162125825882, -0.13114430010318756, -0.10388115793466568, 0.05737285315990448, 0.06432957202196121, 0.0326518714427948, 0.1693384051322937, 0.10345755517482758, -0.042771194130182266, 0.014231901615858078, 0.04525286331772804, 0.08099555969238281, 0.24561424553394318, -0.07478486001491547, 0.13777120411396027, -0.020003067329525948, -0.17604942619800568, 0.06821287423372269, 0.08701446652412415, 0.07066261768341064, 0.0939561203122139, 0.13011965155601501, 0.02880261279642582, 0.03623313829302788, 0.09038466960191727, 0.045518048107624054, 0.022583454847335815, -0.03737230971455574, -0.06662941724061966, -0.044609375298023224, -0.041899174451828, 0.021367410197854042, 0.03182210028171539, -0.14257657527923584, -0.05327456444501877, 0.004599247593432665, 0.04934147000312805, 0.04460717737674713, 0.04937165975570679, -0.1958761215209961, -0.01135042030364275, 0.08069007098674774, 0.0008143498562276363, -0.08933985978364944, 0.0729956179857254, -0.010731075890362263, -0.11471421271562576, 0.046170588582754135, -0.02729932591319084, 0.12909768521785736, -0.07558268308639526, 0.08029930293560028, -0.14030702412128448, -0.06785701215267181, 0.011536509729921818, 0.11896203458309174, -0.2636930048465729, 0.20840856432914734, -0.008380764164030552, -0.049465201795101166, -0.1043824851512909, -0.009141412563621998, 0.0023304640781134367, 0.0944279134273529, 0.1368221640586853, -0.028159884735941887, -0.02392721176147461, 0.024199169129133224, -0.06698837131261826, 0.03141431510448456, 0.08238276839256287, -0.08262956887483597, 0.0013082197401672602, -0.04166802391409874, 0.0039241621270775795, 0.009456396102905273, -0.06101514399051666, 0.01121380366384983, -0.195927694439888, 0.0798632875084877, 0.05245203897356987, 0.06079527735710144, 0.04320540651679039, -0.030458878725767136, -0.12451554834842682, 0.21634705364704132, -0.01915883459150791, -0.09401129186153412, -0.09610380977392197, -0.02037319913506508, 0.01868581213057041, -0.08247993141412735, -0.029365237802267075, -0.05376124754548073, 0.03249189257621765, -0.0736650601029396, -0.1903923898935318, 0.12846902012825012, -0.11052907258272171, -0.028645969927310944, -0.05812210589647293, 0.2216455489397049, -0.030725445598363876, 0.015262283384799957, 0.059073857963085175, -0.026270287111401558, -0.09585471451282501, -0.09591566771268845, -0.007837353274226189, 0.022675657644867897, 0.027353649958968163, -0.013003773055970669, -0.04600683972239494, -0.03253196179866791, -0.07994730770587921, -0.018432755023241043, 0.3114815652370453, 0.10657966881990433, -0.053566571325063705, 0.1609998643398285, 0.08925200253725052, -0.07996044307947159, -0.24130167067050934, -0.11830049753189087, -0.06823843717575073, -0.04297657683491707, -0.04875720292329788, -0.17909106612205505, 0.07042492181062698, -0.01572689227759838, -0.0246592964977026, 0.0796264261007309, -0.34351006150245667, -0.09367087483406067, 0.17016243934631348, -0.044724494218826294, 0.4543110132217407, -0.1197502538561821, -0.10162397474050522, -0.06263615190982819, -0.1313125193119049, 0.18072150647640228, -0.0014751619892194867, 0.10033301264047623, 0.004829281009733677, 0.16866052150726318, 0.05691388249397278, 0.007217222824692726, 0.091414675116539, 0.01798011176288128, -0.0663527175784111, -0.07840543240308762, -0.10491379350423813, -0.03505389019846916, 0.005603624042123556, -0.0021735846530646086, -0.070311039686203, 0.020419610664248466, -0.15853211283683777, -0.06377539038658142, -0.08453743904829025, 0.02356107160449028, 0.030920391902327538, -0.05867331475019455, 0.013508422300219536, -0.06677638739347458, 0.018330350518226624, -0.001379468129016459, 0.19494818150997162, -0.11257137358188629, 0.17101415991783142, 0.04597465693950653, 0.1332368403673172, -0.08957637846469879, -0.0664193332195282, -0.09592875093221664, -0.04850716516375542, 0.07897713780403137, -0.12196645140647888, 0.028588851913809776, 0.10687820613384247, -0.026256760582327843, 0.08356694877147675, 0.0898364707827568, -0.007075367029756308, 0.034319691359996796, 0.09755269438028336, -0.21034450829029083, -0.08239499479532242, -0.0750756561756134, 0.03635190799832344, 0.08414819836616516, 0.09669569879770279, 0.19893206655979156, 0.012653730809688568, -0.042145922780036926, 0.017787763848900795, 0.012123693712055683, -0.02111114002764225, 0.07857295125722885, 0.018965713679790497, 0.005439399741590023, -0.1433452069759369, 0.052979227155447006, 0.010302988812327385, -0.08479844033718109, 0.010003048926591873, 0.1269667148590088, -0.0933220237493515, -0.11143316328525543, -0.03672108054161072, 0.14701539278030396, -0.17777521908283234, -0.0062329513020813465, -0.0675484910607338, -0.13784073293209076, 0.0540735125541687, 0.08879818022251129, 0.04015353322029114, 0.03582773730158806, -0.09314056485891342, -0.023970693349838257, -0.027464643120765686, -0.025517743080854416, 0.06011157110333443, -0.016900639981031418, -0.06295756995677948, 0.07077009975910187, -0.024854036048054695, 0.11451400071382523, -0.08354704827070236, -0.08851433545351028, -0.16065533459186554, 0.05232835188508034, -0.0971856415271759, -0.051306720823049545, -0.09556061774492264, -0.050184283405542374, -0.01728140190243721, -0.018070710822939873, -0.02822391875088215, -0.04367216303944588, -0.10629069805145264, 0.04816720634698868, -0.03879733756184578, 0.02260771207511425, -0.06238244101405144, 0.02379411831498146, 0.03523104265332222, -0.015178869478404522, 0.1198926493525505, 0.12207183241844177, -0.09698602557182312, 0.09035449475049973, -0.16394847631454468, -0.06776027381420135, 0.11319122463464737, 0.01993078552186489, 0.060301147401332855, 0.06648512184619904, 0.011058407835662365, 0.06193515285849571, 0.047181855887174606, 0.04019154608249664, 0.0011482342379167676, -0.0998595729470253, 0.05959494039416313, -0.025470629334449768, -0.11498484015464783, -0.05095286667346954, 0.004914015997201204, 0.01974906586110592, 0.04762796312570572, 0.09024947136640549, -0.07045641541481018, 0.10133861005306244, -0.0719725638628006, 0.042781490832567215, 0.013207492418587208, -0.1567831039428711, -0.009022838436067104, -0.07855422794818878, 0.051177605986595154, 0.018598362803459167, 0.16054493188858032, 0.01182752288877964, 0.026072200387716293, 0.010315481573343277, 0.07682152092456818, 0.04349011927843094, -0.014741134829819202, 0.21185405552387238, 0.10036209225654602, -0.025498010218143463, -0.09448950737714767, 0.10583885759115219, 0.05347297713160515, 0.04882397875189781, 0.13073812425136566, 0.00329616479575634, -0.014263730496168137, 0.08692904561758041, 0.002064551692456007, 0.02444906160235405, -0.11166326701641083, -0.12022344022989273, -0.034913238137960434, 0.03916458785533905, -0.018694007769227028, 0.07413183152675629, 0.13662488758563995, -0.011044695042073727, 0.01868942379951477, -0.006493001710623503, -0.044404536485672, -0.19629815220832825, -0.21793615818023682, -0.07676366716623306, -0.13648608326911926, -0.0023855564650148153, -0.11510718613862991, 0.041815925389528275, 0.008664922788739204, 0.07993809133768082, -0.08146228641271591, 0.05119030550122261, 0.06000566482543945, -0.1558382213115692, 0.07134677469730377, -0.023100191727280617, 0.1049720048904419, -0.07798602432012558, 0.0240509994328022, -0.06541567295789719, 0.07175038009881973, 0.01739879511296749, 0.029955588281154633, -0.01989707536995411, 0.008211890235543251, -0.12742145359516144, -0.08334774523973465, -0.06036959961056709, 0.06557390093803406, 0.020560268312692642, 0.1673227995634079, 0.009943484328687191, -0.023078134283423424, 0.028660951182246208, 0.24841417372226715, -0.0622759610414505, -0.06388328224420547, -0.06632015854120255, 0.20368391275405884, -0.019206369295716286, 0.09927452355623245, -0.04167409613728523, 0.010532806627452374, -0.07129248231649399, 0.3616236448287964, 0.2931338846683502, -0.10953228175640106, 0.007534074131399393, -0.023572852835059166, 0.042554765939712524, 0.11510708183050156, 0.1042989119887352, 0.07050987333059311, 0.29585832357406616, -0.05323539301753044, -0.035456158220767975, -0.00002445635254844092, 0.0010794458212330937, -0.06095365062355995, 0.08202893286943436, 0.03708086162805557, -0.050908174365758896, -0.00022808888752479106, 0.14211015403270721, -0.2293669879436493, 0.05093219503760338, -0.1549203097820282, -0.17726412415504456, -0.08134126663208008, -0.008463856764137745, 0.11377648264169693, 0.03384561836719513, 0.10970432311296463, -0.004765030462294817, -0.0765727236866951, 0.0467694029211998, 0.03101363778114319, -0.19725514948368073, 0.009119519963860512, 0.06916453689336777, -0.02850601077079773, -0.030441200360655785, -0.02334762178361416, 0.06269988417625427, 0.07391422241926193, 0.05389029160141945, -0.01931828074157238, 0.03749391436576843, -0.00035577634116634727, -0.041961804032325745, 0.07208401709794998, -0.0301519688218832, 0.008827175945043564, -0.0679439827799797, 0.08395512402057648, -0.18772968649864197, 0.04201417416334152, 0.018391063436865807, -0.04758477583527565, -0.04135806858539581, 0.04440581426024437, -0.07019207626581192, 0.08052141219377518, 0.07222563028335571, -0.01630892977118492, -0.02569187991321087, -0.013696021400392056, -0.005180103238672018, 0.000036183802876621485, -0.011484390124678612, -0.10217433422803879, -0.1737447828054428, -0.13257969915866852, 0.06753268837928772, 0.018290087580680847, -0.20535635948181152, 0.03693879395723343, -0.12776882946491241, 0.02865419164299965, -0.10825670510530472, 0.07439035177230835, 0.07718028128147125, 0.02790611796081066, -0.0036239037290215492, -0.022755812853574753, 0.03530992195010185, 0.0841168612241745, -0.12845547497272491, -0.06076517328619957 ]
null
null
transformers
# Loki DialoGPT Model
{"tags": ["conversational"]}
text-generation
banden/DialoGPT-small-LokiBot
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Loki DialoGPT Model
[ "# Loki DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Loki DialoGPT Model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Loki DialoGPT Model" ]
[ 0.01486613042652607, 0.030066967010498047, -0.004986276384443045, 0.037759166210889816, 0.15038181841373444, -0.013893868774175644, 0.1857791543006897, 0.07775568217039108, 0.036045853048563004, -0.0182502344250679, 0.13298356533050537, 0.09662327915430069, 0.027916889637708664, 0.06248512491583824, -0.015100874938070774, -0.34966424107551575, 0.054286520928144455, -0.015624943189322948, -0.049614377319812775, 0.11303138732910156, 0.08171427249908447, -0.04237086698412895, 0.0890401303768158, -0.002632239367812872, -0.1230132207274437, 0.03277929872274399, -0.00304834614507854, -0.08525054901838303, 0.11917194724082947, 0.06573276221752167, 0.030652478337287903, 0.017774084582924843, -0.06259123980998993, -0.0788833349943161, 0.048960015177726746, -0.039861276745796204, -0.025605691596865654, 0.022463012486696243, 0.009047086350619793, -0.0462990365922451, 0.1930657923221588, 0.15361908078193665, -0.01467986311763525, 0.021146638318896294, -0.10116122663021088, 0.12015543133020401, 0.028235312551259995, 0.05995951220393181, 0.055512234568595886, 0.0961417481303215, -0.03437409922480583, 0.08432146161794662, -0.08036944270133972, 0.10787800699472427, 0.08657405525445938, -0.40765634179115295, -0.06103263050317764, 0.12423483282327652, 0.04424996301531792, 0.10489892214536667, -0.057087868452072144, 0.09583694487810135, 0.0123214740306139, 0.01058962382376194, -0.06077750772237778, -0.05842345207929611, -0.1679048240184784, -0.000054686315706931055, -0.09737254679203033, 0.03568092733621597, 0.27907872200012207, -0.04824528843164444, 0.04436390846967697, -0.06763115525245667, -0.07617444545030594, 0.027780748903751373, -0.014023186638951302, -0.06570441275835037, -0.07800441235303879, 0.06467516720294952, 0.016019219532608986, -0.10035112500190735, -0.11638735979795456, -0.024363191798329353, -0.19414488971233368, 0.1939847767353058, 0.041939567774534225, 0.04204876720905304, -0.22828644514083862, 0.12519115209579468, -0.0039171623066067696, -0.08820884674787521, -0.02571326680481434, -0.09676405042409897, 0.00816379301249981, 0.018595412373542786, -0.05507577583193779, -0.07432478666305542, 0.05062243714928627, 0.1122737005352974, 0.058119986206293106, 0.01828167773783207, -0.03490214794874191, 0.08692379295825958, 0.03588268533349037, 0.03531067818403244, 0.003972339443862438, -0.14568190276622772, 0.03768519312143326, -0.06945929676294327, 0.05305728316307068, -0.07313504070043564, -0.18377630412578583, -0.06235792487859726, 0.0008708470268175006, 0.03292878717184067, 0.03221816197037697, 0.12984296679496765, -0.015475266613066196, -0.0126345818862319, 0.08714722841978073, -0.048633333295583725, -0.0506463423371315, -0.016107143834233284, -0.022581495344638824, 0.10382140427827835, 0.039696019142866135, 0.043629735708236694, -0.11599461734294891, -0.009134326130151749, -0.05040239170193672, 0.028510594740509987, 0.03079373389482498, -0.03086976706981659, -0.01335159596055746, -0.0209438968449831, -0.008593227714300156, -0.14566917717456818, -0.13811847567558289, 0.013567918911576271, -0.009950094856321812, -0.05144621059298515, -0.14085964858531952, -0.09459368139505386, -0.04303678497672081, 0.046521008014678955, -0.05737517401576042, -0.04728703200817108, -0.037426672875881195, 0.03700326010584831, -0.05917734280228615, 0.1395740807056427, -0.0788402408361435, 0.08104703575372696, -0.13294434547424316, -0.04427710175514221, -0.16378390789031982, 0.1308826059103012, -0.03382956609129906, 0.0695917010307312, -0.0008622508612461388, -0.0219427477568388, -0.13356240093708038, 0.04444122686982155, -0.059054985642433167, 0.2598888576030731, -0.10735499858856201, -0.11027521640062332, 0.29777294397354126, -0.10353975743055344, -0.13651050627231598, 0.13359828293323517, 0.027862878516316414, 0.02331433817744255, 0.11656647175550461, 0.22501645982265472, -0.046393848955631256, 0.02303689531981945, 0.060088351368904114, 0.09303846955299377, -0.08990717679262161, 0.06234386935830116, 0.022273249924182892, -0.028484612703323364, -0.022931737825274467, 0.018908360973000526, 0.03277050703763962, 0.07913781702518463, -0.033572155982255936, -0.010462872684001923, -0.00852642860263586, -0.012860490940511227, 0.08815956115722656, -0.017942199483513832, 0.14921467006206512, -0.041365623474121094, -0.017413143068552017, 0.012458129785954952, 0.01917247846722603, -0.04953534156084061, 0.07661653310060501, -0.044688452035188675, 0.09538550674915314, -0.011364314705133438, 0.05900847539305687, -0.1245562955737114, -0.08304496854543686, -0.026057250797748566, 0.21545538306236267, 0.08719059079885483, 0.12401587516069412, 0.07376852631568909, -0.025590410456061363, -0.06013491377234459, 0.04148067161440849, 0.13907253742218018, -0.023899206891655922, -0.09556692838668823, -0.1570730209350586, 0.061619311571121216, -0.06780988723039627, 0.13891704380512238, -0.1385304033756256, 0.021506329998373985, 0.0030774606857448816, 0.11538216471672058, -0.015383276157081127, 0.0187011007219553, 0.04976410046219826, -0.03261536732316017, -0.10709954798221588, -0.006778129376471043, 0.06252196431159973, -0.04992325231432915, -0.14921332895755768, 0.2972072660923004, -0.1325116604566574, 0.12803682684898376, 0.19622591137886047, -0.19080288708209991, -0.010855929926037788, -0.0631210207939148, -0.020086416974663734, -0.026843681931495667, 0.0816117525100708, -0.02512061782181263, 0.1841832399368286, -0.013322900980710983, 0.144127756357193, -0.046491410583257675, -0.03327372670173645, -0.0623331144452095, -0.04998951777815819, -0.009396704845130444, 0.09989386796951294, 0.06733222305774689, -0.20323505997657776, 0.15889398753643036, 0.12007931619882584, 0.002905449131503701, 0.27055642008781433, 0.08408074080944061, -0.010521376505494118, 0.09740404039621353, 0.0020996578969061375, -0.046663299202919006, -0.1111602634191513, -0.2536253333091736, -0.048179950565099716, 0.059114228934049606, 0.013306035660207272, 0.09323349595069885, -0.08298036456108093, -0.05829113349318504, -0.018648041412234306, -0.02698504365980625, 0.06831858307123184, 0.13406580686569214, 0.002576302969828248, 0.1311359703540802, -0.028014160692691803, -0.088065966963768, 0.06915098428726196, -0.0016338988207280636, -0.08612877130508423, 0.18219560384750366, -0.11936262249946594, -0.18824197351932526, -0.07556237280368805, -0.1286810338497162, -0.10989357531070709, 0.04159403219819069, 0.07383321970701218, -0.11783333122730255, 0.00189957523252815, -0.07454308122396469, 0.05690717324614525, -0.14073526859283447, 0.014232020825147629, -0.03419553488492966, -0.045393459498882294, -0.14667314291000366, -0.07970884442329407, -0.053735148161649704, -0.03604525327682495, -0.12271431088447571, 0.1263425499200821, -0.13460120558738708, 0.016745135188102722, 0.22317014634609222, 0.027134306728839874, 0.04102727398276329, -0.04218797758221626, 0.21808885037899017, -0.11049038916826248, 0.05460190773010254, 0.0931292325258255, -0.029301129281520844, 0.07923130691051483, 0.14175571501255035, 0.005087734200060368, -0.06416650116443634, 0.039457760751247406, -0.03470071405172348, -0.08726956695318222, -0.21997174620628357, -0.07924868166446686, -0.11856480687856674, 0.06826479732990265, 0.016864599660038948, 0.08828774839639664, 0.10954536497592926, 0.07796146720647812, -0.05078720301389694, -0.016860201954841614, 0.09940002858638763, 0.07330556213855743, 0.16895900666713715, -0.03232485428452492, 0.14357908070087433, -0.022663358598947525, -0.10797560214996338, 0.04666043072938919, 0.07088379561901093, 0.10691481083631516, 0.04666727036237717, -0.027559757232666016, 0.007367618381977081, 0.0211698766797781, 0.1316876858472824, 0.034660011529922485, 0.024526402354240417, -0.07195723056793213, -0.008716552518308163, -0.046927470713853836, -0.006399269215762615, 0.08609764277935028, 0.11595253646373749, -0.10105220973491669, 0.026278287172317505, 0.03682911768555641, 0.07835384458303452, 0.08958559483289719, 0.06887181848287582, -0.15407027304172516, -0.07319199293851852, 0.0970940813422203, -0.09740131348371506, -0.07724462449550629, 0.09483039379119873, 0.060623086988925934, -0.15682187676429749, 0.1227918490767479, -0.012621437199413776, 0.08797905594110489, 0.03068847581744194, 0.05049403756856918, -0.10411889106035233, -0.10163535177707672, -0.0024651254061609507, 0.04989268258213997, -0.2212640941143036, 0.17687827348709106, -0.03255782648921013, -0.04666636884212494, -0.05586082115769386, 0.005490650422871113, 0.05464579164981842, 0.19672012329101562, 0.10899200290441513, 0.011682926677167416, 0.09039774537086487, 0.022634845227003098, 0.002406833693385124, 0.02495572902262211, 0.07807300239801407, -0.09652343392372131, -0.028418749570846558, -0.009637805633246899, 0.012983952648937702, -0.054012224078178406, 0.0036665380466729403, -0.02104058489203453, -0.19155414402484894, 0.07581046968698502, 0.0670255646109581, 0.09988244622945786, 0.03081352822482586, -0.055520832538604736, -0.13090679049491882, 0.26712191104888916, -0.08578804135322571, -0.11658891290426254, -0.04425373300909996, -0.022276239469647408, 0.0348195880651474, -0.050507038831710815, 0.006714787799865007, -0.05721201375126839, -0.017190931364893913, -0.09237893670797348, -0.17740942537784576, 0.10729662328958511, -0.08355394005775452, -0.08350963890552521, -0.012997690588235855, 0.18033769726753235, 0.021595144644379616, 0.06100495532155037, 0.017634714022278786, 0.016116665676236153, -0.105299212038517, -0.07753552496433258, -0.01426906418055296, 0.0569017194211483, -0.01859537698328495, 0.06286049634218216, 0.04220940172672272, -0.0451677031815052, -0.059442710131406784, -0.037072066217660904, 0.29144489765167236, 0.13185922801494598, -0.04743441194295883, 0.17260338366031647, 0.1186298206448555, -0.0388813354074955, -0.2627389132976532, -0.10144256055355072, -0.0765811949968338, -0.02852388098835945, 0.0016447646776214242, -0.15317951142787933, 0.14291788637638092, -0.038605403155088425, -0.024938318878412247, 0.1428646743297577, -0.27212443947792053, -0.12434854358434677, 0.13660991191864014, -0.02439216896891594, 0.38087162375450134, -0.09829504042863846, -0.07520009577274323, -0.009589570574462414, -0.14243964850902557, 0.0663929209113121, -0.029514119029045105, 0.11906233429908752, -0.0323016531765461, 0.20070074498653412, 0.01666216179728508, 0.015289394184947014, 0.13275988399982452, 0.03802923485636711, -0.07522693276405334, -0.1204216256737709, -0.047981295734643936, 0.03834594786167145, 0.02431773766875267, 0.03455134853720665, -0.10818493366241455, -0.018552536144852638, -0.13761624693870544, -0.03530719503760338, -0.08357024937868118, 0.052329692989587784, 0.046206288039684296, -0.04915684834122658, -0.01141113881021738, -0.03531961515545845, 0.004835211206227541, 0.047611694782972336, 0.203245609998703, -0.1105724647641182, 0.1832006573677063, 0.05861802399158478, 0.18027010560035706, -0.10548772662878036, 0.04395605996251106, -0.04772373288869858, -0.03338851034641266, 0.10374674946069717, -0.08974427729845047, -0.01865694299340248, 0.10931514203548431, -0.023862745612859726, 0.025255005806684494, 0.07960131764411926, 0.015356387943029404, 0.05917464941740036, 0.08470113575458527, -0.24489635229110718, -0.12299062311649323, -0.03825861215591431, -0.05509456619620323, 0.0789642184972763, 0.09788701683282852, 0.20610694587230682, -0.0807584896683693, -0.029531775042414665, 0.02695406787097454, 0.020210033282637596, -0.028034571558237076, 0.06534469127655029, -0.011855981312692165, 0.009991759434342384, -0.13129495084285736, 0.09604810178279877, 0.0023022755049169064, -0.12845255434513092, 0.061932433396577835, 0.16780222952365875, -0.12125377357006073, -0.12648259103298187, -0.16537152230739594, 0.000039479087718063965, -0.0763055682182312, -0.01196261402219534, 0.021597953513264656, -0.12898394465446472, 0.04721430689096451, 0.009436305612325668, 0.011872677132487297, 0.037814561277627945, -0.13021624088287354, -0.00977497361600399, 0.050588298588991165, -0.015930745750665665, 0.01638910546898842, -0.0011518909595906734, -0.05460795760154724, 0.04596345126628876, -0.05720005929470062, 0.14475534856319427, -0.09090253710746765, -0.11530829221010208, -0.06859371811151505, 0.03579932823777199, -0.06478391587734222, -0.09244459122419357, -0.14884550869464874, -0.06012878566980362, -0.035970840603113174, -0.05913036689162254, -0.04345264285802841, -0.02385770156979561, -0.12396922707557678, 0.03510846197605133, -0.05612960457801819, 0.04205641150474548, -0.06632153689861298, 0.04467068985104561, 0.04369915649294853, -0.018237093463540077, 0.1865409016609192, 0.20043176412582397, -0.14294642210006714, 0.06637680530548096, -0.1498527228832245, -0.06544224172830582, 0.08723019808530807, -0.0008425407577306032, 0.04546883702278137, 0.13918110728263855, -0.04550870880484581, 0.026481471955776215, 0.05820097774267197, 0.06467492133378983, 0.0796303078532219, -0.06979119777679443, 0.02106347493827343, -0.03718730807304382, -0.10476769506931305, -0.058026526123285294, -0.016073239967226982, 0.060418400913476944, 0.004246698692440987, 0.06503680348396301, -0.0918797105550766, 0.05073734372854233, -0.08820690959692001, 0.04166311398148537, 0.03815112262964249, -0.14511393010616302, -0.0255482979118824, -0.07048128545284271, 0.03864092007279396, -0.04036969318985939, 0.18268004059791565, -0.04495810717344284, -0.03542110696434975, 0.033298324793577194, 0.08494384586811066, 0.035112909972667694, 0.008365891873836517, 0.15972556173801422, 0.10416655987501144, -0.04415135085582733, -0.09428881108760834, 0.028926778584718704, 0.03871328756213188, 0.041619863361120224, 0.1323782056570053, -0.045128270983695984, -0.042581524699926376, 0.13597673177719116, -0.04019947722554207, 0.03854338452219963, -0.14833876490592957, -0.13755899667739868, -0.0507751926779747, 0.06922952085733414, -0.0344083271920681, 0.10647465288639069, 0.25197452306747437, 0.007922380231320858, -0.031113769859075546, -0.02142507955431938, -0.04157278314232826, -0.16379745304584503, -0.24136625230312347, -0.05786559730768204, -0.13577137887477875, 0.03057532012462616, -0.12561172246932983, 0.0736498311161995, 0.041099268943071365, 0.12485956400632858, -0.05453050509095192, 0.14786040782928467, 0.09160232543945312, -0.11675551533699036, 0.09346549212932587, -0.040004875510931015, 0.06943925470113754, 0.006995427422225475, -0.02002209611237049, -0.06365171819925308, 0.013348598033189774, -0.021063784137368202, 0.044706594198942184, -0.04032357037067413, 0.024085914716124535, -0.12012656033039093, -0.08057055622339249, -0.043020810931921005, 0.08026106655597687, -0.026609869673848152, 0.07399801164865494, 0.06494064629077911, -0.033030055463314056, 0.006164721213281155, 0.17997804284095764, -0.08715544641017914, -0.07198242843151093, -0.03653552383184433, 0.18780992925167084, 0.01959707960486412, 0.10004328936338425, -0.027109825983643532, -0.011405110359191895, -0.11558588594198227, 0.2169588804244995, 0.35200145840644836, -0.11149173229932785, 0.03159010782837868, -0.019736891612410545, 0.04448391869664192, 0.08851344138383865, 0.05395565181970596, 0.08098806440830231, 0.2453003227710724, -0.06121022254228592, 0.009631899185478687, -0.034255195409059525, -0.027053004130721092, -0.09421605616807938, 0.03929241746664047, 0.05247931182384491, -0.06965744495391846, -0.028886863961815834, 0.10350214689970016, -0.2520119547843933, 0.07059116661548615, -0.14576798677444458, -0.18843625485897064, -0.07589298486709595, -0.0187019482254982, 0.05794772878289223, 0.04519185051321983, 0.05160324275493622, 0.0015868166228756309, -0.056391436606645584, 0.028576243668794632, 0.025954201817512512, -0.18106651306152344, 0.037276655435562134, 0.10601743310689926, -0.09667512774467468, -0.0724271759390831, -0.027807148173451424, 0.04352898895740509, 0.05550481751561165, 0.10564315319061279, -0.00018550478853285313, 0.0014280875911936164, 0.02074241451919079, -0.007832472212612629, -0.021722104400396347, 0.09512357413768768, 0.058806631714105606, -0.1008937656879425, 0.10886753350496292, -0.041866715997457504, 0.05239246040582657, 0.03986401855945587, 0.005639142356812954, -0.03856796771287918, 0.04990500584244728, -0.051926374435424805, 0.07977505773305893, 0.10360269248485565, -0.004626309033483267, -0.030645694583654404, -0.018302522599697113, -0.08947297930717468, -0.004235382191836834, -0.0385100394487381, -0.11175248771905899, -0.18706198036670685, -0.09582236409187317, 0.00869648065418005, 0.03152131289243698, -0.139672189950943, -0.0007627275190316141, -0.1395038515329361, 0.06650012731552124, -0.10482923686504364, 0.09951856732368469, 0.11140243709087372, 0.007602361496537924, 0.020655188709497452, -0.09204163402318954, 0.06752172112464905, 0.1061629205942154, -0.19194641709327698, -0.07145290821790695 ]
null
null
transformers
## Overview This model was trained with data from https://registry.opendata.aws/helpful-sentences-from-reviews/ to predict how "helpful" a review is. The model was fine-tuned from the `distilbert-base-uncased` model ### Labels LABEL_0 - Not helpful LABEL_1 - Helpful ### How to use The following code shows how to make a prediction with this model ```python from transformers import ( AutoTokenizer, AutoModelForSequenceClassification, TextClassificationPipeline, ) tokenizer = AutoTokenizer.from_pretrained("banjtheman/distilbert-base-uncased-helpful-amazon") model = AutoModelForSequenceClassification.from_pretrained( "banjtheman/distilbert-base-uncased-helpful-amazon" ) pipe = TextClassificationPipeline(model=model, tokenizer=tokenizer) result = pipe("This was a Christmas gift for my grandson.") print(result) #[{'label': 'LABEL_0', 'score': 0.998775064945221}] # This is NOT A HELPFUL comment ```
{"license": "apache-2.0"}
text-classification
banjtheman/distilbert-base-uncased-helpful-amazon
[ "transformers", "pytorch", "distilbert", "text-classification", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
## Overview This model was trained with data from URL to predict how "helpful" a review is. The model was fine-tuned from the 'distilbert-base-uncased' model ### Labels LABEL_0 - Not helpful LABEL_1 - Helpful ### How to use The following code shows how to make a prediction with this model
[ "## Overview\r\n\r\nThis model was trained with data from URL to predict how \"helpful\" a review is.\r\n\r\nThe model was fine-tuned from the 'distilbert-base-uncased' model", "### Labels\r\nLABEL_0 - Not helpful \r\nLABEL_1 - Helpful", "### How to use\r\n\r\nThe following code shows how to make a prediction with this model" ]
[ "TAGS\n#transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "## Overview\r\n\r\nThis model was trained with data from URL to predict how \"helpful\" a review is.\r\n\r\nThe model was fine-tuned from the 'distilbert-base-uncased' model", "### Labels\r\nLABEL_0 - Not helpful \r\nLABEL_1 - Helpful", "### How to use\r\n\r\nThe following code shows how to make a prediction with this model" ]
[ 46, 44, 16, 18 ]
[ "passage: TAGS\n#transformers #pytorch #distilbert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n## Overview\r\n\r\nThis model was trained with data from URL to predict how \"helpful\" a review is.\r\n\r\nThe model was fine-tuned from the 'distilbert-base-uncased' model### Labels\r\nLABEL_0 - Not helpful \r\nLABEL_1 - Helpful### How to use\r\n\r\nThe following code shows how to make a prediction with this model" ]
[ -0.041265297681093216, 0.14266382157802582, -0.0013327760389074683, 0.04730002582073212, 0.11242728680372238, -0.03496745228767395, 0.07329621911048889, 0.07766947150230408, 0.004562120884656906, -0.03897024691104889, 0.14361701905727386, 0.15199263393878937, -0.007313192822039127, 0.06383156031370163, -0.06333646178245544, -0.2305273860692978, 0.06364639103412628, 0.02619621530175209, 0.06776201725006104, 0.14936892688274384, 0.12718719244003296, -0.0917428731918335, 0.07252329587936401, 0.04758292809128761, -0.06991630047559738, 0.0040688482113182545, -0.01329295989125967, -0.05252337455749512, 0.0805266723036766, 0.02956857532262802, 0.08905541896820068, 0.05971542000770569, 0.044649094343185425, -0.1730949878692627, 0.00009693867468740791, -0.024767130613327026, -0.0033815773203969, 0.05444516986608505, 0.045486822724342346, -0.0703803300857544, 0.1630537360906601, 0.02888852171599865, 0.026676779612898827, 0.038545235991477966, -0.07121936231851578, -0.08342582732439041, -0.06320749968290329, 0.1400480568408966, 0.14709119498729706, 0.08956654369831085, -0.006806060206145048, 0.18716475367546082, -0.1525040864944458, 0.04375622048974037, 0.10294612497091293, -0.3068658113479614, -0.00820862129330635, 0.12504759430885315, 0.037735793739557266, -0.076897032558918, -0.020655089989304543, 0.048199065029621124, 0.05008687078952789, 0.01705005392432213, 0.045755986124277115, -0.05593712627887726, -0.05996697396039963, 0.01854444295167923, -0.13581480085849762, -0.0798707976937294, 0.28706395626068115, 0.021552693098783493, -0.06591662019491196, -0.0824052020907402, -0.007960162125527859, 0.09019196033477783, -0.06601729989051819, -0.0983857810497284, 0.03644934296607971, 0.038030482828617096, 0.06224840506911278, 0.00939156860113144, -0.12319464981555939, -0.06501135975122452, -0.0881233811378479, 0.16654905676841736, -0.0026654894463717937, 0.012217545881867409, -0.08835362643003464, 0.18903660774230957, -0.04846600070595741, -0.09994985163211823, -0.017172275111079216, -0.10119516402482986, 0.07349105924367905, -0.04588930308818817, -0.0735367089509964, 0.02204355224967003, 0.04507967457175255, 0.1901925951242447, 0.09132097661495209, 0.0095202112570405, 0.0527351014316082, 0.03333025053143501, -0.009078407660126686, 0.134236142039299, -0.07804591208696365, -0.07214406877756119, 0.07560314983129501, 0.0147323627024889, 0.0432298518717289, 0.017428912222385406, -0.1122538298368454, -0.04011103883385658, 0.06677530705928802, 0.09618686139583588, 0.039461154490709305, 0.04600406065583229, -0.09140820056200027, -0.02669667638838291, 0.10108575969934464, -0.0695013478398323, -0.001350230653770268, 0.018077082931995392, -0.10756860673427582, 0.018971841782331467, 0.05261960253119469, 0.052563998848199844, -0.08528342097997665, 0.04016893357038498, -0.07010966539382935, -0.006423019338399172, -0.05691584572196007, -0.07963592559099197, 0.009217812679708004, 0.03393775597214699, 0.05744367837905884, -0.11361398547887802, -0.21567639708518982, -0.003988349810242653, 0.05465570092201233, -0.029253331944346428, -0.054868824779987335, -0.031539615243673325, -0.02071216329932213, -0.0031221292447298765, -0.02667914144694805, 0.029571836814284325, -0.06651705503463745, 0.017747126519680023, -0.03490222617983818, 0.01868579536676407, -0.05089160427451134, 0.05113353207707405, -0.1485745906829834, 0.02328869327902794, -0.11788956075906754, 0.06400896608829498, -0.06084277853369713, 0.11137519031763077, -0.12708334624767303, -0.06903225183486938, 0.009302286431193352, 0.0003263780090492219, 0.02243315987288952, 0.23216132819652557, -0.10339488834142685, -0.018369508907198906, 0.09811673313379288, -0.10671267658472061, -0.17062602937221527, 0.07433539628982544, -0.041579920798540115, 0.11390387266874313, 0.12284689396619797, 0.05132246017456055, 0.13067182898521423, -0.0007151568424887955, 0.04163239896297455, 0.11475770175457001, -0.05511970818042755, -0.06338734924793243, 0.04698748141527176, -0.013600698672235012, -0.23902954161167145, 0.06113205850124359, -0.1367194503545761, 0.012093828059732914, -0.043964583426713943, -0.07642534375190735, -0.05345866084098816, -0.07132193446159363, 0.04933009296655655, 0.020793594419956207, 0.06853151321411133, -0.016467677429318428, -0.05141143500804901, -0.00006461521843448281, 0.11172335594892502, -0.002395871328189969, 0.01626507006585598, -0.051494892686605453, 0.15007412433624268, -0.055434249341487885, 0.01112719438970089, -0.19654953479766846, 0.018730012699961662, 0.008721382357180119, 0.04141049459576607, -0.0010921125067397952, -0.0034376438707113266, 0.030225716531276703, -0.021675897762179375, 0.007122749928385019, -0.04538200795650482, 0.10262245684862137, 0.043268702924251556, -0.09970999509096146, -0.09310135245323181, 0.0008461294346489012, -0.007154556922614574, 0.12163436412811279, -0.15789742767810822, -0.024842500686645508, -0.01659819483757019, 0.09429606050252914, -0.041084520518779755, 0.07542560994625092, 0.04390685260295868, 0.0012089589145034552, -0.042381878942251205, -0.025732139125466347, 0.092672199010849, 0.010487520135939121, -0.1502087563276291, 0.11189479380846024, -0.08704505115747452, 0.09944657236337662, 0.13520091772079468, -0.11003205180168152, -0.054892342537641525, 0.01996861770749092, -0.03478878363966942, -0.008123360574245453, 0.02741096541285515, 0.034686826169490814, 0.1456557661294937, 0.010681377723813057, 0.1061532199382782, -0.05233672633767128, -0.0417601652443409, -0.015181889757514, -0.08286644518375397, 0.017096154391765594, 0.08289726078510284, 0.0931173786520958, -0.1763499528169632, 0.07177914679050446, 0.19975921511650085, -0.037951480597257614, 0.013616385869681835, -0.020055249333381653, -0.04660733789205551, 0.005202228669077158, 0.0033463267609477043, -0.030353659763932228, 0.010216864757239819, -0.10129033774137497, -0.019453229382634163, 0.07357737421989441, -0.002208154881373048, 0.0380171462893486, -0.15813179314136505, 0.0025132710579782724, 0.04028252884745598, -0.004769409541040659, -0.0108097018674016, 0.007241454441100359, -0.050754714757204056, 0.08806458860635757, 0.004785814322531223, -0.16470155119895935, 0.10386241972446442, 0.0019417662406340241, -0.1332818865776062, 0.2265675663948059, -0.052045438438653946, -0.22739368677139282, -0.10734126716852188, -0.07419281452894211, -0.03472018986940384, 0.021898960694670677, 0.046800944954156876, -0.0018674591556191444, -0.11680985987186432, -0.0744633674621582, -0.06090084835886955, -0.06593582034111023, -0.02722134441137314, -0.09442266076803207, -0.010084124282002449, 0.054814811795949936, -0.09627727419137955, -0.040296658873558044, -0.023528140038251877, -0.023623686283826828, 0.022731736302375793, -0.08344145119190216, 0.08231515437364578, 0.17278173565864563, 0.0075379908084869385, 0.023941662162542343, -0.011979905888438225, 0.18877549469470978, -0.01753012277185917, -0.015747522935271263, 0.14359265565872192, -0.015373878180980682, 0.05629704147577286, 0.1253376454114914, 0.044640570878982544, -0.07788494974374771, 0.032453734427690506, -0.049217429012060165, -0.03366858884692192, -0.20045803487300873, -0.15922558307647705, -0.04152178391814232, -0.04260891675949097, 0.053248997777700424, 0.03349127620458603, 0.06043660268187523, 0.11209804564714432, 0.0005482314736582339, 0.03724274784326553, -0.036351606249809265, 0.09951568394899368, 0.16271834075450897, -0.013250373303890228, 0.12042460590600967, -0.08667425066232681, -0.08091259747743607, 0.08645462244749069, -0.027994921430945396, 0.10484354943037033, 0.06094986945390701, 0.043899908661842346, 0.08753970265388489, 0.10040970146656036, 0.06127854064106941, 0.11502905935049057, 0.02930312603712082, -0.036358945071697235, -0.04180168733000755, -0.050664469599723816, -0.06961912661790848, 0.05666608735918999, -0.022574329748749733, -0.011827746406197548, -0.13999980688095093, -0.021079368889331818, 0.026600560173392296, 0.09441839903593063, 0.09912396967411041, -0.27619364857673645, -0.04530941694974899, 0.04453740641474724, -0.021354038268327713, -0.02824203483760357, 0.08947481215000153, -0.12749114632606506, -0.1256248503923416, 0.04888458177447319, 0.003135157050564885, 0.14012809097766876, -0.066078782081604, 0.03589341789484024, -0.08532106131315231, -0.04916439205408096, 0.005572889465838671, 0.16514471173286438, -0.2564302682876587, 0.18331147730350494, 0.0030652652494609356, 0.029262978583574295, -0.06519512832164764, -0.01729915291070938, 0.08843375742435455, 0.22331446409225464, 0.1666240692138672, -0.015945687890052795, 0.07660488039255142, -0.06346379965543747, -0.09665776789188385, 0.021620137616991997, -0.014154093340039253, -0.08055023849010468, 0.05315915867686272, -0.0255594402551651, -0.001172772143036127, 0.029959050938487053, -0.03216983750462532, -0.12686961889266968, -0.09922600537538528, 0.03683531656861305, 0.02867356315255165, 0.024209963157773018, -0.04714525490999222, -0.06045357882976532, -0.02027382142841816, 0.16693629324436188, -0.038378406316041946, -0.08628736436367035, -0.11521779745817184, 0.014063195325434208, 0.10836946219205856, -0.0547095462679863, 0.051565371453762054, -0.0376649908721447, 0.02375386655330658, 0.02459479495882988, -0.08868679404258728, 0.04708463326096535, -0.14032916724681854, -0.06361664086580276, -0.012003420852124691, 0.15451961755752563, 0.03291357308626175, 0.018926428630948067, 0.0639786422252655, -0.0027597048319876194, -0.097117118537426, -0.14204029738903046, -0.08219900727272034, 0.07466752082109451, 0.07898547500371933, 0.04795895144343376, -0.007321501150727272, -0.04740699753165245, -0.05097468942403793, 0.0485181100666523, 0.18175461888313293, 0.09958941489458084, -0.0773005336523056, 0.11251845955848694, 0.15264569222927094, -0.06651642173528671, -0.2554401457309723, -0.127365842461586, -0.04966512322425842, 0.008049268275499344, 0.029766138643026352, -0.09994786977767944, 0.16984322667121887, -0.06976649910211563, -0.05781790614128113, -0.003023758763447404, -0.2299787998199463, -0.08731348067522049, 0.2569669187068939, 0.05972880497574806, 0.28206801414489746, -0.07111795246601105, -0.08183161914348602, -0.04466008022427559, -0.17479103803634644, 0.123577781021595, -0.07463311403989792, 0.038611188530921936, -0.007350623607635498, 0.08198536932468414, 0.04785024747252464, -0.05044933408498764, 0.15038757026195526, 0.010571039281785488, 0.042807675898075104, -0.10057047009468079, -0.024087242782115936, 0.027483638375997543, -0.05627434700727463, 0.16470441222190857, -0.03210972994565964, 0.09426945447921753, -0.149377703666687, -0.06955932080745697, -0.014746107161045074, 0.07946337014436722, -0.019344812259078026, -0.0491274893283844, -0.041895486414432526, 0.0019530606223270297, 0.05463669076561928, 0.001887606573291123, 0.06792020797729492, 0.004387260414659977, -0.00810302048921585, 0.08832376450300217, 0.19760042428970337, -0.053401462733745575, -0.06276864558458328, -0.041206806898117065, -0.05912986025214195, 0.08817611634731293, -0.16761639714241028, 0.04819687083363533, 0.11413990706205368, -0.0009843365987762809, 0.10757727175951004, 0.047462545335292816, -0.018165677785873413, -0.04100542142987251, 0.0720977708697319, -0.19141443073749542, -0.08164604008197784, -0.06549336016178131, 0.01993519999086857, -0.02619527280330658, 0.032601580023765564, 0.08824358880519867, -0.09913666546344757, -0.0351681150496006, 0.003927493002265692, 0.028366191312670708, -0.052455686032772064, 0.12437348067760468, 0.075254887342453, 0.025847019627690315, -0.1224055141210556, 0.08317025750875473, 0.03150821849703789, -0.026266010478138924, -0.0020744623616337776, -0.03868267312645912, -0.16134576499462128, -0.11433354020118713, -0.04595978930592537, 0.16357746720314026, -0.14936701953411102, -0.14931006729602814, -0.07592084258794785, -0.09057819098234177, 0.056677620857954025, 0.01652066968381405, 0.1483117938041687, 0.07051350176334381, -0.0752432718873024, -0.08943170309066772, -0.05180496349930763, 0.07266052812337875, 0.11248921602964401, -0.038603443652391434, -0.1648419201374054, 0.012139885686337948, 0.03334478288888931, 0.0632840171456337, -0.0655180811882019, -0.038560472428798676, -0.0754961222410202, 0.026983605697751045, -0.20154157280921936, 0.02422104962170124, -0.10188028216362, 0.022352268919348717, 0.01065784227102995, -0.014572365209460258, -0.05760570988059044, 0.056299906224012375, -0.09743454307317734, -0.002722550882026553, -0.00925691518932581, 0.0978403240442276, -0.0893198549747467, -0.01664683222770691, 0.072818323969841, -0.021565718576312065, 0.08117540180683136, 0.05208740383386612, -0.034871120005846024, 0.05369425192475319, -0.16691917181015015, -0.02837114967405796, 0.04718024283647537, 0.035922206938266754, 0.028126493096351624, -0.16413436830043793, -0.009919007308781147, 0.09747201949357986, -0.03589131310582161, 0.0013832715339958668, 0.04649602994322777, -0.12679828703403473, -0.08368445187807083, 0.009488210082054138, -0.04544141888618469, -0.02317454293370247, 0.009696612134575844, 0.058306735008955, 0.07158661633729935, 0.19045908749103546, -0.008562983013689518, 0.006607417948544025, -0.08317115157842636, 0.008073156699538231, -0.07718847692012787, -0.05813013017177582, -0.17630860209465027, -0.05833064019680023, 0.006296024192124605, -0.025368547067046165, 0.21817612648010254, 0.03308799862861633, 0.10405851155519485, 0.01887134648859501, 0.13497449457645416, 0.16041617095470428, 0.01735028438270092, 0.24030746519565582, 0.034521184861660004, 0.018206104636192322, -0.05858534574508667, 0.05162098631262779, 0.037408486008644104, -0.011843325570225716, 0.07124760001897812, 0.060267575085163116, -0.05476769059896469, 0.10927759110927582, -0.05888558179140091, -0.003734544152393937, -0.06842713057994843, -0.09532850980758667, -0.013989551924169064, 0.06667737662792206, -0.0004010096308775246, 0.08236029744148254, 0.11619693785905838, -0.038032740354537964, 0.023282311856746674, -0.005632401909679174, -0.04816729202866554, -0.14798539876937866, -0.16901157796382904, -0.07877012342214584, -0.11763598769903183, -0.018833689391613007, -0.0758480355143547, -0.03514125570654869, 0.07101999968290329, 0.015144826844334602, -0.08532164245843887, 0.093721903860569, -0.08702513575553894, -0.04296056181192398, 0.08570674061775208, -0.04044453427195549, -0.04193098470568657, -0.03577132523059845, 0.016655050218105316, -0.052191127091646194, -0.03420546278357506, -0.016497546806931496, -0.00973512977361679, 0.0053748697973787785, 0.021231211721897125, -0.04385216161608696, -0.05747770145535469, -0.05125129967927933, 0.04806451499462128, 0.03593564033508301, 0.09552060067653656, 0.01590035855770111, 0.016682125627994537, 0.014742239378392696, 0.23821444809436798, -0.05382853373885155, -0.10070916265249252, -0.10718495398759842, 0.33314067125320435, 0.06016950309276581, 0.01977282017469406, 0.026262106373906136, -0.0734495148062706, 0.042191196233034134, 0.2846691906452179, 0.1866636574268341, -0.07246572524309158, -0.02050122804939747, -0.042176466435194016, -0.001033347798511386, 0.07288829982280731, 0.11154826730489731, 0.04116278141736984, 0.20923568308353424, -0.09284479171037674, -0.025220204144716263, -0.11817122995853424, -0.05234697461128235, -0.028921637684106827, 0.08580511808395386, 0.06200293079018593, -0.08556309342384338, -0.046005260199308395, 0.14620883762836456, -0.11532316356897354, 0.01765022613108158, -0.06353657692670822, -0.065835140645504, -0.06964749097824097, -0.0306700486689806, 0.03625200316309929, 0.04285764321684837, 0.06732126325368881, -0.06293471157550812, -0.017329838126897812, 0.1297825574874878, -0.00867757759988308, -0.18613693118095398, -0.008056540042161942, 0.07100265473127365, 0.07186407595872879, 0.16181297600269318, 0.02870132587850094, 0.1278453767299652, 0.08913198858499527, 0.024633284658193588, -0.1375143676996231, 0.08009076863527298, 0.030775876715779305, 0.007595620583742857, 0.13320626318454742, -0.009791107848286629, 0.04471273720264435, -0.08381710946559906, 0.008984101936221123, -0.1254633218050003, 0.04329799488186836, 0.01089733000844717, -0.0705486610531807, -0.11765368282794952, 0.05428370460867882, -0.0658351331949234, 0.11847252398729324, 0.09780219942331314, -0.09006263315677643, -0.019781824201345444, -0.04467284306883812, 0.06317944079637527, -0.01744958385825157, -0.05434292554855347, 0.0183025561273098, -0.0948929637670517, 0.01345102209597826, -0.00325187761336565, -0.02115483023226261, -0.16954238712787628, -0.026227664202451706, -0.00635753246024251, -0.009443068876862526, -0.043554093688726425, 0.08260030299425125, 0.08975943177938461, 0.03338024020195007, -0.04020272195339203, 0.0029677057173103094, -0.04357466101646423, 0.060165438801050186, -0.06806882470846176, -0.12312763184309006 ]
null
null
transformers
Model based on [ruGPT-3](https://huggingface.co/sberbank-ai/rugpt3small_based_on_gpt2) for generating songs. Tuned on lyrics collected from [genius](https://genius.com/). Examples of used artists: * [Oxxxymiron](https://genius.com/artists/Oxxxymiron) * [ะœะพั€ะณะตะฝัˆั‚ะตั€ะฝ](https://genius.com/artists/Morgenshtern) * [ะ›ะกะŸ](https://genius.com/artists/Lsp) * [ะ“ั€ะฐะถะดะฐะฝัะบะฐั ะพะฑะพั€ะพะฝะฐ](https://genius.com/artists/Civil-defense) * [ะšะพั€ะพะปัŒ ะธ ะจัƒั‚](https://genius.com/artists/The-king-and-the-jester) * etc
{"language": ["ru"], "tags": ["PyTorch", "Transformers"], "widget": [{"text": "\u0411\u0430\u0442\u044f \u0432\u043e\u0437\u0432\u0440\u0430\u0449\u0430\u0435\u0442\u0441\u044f \u0442\u0440\u0435\u0437\u0432\u044b\u0439, \u0432 \u0440\u0443\u043a\u0435 \u0431\u0443\u0445\u0430\u043d\u043a\u0430", "example_title": "Example 1"}, {"text": "\u041a\u0430\u043a \u0434\u0435\u043b\u0430? \u041a\u0430\u043a \u0434\u0435\u043b\u0430? \u042d\u0442\u043e \u043d\u043e\u0432\u044b\u0439 \u043a\u0430\u0434\u0438\u043b\u043b\u0430\u043a", "example_title": "Example 2"}, {"text": "4:20 \u043d\u0430 \u0447\u0430\u0441\u0430\u0445 \u0438 \u044f \u0434\u0440\u043e\u0447\u0443 \u043d\u0430 \u0442\u0432\u043e\u0451 \u0444\u043e\u0442\u043e", "example_title": "Example 3"}], "inference": {"parameters": {"temperature": 0.9, "k": 50, "p": 0.95, "length": 1500}}}
text-generation
bankholdup/rugpt3_song_writer
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "PyTorch", "Transformers", "ru", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "ru" ]
TAGS #transformers #pytorch #safetensors #gpt2 #text-generation #PyTorch #Transformers #ru #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
Model based on ruGPT-3 for generating songs. Tuned on lyrics collected from genius. Examples of used artists: * Oxxxymiron * ะœะพั€ะณะตะฝัˆั‚ะตั€ะฝ * ะ›ะกะŸ * ะ“ั€ะฐะถะดะฐะฝัะบะฐั ะพะฑะพั€ะพะฝะฐ * ะšะพั€ะพะปัŒ ะธ ะจัƒั‚ * etc
[]
[ "TAGS\n#transformers #pytorch #safetensors #gpt2 #text-generation #PyTorch #Transformers #ru #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 67 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #gpt2 #text-generation #PyTorch #Transformers #ru #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ -0.008647638373076916, -0.03651204705238342, -0.0060465168207883835, 0.018718423321843147, 0.1250101923942566, -0.006818299647420645, 0.15762774646282196, 0.08865608274936676, -0.0364232175052166, 0.02058558538556099, 0.16555942595005035, 0.1380137801170349, -0.02212666906416416, 0.09638597071170807, -0.0534849688410759, -0.2700098156929016, 0.09471429884433746, 0.02237537130713463, -0.060780156403779984, 0.14116628468036652, 0.09550432115793228, -0.0799560472369194, 0.06519033014774323, -0.016893353313207626, -0.13943833112716675, 0.009447026997804642, 0.04815905913710594, -0.15256790816783905, 0.12888750433921814, 0.028853993862867355, 0.154652938246727, 0.0512017086148262, -0.04381987079977989, -0.10484419763088226, 0.03601132333278656, 0.031432490795850754, -0.07471508532762527, 0.06961647421121597, 0.07862500846385956, -0.06677931547164917, 0.1472545713186264, 0.009964427910745144, 0.005019202828407288, 0.02185327745974064, -0.12686850130558014, -0.10793600976467133, -0.012621134519577026, 0.09583478420972824, 0.012253831140697002, 0.12228602916002274, -0.026479177176952362, 0.18886323273181915, -0.11954683810472488, 0.09290266036987305, 0.21040187776088715, -0.32151564955711365, -0.02268112264573574, 0.05382455140352249, 0.09756923466920853, 0.05346802622079849, -0.06007350981235504, 0.06120475009083748, 0.02742333523929119, 0.02148601971566677, 0.04995131120085716, -0.032621774822473526, -0.11138293147087097, 0.02001192234456539, -0.12782664597034454, -0.05123448744416237, 0.18726769089698792, -0.0847025215625763, 0.06894693523645401, -0.06603886187076569, -0.1311526596546173, -0.06541886180639267, 0.0016225315630435944, 0.011028062552213669, -0.0713578462600708, 0.029870517551898956, -0.007088592275977135, -0.02662867307662964, -0.15670084953308105, -0.006318152416497469, -0.14034032821655273, 0.23999850451946259, 0.029061928391456604, 0.044740770012140274, -0.16830562055110931, 0.11384711414575577, 0.07563851028680801, -0.14320749044418335, 0.0715407282114029, -0.11122727394104004, 0.07870915532112122, 0.02725895866751671, -0.058512259274721146, -0.061613988131284714, 0.10128598660230637, 0.1385941058397293, -0.06453298032283783, -0.014273181557655334, 0.01877102255821228, 0.10069063305854797, 0.00679891649633646, 0.08716956526041031, -0.05594881996512413, 0.0490899533033371, 0.043656378984451294, -0.053180620074272156, 0.045932650566101074, -0.08180145174264908, -0.11441491544246674, -0.034239280968904495, 0.0868048369884491, 0.08298005163669586, 0.03948049247264862, 0.09986646473407745, -0.0221202801913023, 0.02601197548210621, 0.10844948887825012, -0.060191355645656586, -0.0004163105331826955, 0.012881673872470856, 0.0419994592666626, 0.10359877347946167, 0.013861721381545067, 0.00715164327993989, -0.06971973180770874, 0.09681273996829987, -0.07179315388202667, -0.01321228127926588, -0.03895660489797592, -0.09848982840776443, 0.04167133569717407, -0.08275027573108673, 0.01499974261969328, -0.20523394644260406, -0.0859498605132103, 0.01858353801071644, -0.01724298857152462, -0.008698932826519012, -0.030542414635419846, 0.031165704131126404, -0.09361206740140915, 0.06530435383319855, -0.04764528572559357, -0.024205893278121948, -0.044703684747219086, 0.12354288250207901, -0.04916541278362274, 0.08949477225542068, -0.10133221745491028, 0.03624150529503822, -0.12615489959716797, -0.016926279291510582, -0.0997496172785759, -0.0010327381314709783, -0.05744883045554161, 0.11491072922945023, 0.020449278876185417, -0.04915481060743332, -0.06733760982751846, 0.05341419577598572, -0.014648642390966415, 0.1751621514558792, -0.06747374683618546, -0.11373066157102585, 0.23714567720890045, -0.16077402234077454, -0.13468487560749054, 0.11451269686222076, 0.020770782604813576, -0.009632333181798458, 0.06206619739532471, 0.1675150841474533, 0.05080275237560272, -0.04872952029109001, -0.027536865323781967, 0.08003782480955124, -0.08732826262712479, -0.043255195021629333, 0.030899010598659515, 0.03031252510845661, -0.06661506742238998, 0.0369449183344841, 0.031956352293491364, 0.09053529053926468, -0.06795109808444977, -0.052146658301353455, -0.07015417516231537, -0.019765105098485947, 0.13396255671977997, 0.025225641205906868, 0.10941018164157867, -0.08237349987030029, -0.10054820030927658, 0.004222761373966932, 0.0257453303784132, -0.00774601474404335, 0.03507795184850693, -0.043500762432813644, 0.15053997933864594, 0.000546977564226836, 0.042202748358249664, -0.15768419206142426, -0.11245717108249664, -0.027256950736045837, 0.0765477642416954, -0.017352348193526268, 0.06754721701145172, 0.08532411605119705, 0.011642186902463436, -0.02180851437151432, -0.04035742208361626, 0.08226411044597626, 0.007748428266495466, -0.05271270498633385, -0.10516470670700073, 0.045111726969480515, -0.07728374004364014, 0.024479888379573822, -0.0686916783452034, 0.052339665591716766, 0.10245878249406815, 0.14782612025737762, -0.003082278184592724, 0.0540597029030323, -0.03495858237147331, 0.03180127590894699, -0.08774665743112564, -0.009328538551926613, 0.07081803679466248, -0.008948576636612415, -0.06075293570756912, 0.17795906960964203, -0.1649857610464096, 0.3323783278465271, 0.23078185319900513, -0.22095663845539093, -0.022254209965467453, 0.032925136387348175, -0.03509417548775673, 0.03347361087799072, 0.0030570554081350565, -0.04560123756527901, 0.027741797268390656, -0.002239232650026679, 0.1293003261089325, -0.049236223101615906, -0.05849235877394676, 0.02268017828464508, -0.06659712642431259, -0.014946013689041138, 0.0879230871796608, 0.050758227705955505, -0.09019037336111069, 0.19460830092430115, 0.23153802752494812, -0.026165898889303207, 0.18430113792419434, -0.015345851890742779, -0.0037957841996103525, 0.060046397149562836, -0.006380774546414614, -0.019159534946084023, -0.008659500628709793, -0.14219731092453003, -0.038103945553302765, 0.06927560269832611, 0.004424684680998325, 0.06407242268323898, -0.1529628038406372, -0.057526614516973495, -0.017166558653116226, 0.016431570053100586, -0.03043980710208416, 0.11660896986722946, 0.035727426409721375, 0.13763634860515594, -0.043791282922029495, -0.046868931502103806, 0.07968171685934067, 0.02390391379594803, -0.06537353992462158, 0.19625581800937653, -0.09982460737228394, -0.34779518842697144, -0.13450872898101807, -0.04006543383002281, -0.043703701347112656, 0.034600138664245605, 0.1155204102396965, -0.14480060338974, -0.022024208679795265, -0.014903442934155464, 0.10254741460084915, -0.028302635997533798, 0.06546014547348022, -0.09526654332876205, 0.0370962917804718, -0.05308641493320465, -0.09375827759504318, -0.04286732152104378, -0.027157463133335114, -0.07626894116401672, 0.16984547674655914, -0.011363288387656212, 0.08028674125671387, 0.17384648323059082, 0.0029832087457180023, 0.0163574256002903, -0.03002408891916275, 0.15715396404266357, -0.09767607599496841, 0.021133946254849434, 0.20989039540290833, -0.021445300430059433, 0.08649960905313492, 0.12400976568460464, 0.022011036053299904, -0.03534834459424019, 0.04054815694689751, -0.046607017517089844, -0.09578371793031693, -0.21904584765434265, -0.11681253463029861, -0.08276978135108948, 0.06593237072229385, 0.019038036465644836, 0.08481762558221817, 0.11509759724140167, 0.10364020615816116, -0.030090922489762306, -0.043415285646915436, 0.009499113075435162, 0.06013680249452591, 0.17426413297653198, -0.016410337761044502, 0.17548438906669617, -0.07542615383863449, -0.15129354596138, 0.07716807723045349, -0.024196650832891464, 0.1302255094051361, 0.009984196163713932, -0.02252243272960186, 0.04396936297416687, 0.13258123397827148, 0.15897075831890106, 0.13550902903079987, 0.025867396965622902, -0.04665861651301384, 0.0004115618939977139, -0.03969207778573036, -0.013788880780339241, 0.02558400109410286, -0.019662000238895416, -0.09877898544073105, -0.05781538411974907, -0.100985586643219, 0.12884202599525452, 0.07927314937114716, 0.04807490482926369, -0.21726608276367188, -0.01663529872894287, 0.08196943998336792, -0.013536647893488407, -0.08321981132030487, 0.0799415335059166, 0.0477159209549427, -0.11385643482208252, 0.03276919201016426, -0.04685071483254433, 0.0758456140756607, -0.011882716789841652, 0.08178414404392242, -0.04142133146524429, -0.06645811349153519, 0.007841183803975582, 0.08812669664621353, -0.25362253189086914, 0.22752420604228973, -0.02989445999264717, -0.0443088635802269, -0.06953729689121246, 0.002898681443184614, 0.02929074317216873, 0.14794722199440002, 0.08639233559370041, 0.006366277113556862, -0.06591226905584335, -0.12035661190748215, -0.04340856149792671, 0.03258180990815163, 0.1121455579996109, -0.036064811050891876, 0.016341887414455414, -0.04501250758767128, 0.005027364939451218, 0.010038972832262516, -0.0013023673091083765, -0.03149312362074852, -0.14948543906211853, 0.0836428627371788, 0.04460005834698677, 0.07973881810903549, -0.05300716683268547, -0.09215235710144043, -0.1246165856719017, 0.15175411105155945, -0.03858286887407303, -0.09919067472219467, -0.11111639440059662, -0.07427042722702026, 0.08266671746969223, -0.08957620710134506, 0.06371024996042252, -0.06369074434041977, 0.048031531274318695, -0.07877316325902939, -0.16218015551567078, 0.12325392663478851, -0.10735712200403214, -0.12634924054145813, -0.014918221160769463, 0.16668830811977386, -0.055714186280965805, 0.01557606179267168, -0.016406157985329628, 0.05598283186554909, -0.13125452399253845, -0.13317890465259552, 0.006222804542630911, -0.05264890193939209, 0.011874393559992313, 0.003101508831605315, -0.038425981998443604, -0.07380400598049164, 0.0038847934920340776, -0.003405991941690445, 0.24811236560344696, 0.19845858216285706, -0.08220536261796951, 0.1364060640335083, 0.13821187615394592, -0.011681266129016876, -0.3538472652435303, -0.111261747777462, -0.11646369844675064, -0.04398343339562416, -0.023143639788031578, -0.11443658918142319, 0.07541467249393463, 0.008669314906001091, -0.060655102133750916, 0.08934452384710312, -0.2618584930896759, -0.09696945548057556, 0.16968923807144165, -0.013546023517847061, 0.3060927391052246, -0.18279093503952026, -0.05959980934858322, -0.04327830672264099, -0.11010909825563431, 0.14070095121860504, -0.1380971223115921, 0.08141566812992096, -0.02506125532090664, 0.05737973004579544, 0.03675849363207817, -0.07363510131835938, 0.10689433664083481, -0.04587986692786217, 0.015238828957080841, -0.11220026016235352, -0.05564834550023079, 0.1216956153512001, 0.007519707083702087, 0.02221044711768627, -0.0910346731543541, 0.03509392589330673, -0.0691923052072525, -0.033124592155218124, -0.09831184893846512, 0.09619525820016861, 0.0036733949091285467, -0.06540130078792572, -0.01771744154393673, -0.044878724962472916, -0.023726265877485275, -0.01112903468310833, 0.2069440484046936, -0.003921871073544025, 0.2061537355184555, 0.1074315756559372, 0.0602337121963501, -0.09018737822771072, 0.008716207928955555, -0.022888904437422752, -0.07557186484336853, 0.09295062720775604, -0.11191629618406296, 0.022217651829123497, 0.0945567712187767, -0.03331995755434036, 0.020123561844229698, 0.11629010736942291, 0.002874640515074134, -0.0021480992436408997, 0.15846282243728638, -0.26903554797172546, -0.030184557661414146, -0.055325184017419815, 0.03254011645913124, 0.04381252080202103, 0.11487522721290588, 0.1661529839038849, -0.0330989733338356, -0.03491433337330818, -0.004072023555636406, 0.006601514760404825, -0.04454866424202919, 0.07762482017278671, 0.058831389993429184, 0.04772583022713661, -0.12032811343669891, 0.028356999158859253, 0.01825951784849167, -0.12638503313064575, -0.0023075053468346596, 0.13826751708984375, -0.1302686482667923, -0.1447412222623825, -0.015337536111474037, 0.011016505770385265, -0.12374426424503326, -0.04359166696667671, -0.055723149329423904, -0.17048059403896332, 0.04753165692090988, 0.16114185750484467, 0.08326844871044159, 0.0765388160943985, -0.022258244454860687, -0.006738380063325167, -0.04334175959229469, 0.029774954542517662, 0.02339782565832138, 0.04726779833436012, -0.13305120170116425, 0.080879807472229, 0.015662578865885735, 0.0930185467004776, -0.10317522287368774, 0.000054878564696991816, -0.17142529785633087, 0.01465549599379301, -0.08954669535160065, -0.09437432885169983, -0.06543149799108505, -0.06362663209438324, -0.022440174594521523, -0.07021176815032959, -0.03827290236949921, -0.03703268617391586, -0.09692531824111938, 0.023891301825642586, -0.030889863148331642, 0.019678713753819466, -0.1312156617641449, -0.008655255660414696, 0.0956750139594078, -0.057546623051166534, 0.1377263367176056, 0.12510743737220764, -0.07214770466089249, 0.1077423244714737, -0.12905888259410858, -0.05858822911977768, 0.10348512977361679, -0.008805516175925732, 0.015115086920559406, 0.04461446404457092, 0.029256334528326988, 0.043299488723278046, 0.03611794486641884, 0.07661139219999313, 0.024087805300951004, -0.10502835363149643, 0.10958848893642426, -0.012377827428281307, -0.12057610601186752, -0.03486701101064682, -0.054854825139045715, 0.07051923871040344, -0.015579869970679283, 0.14652910828590393, -0.063470758497715, 0.08217188715934753, -0.08003320544958115, 0.04264044389128685, -0.011064857244491577, -0.19657254219055176, -0.07333298027515411, -0.037550654262304306, 0.028933178633451462, -0.012896346859633923, 0.24410897493362427, 0.04449862614274025, -0.03132988139986992, 0.04710838943719864, 0.10066860169172287, 0.015535727143287659, 0.03952853009104729, 0.1732417196035385, 0.0945267304778099, -0.04881427437067032, -0.11950697004795074, 0.01716795191168785, 0.0252943467348814, -0.1009189710021019, 0.13053613901138306, 0.054942503571510315, 0.011201920919120312, 0.07081196457147598, 0.011840474791824818, 0.02243036963045597, -0.1296689659357071, -0.1698148399591446, -0.06624817848205566, 0.02944430522620678, -0.04807887598872185, 0.07985231280326843, 0.20355196297168732, 0.019096670672297478, 0.005024853628128767, -0.03622918948531151, -0.026738660410046577, -0.15624304115772247, -0.07433866709470749, -0.07070262730121613, -0.13972942531108856, 0.004709141328930855, -0.08188934624195099, 0.0025202245451509953, 0.08614427596330643, 0.04928900673985481, -0.03647598624229431, 0.15754780173301697, 0.10829436033964157, -0.03965985029935837, 0.05954938381910324, -0.005976430606096983, 0.023746294900774956, -0.007487332448363304, -0.027297431603074074, -0.10930893570184708, 0.011669928207993507, -0.030308660119771957, 0.0439448282122612, -0.10565672069787979, 0.02754775620996952, -0.12858745455741882, -0.10082519799470901, -0.04262761026620865, 0.057627249509096146, -0.01142420805990696, 0.1199890747666359, -0.008290976285934448, -0.012187622487545013, 0.021825477480888367, 0.23466604948043823, -0.07477783411741257, -0.13488346338272095, -0.028599847108125687, 0.1702715903520584, 0.012889954261481762, 0.09558697789907455, -0.028422171249985695, -0.021595396101474762, -0.11133092641830444, 0.2931497097015381, 0.33339744806289673, -0.033057790249586105, 0.07214156538248062, 0.015878265723586082, 0.0282759927213192, 0.05812513828277588, 0.10319389402866364, 0.11739486455917358, 0.2945913076400757, -0.09746193885803223, -0.01413646899163723, -0.04186810553073883, -0.003395375330001116, -0.10882306098937988, 0.08146512508392334, 0.030783645808696747, -0.05547504499554634, -0.06831468641757965, 0.07635264098644257, -0.19533491134643555, 0.07808779180049896, -0.07479675859212875, -0.18577119708061218, -0.041590724140405655, 0.029929963871836662, 0.18229275941848755, -0.009088757447898388, 0.0977732390165329, -0.005523467902094126, -0.08876261860132217, 0.02748659811913967, 0.031232524663209915, -0.15125460922718048, 0.025420907884836197, 0.08090579509735107, -0.06508573889732361, 0.0524565652012825, -0.03819404914975166, 0.04821470007300377, 0.10228719562292099, 0.020359763875603676, -0.07106728106737137, 0.05405664071440697, 0.02468743734061718, -0.0922643169760704, -0.018851717934012413, 0.01569155417382717, 0.021632006391882896, -0.10404114425182343, 0.06474205106496811, -0.14968213438987732, 0.04734135791659355, -0.04412945732474327, 0.004465137608349323, -0.02386661060154438, 0.030613824725151062, -0.03697443753480911, 0.059354063123464584, 0.07391142100095749, -0.026062138378620148, -0.016863640397787094, -0.05175596475601196, -0.01332868356257677, -0.007894502952694893, -0.06264658272266388, -0.06277815997600555, -0.15686555206775665, -0.05939681828022003, 0.12709462642669678, 0.024153413251042366, -0.2096116989850998, -0.011044737882912159, -0.08583284169435501, 0.05414392054080963, -0.17749841511249542, 0.04353443160653114, 0.10069365799427032, -0.0020100348629057407, 0.013629460707306862, -0.009095733053982258, 0.012970766052603722, 0.08236199617385864, -0.11897346377372742, -0.07148967683315277 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.7523 - Matthews Correlation: 0.5259 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.533 | 1.0 | 535 | 0.5318 | 0.3887 | | 0.3562 | 2.0 | 1070 | 0.5145 | 0.5100 | | 0.2429 | 3.0 | 1605 | 0.6558 | 0.4888 | | 0.1831 | 4.0 | 2140 | 0.7523 | 0.5259 | | 0.1352 | 5.0 | 2675 | 0.8406 | 0.5182 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5258663312307151, "name": "Matthews Correlation"}]}]}]}
text-classification
banri/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.7523 * Matthews Correlation: 0.5259 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ -0.09953347593545914, 0.09874087572097778, -0.002511604456230998, 0.12091247737407684, 0.16499070823192596, 0.034331291913986206, 0.1324433982372284, 0.12642896175384521, -0.08521388471126556, 0.021414093673229218, 0.12043958902359009, 0.1608530431985855, 0.02341119386255741, 0.11020602285861969, -0.04826747626066208, -0.26533666253089905, -0.01393829844892025, 0.04971835017204285, -0.05632239580154419, 0.1354067325592041, 0.09114500880241394, -0.12023554742336273, 0.09020212292671204, 0.0099337138235569, -0.19124649465084076, 0.0011048978194594383, -0.0010920170461758971, -0.05103665590286255, 0.14910998940467834, 0.02660513110458851, 0.12270399183034897, 0.005109584890305996, 0.08301815390586853, -0.20412342250347137, 0.010900583118200302, 0.046728942543268204, 0.0031716071534901857, 0.09235342592000961, 0.045432198792696, 0.0016819247975945473, 0.11801120638847351, -0.07893169671297073, 0.05534156784415245, 0.024939799681305885, -0.11975694447755814, -0.2128710299730301, -0.0778748020529747, 0.036783378571271896, 0.07661557197570801, 0.10431034117937088, -0.008242898620665073, 0.1184171512722969, -0.08168778568506241, 0.09322264790534973, 0.2265443205833435, -0.28452247381210327, -0.06619590520858765, 0.04379666596651077, 0.015058509074151516, 0.04781145974993706, -0.1015576496720314, -0.035203490406274796, 0.04749462753534317, 0.0519600510597229, 0.12782640755176544, -0.027968812733888626, -0.11474987864494324, 0.006442997604608536, -0.14108677208423615, -0.03252919390797615, 0.16779018938541412, 0.04112625494599342, -0.027329619973897934, -0.05655708163976669, -0.0565880686044693, -0.15023289620876312, -0.035420678555965424, -0.016137896105647087, 0.049116719514131546, -0.02263779006898403, -0.04222801327705383, -0.0091459471732378, -0.1084260642528534, -0.0649506226181984, -0.07226604968309402, 0.11088686436414719, 0.038162484765052795, 0.0058541507460176945, -0.030373068526387215, 0.1126675009727478, -0.005879268515855074, -0.12169726938009262, 0.02592085860669613, 0.022252557799220085, 0.012412738986313343, -0.039657268673181534, -0.052843380719423294, -0.06314884126186371, 0.01163672935217619, 0.13044480979442596, -0.05523933470249176, 0.04299374297261238, 0.051151104271411896, 0.0497078113257885, -0.09264211356639862, 0.18815849721431732, -0.03412971273064613, -0.0246804878115654, -0.0013784491457045078, 0.05078420042991638, 0.015825295820832253, -0.012068700976669788, -0.12227529287338257, 0.007106361445039511, 0.08839566260576248, 0.009266041219234467, -0.0627312883734703, 0.07418110966682434, -0.060031548142433167, -0.024304281920194626, -0.000583169050514698, -0.09094366431236267, 0.021768907085061073, 0.0020300941541790962, -0.07044500857591629, -0.021747708320617676, 0.03545186296105385, 0.016193050891160965, -0.01973368413746357, 0.10824801027774811, -0.08700212091207504, 0.02766178362071514, -0.09214701503515244, -0.11091382056474686, 0.016382884234189987, -0.10616400837898254, 0.02209707722067833, -0.09194151312112808, -0.17804844677448273, -0.016779260709881783, 0.06123820319771767, -0.0249736700206995, -0.057835739105939865, -0.059264104813337326, -0.0670977234840393, 0.010518141090869904, -0.007938647642731667, 0.11550603061914444, -0.06454900652170181, 0.09306663274765015, 0.02486114203929901, 0.06328190118074417, -0.03969729691743851, 0.060109253972768784, -0.10152862221002579, 0.014346135780215263, -0.15077118575572968, 0.04044334962964058, -0.051775697618722916, 0.06826219707727432, -0.0828319564461708, -0.10620411485433578, 0.004799502901732922, -0.002012147568166256, 0.06199968606233597, 0.09603439271450043, -0.18445594608783722, -0.08309091627597809, 0.1646650731563568, -0.07347718626260757, -0.12144056707620621, 0.12104693800210953, -0.05849684029817581, 0.06013638526201248, 0.05822913348674774, 0.18045574426651, 0.08632402122020721, -0.07972324639558792, 0.0034902701154351234, 0.023290645331144333, 0.0507633276283741, -0.0611066073179245, 0.06806862354278564, 0.0016729767667129636, 0.021276770159602165, 0.035037096589803696, -0.028808169066905975, 0.06409851461648941, -0.0885707437992096, -0.0994555801153183, -0.03766985982656479, -0.08175698667764664, 0.047097474336624146, 0.07800299674272537, 0.0673503428697586, -0.09371918439865112, -0.07792548090219498, 0.05101942643523216, 0.08089784532785416, -0.05952814593911171, 0.02505442686378956, -0.05018920451402664, 0.07249259203672409, -0.023010022938251495, -0.021495360881090164, -0.17935718595981598, -0.03304700180888176, 0.007671220228075981, 0.0016373213147744536, 0.018306195735931396, 0.03491630405187607, 0.06346556544303894, 0.06125207617878914, -0.04979454353451729, -0.019516687840223312, -0.034941304475069046, 0.00024957084679044783, -0.12484022974967957, -0.1978163868188858, -0.029437387362122536, -0.022979283705353737, 0.15926769375801086, -0.20804807543754578, 0.05219445005059242, -0.012930957600474358, 0.06893651932477951, 0.01259586587548256, -0.006784488912671804, -0.037707116454839706, 0.07587210088968277, -0.04283028841018677, -0.05062086880207062, 0.08230898529291153, 0.014233949594199657, -0.09096012264490128, -0.04920530319213867, -0.09941975027322769, 0.1581258773803711, 0.12762826681137085, -0.11041786521673203, -0.07833212614059448, -0.023729141801595688, -0.06755886971950531, -0.03483648970723152, -0.046310633420944214, 0.025552386417984962, 0.18936489522457123, -0.004262363072484732, 0.14987918734550476, -0.0682009607553482, -0.04397238418459892, 0.017608368769288063, -0.037649307399988174, 0.015210999175906181, 0.13444697856903076, 0.1345853954553604, -0.05895493924617767, 0.15505675971508026, 0.14793606102466583, -0.08454471826553345, 0.15085500478744507, -0.04173930361866951, -0.06567565351724625, -0.017701338976621628, -0.02857152186334133, -0.010895025916397572, 0.09964711219072342, -0.15596097707748413, -0.0019119561184197664, 0.03000222146511078, 0.015877656638622284, 0.02510448545217514, -0.22794215381145477, -0.04155518114566803, 0.0389273501932621, -0.0450444333255291, -0.006036035716533661, -0.007226346060633659, 0.005228179506957531, 0.10049860179424286, -0.00018409325275570154, -0.08764992654323578, 0.03679626062512398, 0.0018983642803505063, -0.08439407497644424, 0.21584919095039368, -0.08084686845541, -0.17281745374202728, -0.13095782697200775, -0.07243042439222336, -0.04566062614321709, -0.0010706400498747826, 0.0691693052649498, -0.0967247262597084, -0.026842236518859863, -0.07250673323869705, 0.02798851579427719, 0.007345893885940313, 0.022695617750287056, 0.004121406003832817, 0.007244843058288097, 0.06366469711065292, -0.11213372647762299, -0.015016691759228706, -0.0575074627995491, -0.044975247234106064, 0.04634639620780945, 0.027689194306731224, 0.11020880937576294, 0.15379390120506287, -0.012518510222434998, 0.012555653229355812, -0.03148170933127403, 0.2373463362455368, -0.05985097959637642, -0.020610256120562553, 0.14333312213420868, -0.009508912451565266, 0.052515190094709396, 0.11528576165437698, 0.0747729167342186, -0.07828541100025177, 0.0047271978110075, 0.03834423050284386, -0.034632548689842224, -0.23223504424095154, -0.05309094488620758, -0.05468036234378815, 0.012766015715897083, 0.09088542312383652, 0.023131858557462692, 0.029285304248332977, 0.06966236978769302, 0.04191255196928978, 0.07321835309267044, -0.03509543091058731, 0.05163390561938286, 0.12911584973335266, 0.02997245267033577, 0.12461114674806595, -0.0458318330347538, -0.06507453322410583, 0.04078632593154907, -0.008546932600438595, 0.2244189828634262, 0.01144444476813078, 0.13197962939739227, 0.0672767236828804, 0.16255813837051392, -0.009635847993195057, 0.07562866061925888, -0.009486976079642773, -0.037601273506879807, -0.01691015623509884, -0.03981604427099228, -0.038888368755578995, 0.024911319836974144, -0.06262046843767166, 0.06230686232447624, -0.1238240972161293, 0.014674068428575993, 0.05898761749267578, 0.2494848221540451, 0.03398442640900612, -0.317575603723526, -0.09669401496648788, 0.001925171003676951, -0.03102806955575943, -0.01730147935450077, 0.027163192629814148, 0.09546197950839996, -0.09838628023862839, 0.028661495074629784, -0.07474260032176971, 0.09572183340787888, -0.056389112025499344, 0.0516512431204319, 0.08281200379133224, 0.08809110522270203, 0.011251848191022873, 0.09231287240982056, -0.2882535457611084, 0.2763582170009613, -0.0006312874611467123, 0.0553419403731823, -0.07551172375679016, 0.008627848699688911, 0.04104002192616463, 0.06337206810712814, 0.08036306500434875, -0.012492210604250431, -0.019014839082956314, -0.1881796270608902, -0.06654578447341919, 0.026966385543346405, 0.06983157992362976, -0.042700767517089844, 0.08395274728536606, -0.032174501568078995, 0.00960498582571745, 0.07390110939741135, 0.0026269557420164347, -0.05285651981830597, -0.10823807865381241, -0.00419450830668211, 0.019751714542508125, -0.0604015551507473, -0.06156456097960472, -0.12159401923418045, -0.12967482209205627, 0.1563071310520172, -0.03687328100204468, -0.03934948891401291, -0.10622747987508774, 0.08209172636270523, 0.05936790630221367, -0.08948037028312683, 0.042552195489406586, 0.0031793885864317417, 0.07605233788490295, 0.020437801256775856, -0.07035695016384125, 0.10311683267354965, -0.07421185821294785, -0.15605419874191284, -0.06501059234142303, 0.10722552239894867, 0.03453879803419113, 0.06701719015836716, -0.014527936466038227, 0.0036785739939659834, -0.04561163857579231, -0.08830267190933228, 0.021457381546497345, 0.0034411305096000433, 0.07714685052633286, 0.01873295195400715, -0.07657564431428909, 0.012000871822237968, -0.0651417076587677, -0.033748023211956024, 0.2058180272579193, 0.22602058947086334, -0.0996016338467598, 0.02484969235956669, 0.025279255583882332, -0.07408372312784195, -0.19799210131168365, 0.03336464613676071, 0.054752811789512634, 0.008055822923779488, 0.0453217551112175, -0.18585991859436035, 0.12986215949058533, 0.10611351579427719, -0.011682478711009026, 0.10396619141101837, -0.3257942497730255, -0.1207713782787323, 0.13650722801685333, 0.13745655119419098, 0.09705239534378052, -0.13265833258628845, -0.02269246056675911, -0.017406022176146507, -0.13992424309253693, 0.11496812850236893, -0.090373694896698, 0.1212964728474617, -0.03681858256459236, 0.07597478479146957, 0.0026575999800115824, -0.05805017799139023, 0.12036504596471786, 0.02447415143251419, 0.09468253701925278, -0.05885415896773338, -0.0346006378531456, 0.03072948008775711, -0.04217635840177536, 0.03359675034880638, -0.09899406880140305, 0.028582602739334106, -0.10210000723600388, -0.02577287144958973, -0.06836604326963425, 0.04421823471784592, -0.046214211732149124, -0.06974469125270844, -0.03741461783647537, 0.026357466354966164, 0.048025988042354584, -0.007492511998862028, 0.12165822833776474, 0.023274313658475876, 0.14792604744434357, 0.09826693683862686, 0.07475673407316208, -0.06614682078361511, -0.08220246434211731, -0.02739529311656952, -0.010869084857404232, 0.04992781952023506, -0.1360960453748703, 0.020313922315835953, 0.15213249623775482, 0.019550077617168427, 0.15458178520202637, 0.08367957174777985, -0.021696390584111214, -0.0013909009285271168, 0.05894676223397255, -0.16482074558734894, -0.095497265458107, -0.01692536473274231, -0.0687151551246643, -0.12071292847394943, 0.04411275312304497, 0.09410307556390762, -0.06731542944908142, -0.0070029981434345245, -0.004590475466102362, 0.014873005449771881, -0.05009814724326134, 0.1844250112771988, 0.06193801015615463, 0.04888257756829262, -0.09574009478092194, 0.07230930775403976, 0.04460417479276657, -0.07260123640298843, 0.003555342322215438, 0.0730445459485054, -0.08470243215560913, -0.054818421602249146, 0.06414514034986496, 0.19110994040966034, -0.04363720491528511, -0.0468888059258461, -0.14587534964084625, -0.12429648637771606, 0.07791906595230103, 0.14010050892829895, 0.11792059242725372, 0.011102552525699139, -0.06639044731855392, 0.003767638234421611, -0.10668066143989563, 0.10225244611501694, 0.04751317575573921, 0.061346109956502914, -0.14212389290332794, 0.14342623949050903, 0.019682904705405235, 0.0481918640434742, -0.01839926466345787, 0.0225361417979002, -0.1020636260509491, 0.008319245651364326, -0.09447463601827621, -0.019602062180638313, -0.0297740139067173, 0.011695586144924164, -0.005361055489629507, -0.04652413725852966, -0.05381449684500694, 0.010594788007438183, -0.10772327333688736, -0.023150067776441574, 0.029049210250377655, 0.07299237698316574, -0.1094348207116127, -0.03556108847260475, 0.030979042872786522, -0.06228852644562721, 0.07504276186227798, 0.042609747499227524, 0.014938700012862682, 0.05030939728021622, -0.1388765275478363, 0.020099302753806114, 0.07358679920434952, 0.028721114620566368, 0.060426343232393265, -0.10016195476055145, -0.00950651615858078, -0.009530103765428066, 0.039311885833740234, 0.02166745252907276, 0.07522766292095184, -0.1405515968799591, 0.004839055240154266, -0.023800915107131004, -0.08430628478527069, -0.06710097193717957, 0.027594968676567078, 0.08822209388017654, 0.018141360953450203, 0.19948069751262665, -0.07641546428203583, 0.05125359818339348, -0.21803437173366547, 0.006752499379217625, -0.0058495057746768, -0.10863523185253143, -0.09944412857294083, -0.07315931469202042, 0.05366877093911171, -0.06116030365228653, 0.15022185444831848, 0.04569690302014351, 0.01782877743244171, 0.024864260107278824, -0.01193123310804367, 0.01258366834372282, 0.01141111645847559, 0.18875959515571594, 0.030172210186719894, -0.03526976332068443, 0.05705257132649422, 0.044683922082185745, 0.10286445170640945, 0.11252806335687637, 0.20122970640659332, 0.14473356306552887, -0.009614331647753716, 0.09224250167608261, 0.043594636023044586, -0.0571451298892498, -0.15734533965587616, 0.0523236021399498, -0.03337396681308746, 0.10859769582748413, -0.020434541627764702, 0.22116538882255554, 0.06473782658576965, -0.1704503446817398, 0.051612015813589096, -0.05156480893492699, -0.08757349848747253, -0.11537886410951614, -0.049788981676101685, -0.07638025283813477, -0.13064563274383545, -0.004418745171278715, -0.11596781760454178, -0.0030279008205980062, 0.1252775639295578, 0.004495105240494013, -0.026288362219929695, 0.1577172875404358, 0.013742298819124699, 0.022418297827243805, 0.05891863629221916, 0.008379055187106133, -0.03827941045165062, -0.1397867351770401, -0.058738671243190765, -0.013102419674396515, -0.008189517073333263, 0.0301776472479105, -0.061405882239341736, -0.04409507289528847, 0.03033839538693428, -0.02161327563226223, -0.0961952656507492, 0.0058813439682126045, 0.0114591708406806, 0.05236304923892021, 0.04504939168691635, 0.009300000965595245, 0.017563441768288612, -0.0029333671554923058, 0.19917263090610504, -0.07183873653411865, -0.0666523203253746, -0.10286440700292587, 0.2338429093360901, 0.034976303577423096, -0.017381610348820686, 0.03417123854160309, -0.06636831164360046, 0.0040776897221803665, 0.24897538125514984, 0.21798938512802124, -0.08088204264640808, -0.005363496020436287, 0.017881739884614944, -0.007192263379693031, -0.020721865817904472, 0.09718649834394455, 0.14278611540794373, 0.04755014181137085, -0.09300366789102554, -0.04347268119454384, -0.05805261433124542, -0.01779104396700859, -0.03378866985440254, 0.0696251168847084, 0.052194997668266296, 0.010079930536448956, -0.03599933162331581, 0.05708372965455055, -0.06848310679197311, -0.08963042497634888, 0.05697747692465782, -0.21786916255950928, -0.16666465997695923, -0.016561424359679222, 0.10220707952976227, 0.0009916772833094, 0.062436070293188095, -0.02936011739075184, -0.004481164738535881, 0.0916908010840416, -0.01840386539697647, -0.09784910827875137, -0.0714891329407692, 0.08569670468568802, -0.11273930966854095, 0.21685214340686798, -0.04774317145347595, 0.05394842475652695, 0.12543116509914398, 0.0677146464586258, -0.06339026987552643, 0.06550512462854385, 0.04276767000555992, -0.04071614146232605, 0.02216651663184166, 0.06880888342857361, -0.03338196873664856, 0.06525374948978424, 0.04850692301988602, -0.13922524452209473, 0.023588009178638458, -0.04621881991624832, -0.06968139857053757, -0.04311145469546318, -0.022173874080181122, -0.05998322367668152, 0.1285596489906311, 0.21853448450565338, -0.02487208880484104, -0.009410974569618702, -0.07095746695995331, 0.00939994491636753, 0.055758699774742126, 0.023770613595843315, -0.0571277029812336, -0.21143896877765656, 0.01634996384382248, 0.044792741537094116, -0.017037682235240936, -0.25224122405052185, -0.10219690948724747, 0.0042416369542479515, -0.07179052382707596, -0.09612056612968445, 0.07293257117271423, 0.08916149288415909, 0.05552821233868599, -0.056204959750175476, -0.04866582155227661, -0.07454274594783783, 0.14954450726509094, -0.1451619565486908, -0.09071090817451477 ]
null
null
transformers
# Multi-dialect-Arabic-BERT This is a repository of Multi-dialect Arabic BERT model. By [Mawdoo3-AI](https://ai.mawdoo3.com/). <p align="center"> <br> <img src="https://raw.githubusercontent.com/mawdoo3/Multi-dialect-Arabic-BERT/master/multidialct_arabic_bert.png" alt="Background reference: http://www.qfi.org/wp-content/uploads/2018/02/Qfi_Infographic_Mother-Language_Final.pdf" width="500"/> <br> <p> ### About our Multi-dialect-Arabic-BERT model Instead of training the Multi-dialect Arabic BERT model from scratch, we initialized the weights of the model using [Arabic-BERT](https://github.com/alisafaya/Arabic-BERT) and trained it on 10M arabic tweets from the unlabled data of [The Nuanced Arabic Dialect Identification (NADI) shared task](https://sites.google.com/view/nadi-shared-task). ### To cite this work ``` @misc{talafha2020multidialect, title={Multi-Dialect Arabic BERT for Country-Level Dialect Identification}, author={Bashar Talafha and Mohammad Ali and Muhy Eddin Za'ter and Haitham Seelawi and Ibraheem Tuffaha and Mostafa Samir and Wael Farhan and Hussein T. Al-Natsheh}, year={2020}, eprint={2007.05612}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Usage The model weights can be loaded using `transformers` library by HuggingFace. ```python from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("bashar-talafha/multi-dialect-bert-base-arabic") model = AutoModel.from_pretrained("bashar-talafha/multi-dialect-bert-base-arabic") ``` Example using `pipeline`: ```python from transformers import pipeline fill_mask = pipeline( "fill-mask", model="bashar-talafha/multi-dialect-bert-base-arabic ", tokenizer="bashar-talafha/multi-dialect-bert-base-arabic " ) fill_mask(" ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ [MASK] ") ``` ``` [{'sequence': '[CLS] ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ ุงู„ูƒูˆูŠุช [SEP]', 'score': 0.08296813815832138, 'token': 3226}, {'sequence': '[CLS] ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ ุฏุจูŠ [SEP]', 'score': 0.05123933032155037, 'token': 4747}, {'sequence': '[CLS] ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ ู…ุณู‚ุท [SEP]', 'score': 0.046838656067848206, 'token': 13205}, {'sequence': '[CLS] ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ ุงู„ู‚ุงู‡ุฑุฉ [SEP]', 'score': 0.03234650194644928, 'token': 4003}, {'sequence': '[CLS] ุณุงูุฑ ุงู„ุฑุญุงู„ุฉ ู…ู† ู…ุทุงุฑ ุงู„ุฑูŠุงุถ [SEP]', 'score': 0.02606341242790222, 'token': 2200}] ``` ### Repository Please check the [original repository](https://github.com/mawdoo3/Multi-dialect-Arabic-BERT) for more information.
{"language": "ar", "datasets": ["nadi"], "thumbnail": "https://raw.githubusercontent.com/mawdoo3/Multi-dialect-Arabic-BERT/master/multidialct_arabic_bert.png"}
fill-mask
bashar-talafha/multi-dialect-bert-base-arabic
[ "transformers", "pytorch", "jax", "bert", "fill-mask", "ar", "dataset:nadi", "arxiv:2007.05612", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:05+00:00
[ "2007.05612" ]
[ "ar" ]
TAGS #transformers #pytorch #jax #bert #fill-mask #ar #dataset-nadi #arxiv-2007.05612 #autotrain_compatible #endpoints_compatible #has_space #region-us
# Multi-dialect-Arabic-BERT This is a repository of Multi-dialect Arabic BERT model. By Mawdoo3-AI. <p align="center"> <br> <img src="URL alt="Background reference: URL width="500"/> <br> <p> ### About our Multi-dialect-Arabic-BERT model Instead of training the Multi-dialect Arabic BERT model from scratch, we initialized the weights of the model using Arabic-BERT and trained it on 10M arabic tweets from the unlabled data of The Nuanced Arabic Dialect Identification (NADI) shared task. ### To cite this work ### Usage The model weights can be loaded using 'transformers' library by HuggingFace. Example using 'pipeline': ### Repository Please check the original repository for more information.
[ "# Multi-dialect-Arabic-BERT\nThis is a repository of Multi-dialect Arabic BERT model.\n\nBy Mawdoo3-AI. \n\n<p align=\"center\">\n <br>\n <img src=\"URL alt=\"Background reference: URL width=\"500\"/>\n <br>\n<p>", "### About our Multi-dialect-Arabic-BERT model\nInstead of training the Multi-dialect Arabic BERT model from scratch, we initialized the weights of the model using Arabic-BERT and trained it on 10M arabic tweets from the unlabled data of The Nuanced Arabic Dialect Identification (NADI) shared task.", "### To cite this work", "### Usage\nThe model weights can be loaded using 'transformers' library by HuggingFace.\n\n\n\nExample using 'pipeline':", "### Repository\nPlease check the original repository for more information." ]
[ "TAGS\n#transformers #pytorch #jax #bert #fill-mask #ar #dataset-nadi #arxiv-2007.05612 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Multi-dialect-Arabic-BERT\nThis is a repository of Multi-dialect Arabic BERT model.\n\nBy Mawdoo3-AI. \n\n<p align=\"center\">\n <br>\n <img src=\"URL alt=\"Background reference: URL width=\"500\"/>\n <br>\n<p>", "### About our Multi-dialect-Arabic-BERT model\nInstead of training the Multi-dialect Arabic BERT model from scratch, we initialized the weights of the model using Arabic-BERT and trained it on 10M arabic tweets from the unlabled data of The Nuanced Arabic Dialect Identification (NADI) shared task.", "### To cite this work", "### Usage\nThe model weights can be loaded using 'transformers' library by HuggingFace.\n\n\n\nExample using 'pipeline':", "### Repository\nPlease check the original repository for more information." ]
[ 59, 73, 81, 6, 34, 16 ]
[ "passage: TAGS\n#transformers #pytorch #jax #bert #fill-mask #ar #dataset-nadi #arxiv-2007.05612 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Multi-dialect-Arabic-BERT\nThis is a repository of Multi-dialect Arabic BERT model.\n\nBy Mawdoo3-AI. \n\n<p align=\"center\">\n <br>\n <img src=\"URL alt=\"Background reference: URL width=\"500\"/>\n <br>\n<p>### About our Multi-dialect-Arabic-BERT model\nInstead of training the Multi-dialect Arabic BERT model from scratch, we initialized the weights of the model using Arabic-BERT and trained it on 10M arabic tweets from the unlabled data of The Nuanced Arabic Dialect Identification (NADI) shared task.### To cite this work### Usage\nThe model weights can be loaded using 'transformers' library by HuggingFace.\n\n\n\nExample using 'pipeline':### Repository\nPlease check the original repository for more information." ]
[ -0.04786442220211029, 0.14846716821193695, -0.0008238328737206757, 0.05153226852416992, 0.13231772184371948, 0.03876212239265442, 0.2706740200519562, 0.07417917996644974, 0.07446876913309097, -0.025244159623980522, 0.061297301203012466, 0.019946692511439323, 0.09599395841360092, 0.10672583431005478, 0.06938017159700394, -0.22578655183315277, -0.0013659228570759296, -0.002580532804131508, 0.01097399927675724, 0.08824767917394638, 0.08895900845527649, -0.0509970486164093, 0.12230384349822998, 0.005942855030298233, -0.12146826833486557, 0.0804905965924263, -0.05031521990895271, -0.05389367416501045, 0.09824386984109879, 0.08877614885568619, 0.1339929848909378, 0.016357295215129852, 0.036754004657268524, -0.15321221947669983, 0.03553179278969765, 0.02902158908545971, -0.028130145743489265, 0.0606159009039402, 0.03632557392120361, -0.08582005649805069, 0.13749724626541138, -0.032465964555740356, 0.02736922912299633, 0.04242200031876564, -0.09407447278499603, -0.08261138200759888, -0.02819940075278282, 0.09111998230218887, 0.04382957145571709, 0.03370998427271843, -0.01986292377114296, 0.12074439972639084, -0.03834621608257294, 0.06671881675720215, 0.1639835685491562, -0.20432297885417938, -0.046016424894332886, 0.15664240717887878, 0.026317929849028587, 0.07685156166553497, -0.07843347638845444, 0.04260751232504845, 0.027588794007897377, 0.0017997754039242864, 0.053514450788497925, -0.05630910396575928, 0.005631193984299898, -0.01867326721549034, -0.08740534633398056, -0.03748924657702446, 0.15529072284698486, 0.037035021930933, -0.04079277068376541, -0.11490052193403244, -0.05785125121474266, 0.12602756917476654, -0.040914665907621384, 0.006186922080814838, 0.01149909570813179, 0.0013068306725472212, 0.02011234685778618, -0.12245739996433258, -0.052816152572631836, -0.05606581270694733, -0.08149320632219315, 0.1127198189496994, 0.017308635637164116, 0.029906252399086952, 0.04130786657333374, 0.07149910926818848, -0.14145326614379883, -0.10638470202684402, 0.05157698318362236, -0.06803783029317856, -0.02193032205104828, 0.020628266036510468, 0.021261049434542656, -0.07467678189277649, -0.00325607368722558, 0.030014708638191223, -0.003614310408011079, 0.048607148230075836, -0.09004298597574234, 0.034917548298835754, 0.003979464061558247, 0.13877300918102264, -0.0843537449836731, -0.07937318086624146, -0.013791327364742756, 0.08059248328208923, -0.0495307594537735, 0.0051770927384495735, -0.09174259752035141, -0.04375087842345238, -0.046402160078287125, 0.030220676213502884, -0.034968484193086624, 0.08251935243606567, -0.05931883305311203, -0.0707859992980957, 0.20382985472679138, -0.1343238651752472, 0.014339336194097996, 0.024485934525728226, -0.05630676448345184, 0.0638899952173233, 0.12623029947280884, 0.03569602221250534, -0.07200589030981064, -0.07776403427124023, -0.07723679393529892, 0.04994852840900421, -0.0749870166182518, -0.15683822333812714, 0.013303522020578384, -0.13926169276237488, -0.025413522496819496, -0.13191044330596924, -0.14873677492141724, -0.000622159568592906, 0.06159166991710663, 0.0038057134952396154, -0.01587553136050701, -0.07956922799348831, 0.04662042483687401, -0.007285727187991142, -0.0006107710069045424, 0.013421710580587387, -0.02084496058523655, 0.061263930052518845, -0.0352509468793869, 0.06812215596437454, -0.0052153486758470535, 0.03756840527057648, -0.0585705004632473, 0.00710039958357811, -0.19444188475608826, 0.19042228162288666, -0.038783155381679535, 0.00988733023405075, -0.18622969090938568, 0.03147485852241516, -0.06312315911054611, 0.025056030601263046, 0.03879150003194809, 0.2105935513973236, -0.17785955965518951, -0.10178432613611221, 0.1861133575439453, -0.0507572740316391, -0.11791032552719116, 0.14624980092048645, -0.06176484003663063, 0.04954897612333298, 0.1214422658085823, 0.11154663562774658, -0.028116796165704727, -0.15089377760887146, -0.05262630432844162, 0.059978194534778595, 0.07035127282142639, 0.00772226694971323, 0.12953920662403107, -0.03431502729654312, -0.0816958099603653, -0.012883470393717289, -0.03238311782479286, 0.1239968091249466, -0.0073659662157297134, -0.06125391274690628, 0.03798297792673111, -0.10252250730991364, 0.1137363538146019, 0.03658721596002579, 0.03743981197476387, 0.03077615424990654, -0.11028007417917252, 0.09531612694263458, 0.07121419906616211, -0.08937004953622818, 0.016757765784859657, -0.06564803421497345, -0.007232989650219679, -0.20587952435016632, -0.00850912556052208, -0.12275850027799606, -0.07092026621103287, 0.003762359032407403, 0.03279094770550728, 0.10780978947877884, -0.05907274782657623, 0.08649636059999466, 0.05292188376188278, -0.07571958750486374, -0.00014137047401163727, 0.0485689640045166, -0.030288483947515488, -0.08248290419578552, -0.07833869755268097, -0.052204448729753494, -0.057369038462638855, 0.08902551978826523, -0.142148956656456, 0.04729302227497101, -0.1099342554807663, 0.095624640583992, 0.013367388397455215, -0.017342885956168175, 0.048334505409002304, 0.04178930073976517, 0.027691079303622246, -0.02463160641491413, 0.050411589443683624, -0.009013826958835125, -0.16682298481464386, 0.13152503967285156, -0.14421972632408142, 0.059558577835559845, 0.1436360478401184, 0.0029005524702370167, -0.009633020497858524, -0.07715852558612823, -0.014804118312895298, -0.027847254648804665, -0.0035396150778979063, -0.025901328772306442, 0.17750409245491028, 0.03856441006064415, 0.13743214309215546, -0.10834818333387375, 0.02704128436744213, -0.00837027933448553, -0.03866634517908096, -0.0616861991584301, 0.09227075427770615, 0.029757440090179443, -0.0505148284137249, 0.11412320286035538, 0.07554807513952255, -0.02131601795554161, 0.22504174709320068, 0.05746091157197952, -0.03983902558684349, 0.008765263482928276, -0.0314643494784832, -0.027171079069375992, -0.009798235259950161, -0.12200815975666046, -0.018229257315397263, 0.05734775587916374, -0.011721966788172722, -0.0009141371701844037, -0.0474611297249794, 0.007547453977167606, 0.01967012695968151, -0.06302964687347412, -0.043947383761405945, 0.013993272557854652, 0.003671756014227867, 0.07866600900888443, 0.007987641729414463, 0.05267764627933502, -0.005524108186364174, 0.0008830238366499543, -0.12027434259653091, 0.1779746413230896, -0.027460437268018723, -0.1455710530281067, -0.10577021539211273, -0.17930780351161957, -0.13369573652744293, 0.026721367612481117, 0.028892217203974724, -0.08353786915540695, -0.024518901482224464, -0.0660150796175003, 0.13434246182441711, 0.033789873123168945, -0.0613582581281662, -0.025543179363012314, -0.07722778618335724, 0.018874479457736015, -0.14244775474071503, -0.02965209260582924, 0.0435883067548275, -0.11211621016263962, -0.023481057956814766, -0.14118197560310364, 0.025356421247124672, 0.06798859685659409, -0.03352762386202812, 0.028798788785934448, -0.02582218125462532, 0.2026478797197342, -0.051438622176647186, 0.05696875974535942, 0.1423131376504898, -0.006204657256603241, 0.06568460166454315, 0.13098177313804626, 0.06655890494585037, -0.022439245134592056, 0.015255625359714031, 0.06736184656620026, -0.07314689457416534, -0.1715782880783081, -0.09525108337402344, -0.056159283965826035, -0.008620047010481358, 0.09230834245681763, 0.016006868332624435, 0.006425915285944939, 0.1095895767211914, -0.06689553707838058, 0.07571746408939362, 0.04291392117738724, 0.05471130087971687, -0.05799410492181778, 0.015959030017256737, 0.10212775319814682, -0.0883118286728859, -0.0666317343711853, 0.06469332426786423, 0.08407334983348846, 0.018070802092552185, -0.020744863897562027, 0.08563311398029327, 0.05464065819978714, 0.04224039986729622, 0.06974700093269348, 0.15002602338790894, -0.06261564046144485, -0.0025883300695568323, -0.025479095056653023, -0.08989766985177994, -0.07334834337234497, 0.05746240168809891, 0.13379694521427155, 0.0026781675405800343, 0.0017464762786403298, -0.11124534159898758, 0.053373727947473526, 0.1297604888677597, -0.008564026094973087, -0.3231552243232727, 0.008349566720426083, 0.04024193808436394, -0.014069564640522003, -0.09421233832836151, -0.05117306113243103, -0.005647231359034777, -0.13945652544498444, 0.1196109801530838, -0.013969178311526775, 0.11740563064813614, -0.005904032848775387, -0.02374807931482792, -0.07965793460607529, 0.03669553995132446, -0.05555877462029457, 0.08817939460277557, -0.2724568545818329, 0.17254383862018585, 0.057188864797353745, 0.03547811508178711, -0.08531248569488525, -0.013476746156811714, 0.12320473790168762, 0.019014829769730568, 0.16402247548103333, -0.006169482599943876, 0.2466878741979599, -0.07711675018072128, -0.1270812749862671, 0.03464905172586441, -0.051033519208431244, -0.11697499454021454, 0.04141003265976906, 0.03571897745132446, 0.00026487279683351517, -0.020414430648088455, 0.1952374428510666, -0.08658009767532349, -0.15388843417167664, -0.05006632208824158, -0.014724901877343655, 0.017454778775572777, 0.011359885334968567, -0.08163829892873764, -0.11120449751615524, 0.08594690263271332, 0.034670550376176834, -0.14910563826560974, -0.10417216271162033, -0.024524103850126266, 0.10178907960653305, -0.11050879210233688, -0.030478598549962044, -0.0748811885714531, 0.06774070858955383, -0.12745651602745056, -0.10960119962692261, 0.03321050852537155, -0.03362826630473137, 0.03766927495598793, 0.023494409397244453, 0.1079292967915535, 0.05542691424489021, 0.05846651270985603, 0.002426060615107417, 0.007201053202152252, -0.014499071054160595, -0.051286567002534866, -0.05839831009507179, 0.11532989889383316, 0.09224569797515869, 0.15254506468772888, -0.08673352003097534, -0.09517159312963486, -0.08970719575881958, 0.02312411181628704, 0.11828206479549408, 0.17326141893863678, -0.009925369173288345, 0.0651465356349945, 0.2568816840648651, -0.06491915881633759, -0.24479587376117706, -0.11499396711587906, 0.07768310606479645, 0.07255514711141586, 0.012903480790555477, -0.18703089654445648, 0.07504567503929138, -0.022905664518475533, 0.0072785429656505585, -0.10972172021865845, -0.19264870882034302, -0.10729876905679703, 0.15743273496627808, 0.046730443835258484, 0.10198050737380981, -0.12492790073156357, -0.011223712004721165, -0.09853795915842056, -0.05079957842826843, 0.1289598047733307, -0.03596825152635574, 0.09788180142641068, 0.012248478829860687, 0.025802411139011383, 0.03531840071082115, -0.030615292489528656, 0.13234949111938477, 0.026613889262080193, 0.016639946028590202, -0.07210353761911392, -0.07660359889268875, -0.026027437299489975, -0.04793662577867508, 0.14142799377441406, -0.0936785638332367, -0.006693713366985321, -0.19785119593143463, -0.08700872957706451, -0.0643642246723175, 0.03858131170272827, 0.0051352656446397305, -0.05491013824939728, 0.030891787260770798, 0.007425794377923012, 0.07616035640239716, 0.0478619709610939, 0.09501960128545761, -0.0739152580499649, 0.11849136650562286, 0.18789497017860413, 0.19961099326610565, -0.03102576918900013, -0.03811225667595863, -0.027269819751381874, -0.019392861053347588, 0.07977549731731415, -0.22191305458545685, 0.03450358286499977, 0.07153691351413727, -0.019832663238048553, 0.14884205162525177, 0.012690113857388496, -0.10200690478086472, 0.004080280661582947, 0.07854223996400833, -0.09021313488483429, -0.10941701382398605, -0.04881076142191887, -0.06481719762086868, -0.13435408473014832, -0.08896942436695099, 0.13217701017856598, -0.0680869072675705, -0.030149998143315315, -0.0035719764418900013, 0.011398661881685257, -0.12162096053361893, 0.09036213904619217, 0.047981344163417816, 0.037629928439855576, -0.08764952421188354, 0.11984624713659286, 0.08981072902679443, -0.07998181879520416, 0.027837341651320457, 0.13211697340011597, -0.07025889307260513, -0.0546722374856472, -0.0744372084736824, 0.2590835988521576, -0.12134715169668198, -0.03007739968597889, -0.002389129949733615, -0.037846751511096954, -0.036122169345617294, 0.07469384372234344, 0.04407497122883797, 0.01860114373266697, -0.07041899859905243, -0.0331677608191967, -0.04688185825943947, 0.08598611503839493, 0.08579348772764206, -0.027046911418437958, -0.03689282760024071, 0.050468843430280685, 0.02561226859688759, 0.119601309299469, -0.04121134430170059, -0.0027865904849022627, -0.09663590788841248, 0.007652849890291691, -0.09414278715848923, 0.0171838216483593, -0.06923043727874756, 0.01573360525071621, -0.04907316714525223, -0.025758320465683937, 0.008787221275269985, -0.043439194560050964, -0.07881339639425278, 0.0031456374563276768, -0.013017498888075352, 0.1195196807384491, -0.08836156129837036, -0.005823582876473665, 0.04343109950423241, -0.029486054554581642, 0.06287407130002975, 0.03871894255280495, -0.030812930315732956, 0.033133428543806076, -0.06039155647158623, -0.008110818453133106, 0.0035129885654896498, 0.0005750837153755128, 0.021113082766532898, -0.02532823756337166, 0.054663218557834625, 0.0017386487452313304, 0.03353126719594002, -0.01688232272863388, 0.1052614152431488, -0.04451948404312134, 0.036257196217775345, 0.007659601978957653, -0.03577626496553421, -0.05155055224895477, 0.10360932350158691, 0.049809183925390244, 0.12137681990861893, 0.03415418043732643, -0.03797628730535507, 0.018392177298665047, -0.13742481172084808, -0.009812694974243641, -0.022423718124628067, -0.06498335301876068, 0.009191141463816166, -0.10841212421655655, 0.045470234006643295, -0.0012137877056375146, 0.16026300191879272, 0.08981799334287643, -0.00876986887305975, 0.002669990062713623, 0.029522733762860298, 0.05648748576641083, -0.0507672019302845, 0.13737230002880096, -0.023475226014852524, 0.001935587264597416, 0.05518807843327522, 0.058427903801202774, -0.010091746225953102, 0.13255587220191956, -0.006938132457435131, 0.1499171108007431, 0.05084554851055145, 0.055804543197155, 0.02668367326259613, -0.0013078219490125775, -0.017863944172859192, -0.0800524652004242, 0.050437189638614655, -0.01583143137395382, -0.095233254134655, 0.10464757680892944, 0.16147159039974213, -0.15779709815979004, 0.09916006028652191, 0.09130492806434631, -0.06511422246694565, -0.12422680854797363, -0.1794019192457199, -0.03986254706978798, -0.06912262737751007, -0.008832600899040699, -0.10244698077440262, -0.0016240618424490094, 0.19303564727306366, 0.019798623397946358, 0.03406870365142822, 0.18175643682479858, -0.1403936743736267, -0.09453173726797104, 0.05110059306025505, -0.016816167160868645, 0.051031000912189484, 0.008375018835067749, 0.04963891953229904, 0.07947038859128952, 0.03613390401005745, 0.06808245182037354, 0.0187518447637558, 0.03977275267243385, 0.093634694814682, -0.0355207584798336, -0.11428883671760559, -0.060382381081581116, 0.0036157099530100822, 0.04439928010106087, 0.23644199967384338, 0.029013698920607567, -0.03533885255455971, -0.04214892536401749, 0.09587851911783218, -0.012685701251029968, -0.060192596167325974, -0.16606172919273376, 0.17387522757053375, -0.024071866646409035, -0.017372671514749527, -0.021159840747714043, -0.05422715097665787, -0.042778484523296356, 0.2012624740600586, 0.24567897617816925, -0.05315317586064339, -0.0018135583959519863, -0.016562320291996002, 0.031266260892152786, 0.02138696424663067, 0.06068487837910652, 0.024261945858597755, 0.13491325080394745, -0.1151886060833931, 0.08211196959018707, -0.02727261558175087, -0.04617786779999733, -0.029336344450712204, 0.07260158658027649, -0.02195156365633011, -0.031177956610918045, -0.008834820240736008, 0.07965876162052155, -0.06451698392629623, -0.15695339441299438, -0.04404335841536522, -0.10913686454296112, -0.04404276981949806, -0.05897635966539383, -0.11562750488519669, 0.10695110261440277, 0.017004484310746193, -0.009012825787067413, 0.0030114916153252125, 0.09604346007108688, 0.016907433047890663, -0.13260266184806824, -0.08549502491950989, 0.08124358206987381, 0.03652303293347359, 0.12816649675369263, 0.01645718701183796, 0.004581952001899481, 0.06728313118219376, -0.027458904311060905, -0.0126733323559165, 0.054599273949861526, 0.0034572924487292767, 0.05160944163799286, 0.005899534560739994, 0.06820080429315567, -0.038386281579732895, 0.013546318747103214, -0.03728602081537247, -0.0687447339296341, 0.05984751507639885, -0.010522205382585526, -0.07415201514959335, -0.0532803051173687, 0.10488267987966537, -0.1200612410902977, 0.11611095815896988, 0.19364085793495178, 0.011894227005541325, -0.02129579521715641, -0.037757206708192825, 0.030104899778962135, 0.025765400379896164, -0.05965176224708557, -0.05989539250731468, -0.10293331742286682, -0.017693795263767242, -0.10781671851873398, 0.004134840797632933, -0.27389317750930786, 0.01597508043050766, -0.05546817556023598, -0.019132593646645546, -0.05509631708264351, 0.046835433691740036, 0.043438732624053955, 0.03671870008111, -0.0038149470929056406, -0.3176715672016144, 0.028974656015634537, 0.07332773506641388, -0.12545360624790192, -0.0950804203748703 ]
null
null
transformers
# BatteryBERT-cased for Battery Abstract Classification **Language model:** batterybert-cased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 11 base_LM_model = "batterybert-cased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 97.29, "Test accuracy": 96.85, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batterybert-cased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batterybert-cased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryBERT-cased for Battery Abstract Classification Language model: batterybert-cased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-cased for Battery Abstract Classification \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryBERT-cased for Battery Abstract Classification \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 69, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryBERT-cased for Battery Abstract Classification \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.06621818244457245, 0.047621406614780426, -0.0034733200445771217, 0.04851040616631508, 0.25448092818260193, 0.029042290523648262, 0.17969052493572235, 0.03427138552069664, -0.10801959037780762, -0.038759876042604446, 0.042869798839092255, 0.03317273035645485, 0.04895966127514839, 0.06076286733150482, -0.03576920926570892, -0.18589478731155396, 0.03891358897089958, 0.004405754152685404, 0.026488736271858215, 0.12215378880500793, 0.14108961820602417, -0.09149077534675598, 0.11846058815717697, 0.021844904869794846, -0.11352993547916412, 0.02782781422138214, 0.02233487367630005, -0.1409195512533188, 0.13387560844421387, -0.05997949838638306, 0.16758635640144348, 0.03564565256237984, 0.040654465556144714, -0.1519060581922531, 0.03281746059656143, 0.01432329136878252, 0.007855506613850594, 0.07127919048070908, 0.021707085892558098, -0.09867405146360397, 0.10533203184604645, -0.010418154299259186, 0.0799325630068779, 0.02340642735362053, -0.09036216139793396, -0.17769812047481537, -0.026185305789113045, 0.021391984075307846, 0.10733853280544281, 0.12017195671796799, -0.011212167330086231, 0.0703529641032219, -0.045972246676683426, 0.06938337534666061, 0.21604947745800018, -0.118291474878788, 0.006488127168267965, 0.11050522327423096, 0.0239714402705431, -0.03090936318039894, -0.06184836104512215, -0.031835708767175674, 0.0061904857866466045, 0.049113765358924866, 0.07656661421060562, -0.03516419976949692, -0.15550296008586884, 0.10453162342309952, -0.1002286821603775, 0.007618871051818132, 0.23486433923244476, -0.08046972751617432, -0.0316946841776371, -0.04530306160449982, -0.09161841124296188, -0.07932236790657043, -0.05149874463677406, 0.009806238114833832, -0.03350827097892761, 0.015977274626493454, -0.017975829541683197, 0.05756789445877075, -0.05427098646759987, -0.07054838538169861, -0.11943972855806351, 0.11226348578929901, 0.006621957290917635, 0.0010640877299010754, -0.11233773082494736, 0.11812730133533478, 0.0738573968410492, -0.15862184762954712, 0.013722646981477737, -0.05679704621434212, -0.02156849205493927, -0.030811870470643044, 0.007260884623974562, -0.002764670643955469, 0.08171770721673965, 0.09133724123239517, -0.012562344782054424, 0.00513225793838501, 0.02352764829993248, -0.01448792964220047, 0.0013654041104018688, 0.1488543450832367, -0.02876461111009121, -0.06535280495882034, 0.025547606870532036, -0.009816269390285015, 0.01992890238761902, -0.03555261716246605, -0.1310673952102661, -0.013935025781393051, 0.08468131721019745, 0.05517914518713951, -0.0343225933611393, 0.06417962163686752, -0.07551367580890656, -0.05503300949931145, 0.07343651354312897, -0.10248373448848724, -0.019605282694101334, 0.0014749227557331324, 0.038198597729206085, 0.06185488402843475, -0.017629601061344147, 0.01978939026594162, -0.024024954065680504, 0.07875630259513855, -0.0883207693696022, -0.0528113953769207, -0.10478920489549637, -0.08841358125209808, -0.008409897796809673, 0.08374156057834625, 0.07205545902252197, -0.12471245974302292, -0.2658311128616333, 0.0775371640920639, 0.005365531425923109, 0.0015934424009174109, -0.059461742639541626, -0.061421655118465424, -0.07721325010061264, 0.011785288341343403, -0.006990056484937668, 0.051190368831157684, -0.07128490507602692, 0.046624500304460526, 0.10443320125341415, 0.032876864075660706, -0.06411782652139664, 0.04114047810435295, -0.13308577239513397, -0.006613691337406635, -0.19849702715873718, 0.016228221356868744, -0.09388832002878189, 0.08234661817550659, -0.07504178583621979, -0.06145946681499481, 0.1345883309841156, -0.01921452395617962, 0.01335680391639471, 0.14689680933952332, -0.07037973403930664, -0.15526923537254333, 0.03905908390879631, -0.1380842626094818, -0.09674271196126938, 0.038561511784791946, -0.043075080960989, 0.025004146620631218, 0.06764543801546097, 0.04026537388563156, 0.08355720341205597, -0.1289396733045578, -0.10801009833812714, 0.015344907529652119, 0.0075369421392679214, -0.02179013378918171, 0.09756563603878021, 0.061219483613967896, -0.06017381325364113, 0.07083970308303833, -0.03982628509402275, 0.07345293462276459, -0.040965430438518524, -0.08716632425785065, -0.021567165851593018, -0.01974073424935341, 0.12346234172582626, 0.02793174423277378, 0.009262913838028908, -0.0018572438275441527, -0.09151975065469742, 0.16803359985351562, 0.0888492539525032, -0.04621293395757675, 0.004296836443245411, -0.14898811280727386, 0.08987423777580261, 0.0390339121222496, 0.010902379639446735, -0.11071399599313736, -0.1094307079911232, 0.03230677545070648, 0.06098165735602379, -0.0482352077960968, -0.06733822077512741, 0.008120827376842499, 0.033535879105329514, -0.10532969981431961, -0.012794224545359612, -0.08184532821178436, 0.008825032040476799, 0.00154738943092525, -0.22318574786186218, 0.05812856927514076, -0.03493442013859749, 0.1442514806985855, -0.0865638330578804, 0.05170699954032898, 0.14573614299297333, 0.07743364572525024, -0.023622147738933563, 0.00459142355248332, -0.021673716604709625, 0.04177147150039673, -0.07429216802120209, -0.04020160064101219, 0.0001977212232304737, 0.0005157170235179365, -0.05020802095532417, 0.024328747764229774, -0.14455710351467133, 0.12058588862419128, 0.18200398981571198, -0.007550790905952454, -0.04129454120993614, -0.050305161625146866, -0.00047249693307094276, -0.021419497206807137, -0.10043956339359283, -0.0038829767145216465, 0.2359028160572052, -0.017199493944644928, 0.07863670587539673, -0.03163262829184532, -0.014618202112615108, 0.0017956142546609044, 0.0319330208003521, 0.04567126929759979, 0.10944490879774094, 0.11056370288133621, -0.023599347099661827, 0.1078195869922638, 0.04652164876461029, -0.07282153517007828, 0.14921410381793976, -0.0675898939371109, -0.01907946541905403, 0.00033818589872680604, -0.11626721918582916, -0.003464051755145192, 0.2215927690267563, -0.18478776514530182, 0.012179587967693806, 0.06139039993286133, -0.040054868906736374, 0.013509178534150124, -0.08661564439535141, -0.017117593437433243, -0.00589701859280467, -0.012413297779858112, -0.09619757533073425, 0.009155791252851486, 0.0481075793504715, 0.056729789823293686, -0.010964863933622837, -0.10287106782197952, 0.06643438339233398, 0.019666051492094994, -0.10902542620897293, 0.23507706820964813, -0.11507163196802139, -0.24968108534812927, -0.13741439580917358, -0.09477037936449051, -0.0011103339493274689, -0.014758100733160973, 0.018086586147546768, -0.10861155390739441, -0.1092645525932312, 0.00846624281257391, 0.13408319652080536, -0.02009875327348709, 0.08313285559415817, 0.0705425962805748, 0.024819649755954742, -0.03031720034778118, -0.10374076664447784, 0.0033757758792489767, 0.04499628022313118, -0.041943710297346115, 0.036547716706991196, 0.03962421417236328, 0.036988817155361176, 0.18415021896362305, 0.0032186112366616726, 0.013773531652987003, -0.058398496359586716, 0.14285418391227722, -0.11037968099117279, -0.0363006666302681, 0.2155192643404007, 0.03503413125872612, -0.007921580225229263, 0.16138291358947754, -0.004110304638743401, -0.015125435777008533, 0.06848575174808502, -0.03110889345407486, -0.057569317519664764, -0.2519897222518921, -0.1360781043767929, -0.06116369739174843, 0.09867964684963226, -0.03303831070661545, 0.029082641005516052, 0.020385688170790672, 0.06174520403146744, 0.005537411663681269, 0.06505127996206284, -0.038140345364809036, 0.12905585765838623, 0.23784776031970978, 0.05587189272046089, 0.09355048835277557, -0.01992841809988022, -0.046352460980415344, 0.057118091732263565, -0.04826094210147858, 0.12543809413909912, -0.0007111481972970068, 0.05205812677741051, -0.0015106622595340014, 0.030410369858145714, 0.07105381786823273, 0.0569237545132637, 0.004565062467008829, -0.016962287947535515, 0.02392260544002056, -0.0370173305273056, 0.04257979616522789, -0.013252433389425278, -0.09574691951274872, -0.01369596179574728, -0.043545059859752655, 0.08371993899345398, 0.07340396195650101, -0.03647059202194214, 0.02275005169212818, -0.25396591424942017, -0.06275516748428345, 0.05131927877664566, -0.0620892159640789, -0.09985522925853729, 0.032524578273296356, -0.019932495430111885, -0.13039281964302063, 0.07247651368379593, -0.025102950632572174, 0.07119417935609818, -0.10445796698331833, 0.056832898408174515, 0.04579072818160057, 0.06070248410105705, 0.00006773286440875381, 0.12116114795207977, -0.34706053137779236, 0.13486018776893616, 0.0407765693962574, 0.05239779129624367, -0.028753826394677162, 0.042485710233449936, -0.031161939725279808, 0.05612201243638992, 0.03396034613251686, -0.003832827555015683, 0.03886588290333748, 0.017811181023716927, -0.12286674976348877, 0.062360282987356186, 0.017562925815582275, -0.04150748997926712, 0.03913635388016701, -0.02682027779519558, 0.008721181191504002, 0.038567978888750076, -0.07911636680364609, -0.09824403375387192, -0.12668898701667786, 0.07780171930789948, -0.03560209646821022, 0.12987656891345978, -0.10311085730791092, -0.11093663424253464, -0.007346123922616243, 0.11884844303131104, -0.11751449853181839, -0.10476324707269669, -0.1476205438375473, 0.011728033423423767, 0.0533977709710598, -0.07459352910518646, 0.08178841322660446, -0.0008464416605420411, 0.09158439934253693, -0.041966453194618225, -0.11446665972471237, 0.1078641414642334, -0.07486523687839508, -0.17778822779655457, -0.046076126396656036, 0.10147777944803238, 0.10084649175405502, 0.06349960714578629, -0.01756991632282734, 0.0005949657643213868, -0.08906356245279312, -0.12778569757938385, -0.015303011983633041, -0.010065206326544285, 0.06294752657413483, 0.009177742525935173, 0.07179853320121765, -0.09326373040676117, -0.024282827973365784, -0.055283207446336746, 0.07582321763038635, 0.16228808462619781, -0.031334757804870605, 0.10043438524007797, 0.07008957117795944, -0.04422170668840408, -0.2421455979347229, -0.052089352160692215, 0.007087660022079945, 0.041221313178539276, -0.04679911956191063, -0.25287938117980957, 0.15783172845840454, -0.03808191046118736, -0.06595077365636826, -0.044255129992961884, -0.1407458782196045, -0.13596102595329285, 0.21469049155712128, 0.03021518886089325, 0.10845726728439331, -0.08296641707420349, -0.014832915738224983, -0.08674325048923492, -0.20298254489898682, 0.21197061240673065, -0.13732412457466125, 0.043487079441547394, -0.015696095302700996, 0.18544207513332367, -0.016589343547821045, -0.04091272130608559, 0.12501531839370728, 0.03494300693273544, 0.002116600051522255, -0.00927862711250782, -0.09208962321281433, 0.21565556526184082, 0.0413234643638134, 0.008113743737339973, -0.04808570444583893, 0.04582080990076065, -0.09775489568710327, -0.051251161843538284, -0.025074932724237442, 0.11519122868776321, -0.02777043916285038, -0.15172037482261658, -0.10653788596391678, 0.027940254658460617, -0.034185558557510376, -0.056308597326278687, 0.20417039096355438, 0.074412040412426, 0.027041956782341003, 0.020993219688534737, 0.21170847117900848, -0.015391665510833263, 0.037217166274785995, 0.00771663011983037, -0.023790692910552025, 0.0592755563557148, -0.17061255872249603, 0.03016551025211811, 0.09736528992652893, 0.03782124072313309, 0.038796234875917435, 0.10527550429105759, -0.05796471983194351, -0.024923935532569885, 0.103916697204113, -0.22924353182315826, -0.04930713400244713, -0.09952793270349503, 0.14323846995830536, -0.13408727943897247, 0.08181021362543106, 0.1127849891781807, -0.04964422434568405, -0.0644385814666748, 0.04625800624489784, 0.03077717125415802, -0.04096051678061485, 0.12313349545001984, 0.051271941512823105, 0.11297930777072906, -0.13784164190292358, 0.04877679422497749, 0.061197429895401, 0.029105691239237785, -0.028104225173592567, 0.07243169099092484, -0.13276852667331696, -0.008577202446758747, 0.030240552499890327, 0.19004279375076294, -0.12306849658489227, -0.051543883979320526, -0.19445106387138367, -0.16245903074741364, 0.03062017448246479, 0.1422191560268402, 0.11247912794351578, 0.09357725828886032, -0.055706936866045, 0.015463094227015972, -0.10426623374223709, 0.08911652117967606, 0.007122712209820747, 0.027547864243388176, -0.1520310640335083, 0.030237123370170593, -0.009600087068974972, 0.04514516517519951, -0.06747927516698837, 0.002183435019105673, -0.15205314755439758, 0.024708708748221397, -0.06982821971178055, -0.06576376408338547, -0.04727083072066307, -0.022890564054250717, 0.014978953637182713, -0.04385649040341377, -0.020196855068206787, 0.03973265364766121, -0.04270036146044731, 0.06557904928922653, 0.022937029600143433, 0.05473771691322327, -0.10740836709737778, 0.06947777420282364, 0.07503653317689896, -0.062471963465213776, 0.12222590297460556, 0.09131573140621185, 0.07381204515695572, 0.11025924980640411, -0.16072800755500793, 0.014868568629026413, 0.061488449573516846, 0.048182275146245956, 0.056492406874895096, -0.12903650104999542, -0.0030098112765699625, 0.0034865487832576036, 0.0279680248349905, 0.05028466507792473, 0.045100148767232895, -0.10392841696739197, 0.05982891842722893, -0.04588760435581207, -0.013992260210216045, -0.052886802703142166, 0.018193451687693596, 0.13388857245445251, 0.03389599546790123, 0.12458973377943039, -0.061462994664907455, 0.00772865442559123, -0.08043811470270157, 0.02918589860200882, -0.03201936557888985, -0.06998280435800552, -0.08165893703699112, -0.04645754396915436, 0.047838520258665085, 0.010811582207679749, 0.21001192927360535, 0.025252940133213997, -0.013333373703062534, 0.06028097867965698, 0.16103166341781616, 0.07490399479866028, 0.04241860657930374, 0.11956924945116043, 0.10741341859102249, 0.05138968676328659, 0.08476223051548004, 0.005212985910475254, 0.12590712308883667, 0.09687639772891998, 0.14897745847702026, 0.034044619649648666, -0.01884663850069046, 0.04730920493602753, 0.01648664101958275, -0.09483591467142105, -0.1591184288263321, 0.019486935809254646, -0.07617254555225372, 0.03296518325805664, -0.03984476253390312, 0.19270585477352142, 0.11635913699865341, -0.09961041063070297, 0.05566047877073288, -0.02498583495616913, -0.09623635560274124, -0.10031957179307938, 0.006806078366935253, -0.04081057012081146, -0.1916644275188446, 0.029558688402175903, -0.13581620156764984, -0.09519702941179276, 0.13820290565490723, -0.0014148802729323506, -0.029653199017047882, 0.09417541325092316, 0.18002817034721375, -0.023174051195383072, 0.0251685231924057, -0.008743716403841972, -0.011045694351196289, -0.025114241987466812, 0.0025731550995260477, -0.08594769239425659, 0.06373131275177002, 0.005338930990546942, -0.0119386101141572, -0.0378752239048481, 0.013893665745854378, -0.014125934801995754, -0.0628848746418953, 0.011578378267586231, -0.008447673171758652, -0.07960177958011627, 0.15806077420711517, -0.01450074091553688, 0.06658967584371567, 0.04367147386074066, 0.28019431233406067, -0.06468916684389114, -0.15986208617687225, -0.10123956203460693, 0.10822421312332153, -0.012901391834020615, 0.03980657458305359, 0.05395132675766945, -0.08162985742092133, -0.09995594620704651, 0.1798306554555893, 0.10627254098653793, 0.009319706819951534, -0.010370256379246712, -0.0018473585369065404, -0.012523874640464783, -0.007746622432023287, 0.021615486592054367, 0.07718979567289352, 0.18355078995227814, -0.07085757702589035, -0.06144194304943085, -0.054381709545850754, -0.05576818063855171, -0.04584623500704765, 0.04925195127725601, 0.04149569198489189, -0.04091464728116989, -0.018756408244371414, 0.0546841099858284, -0.17609649896621704, -0.10958579927682877, -0.08952927589416504, -0.10282915830612183, -0.09070415049791336, -0.06010788679122925, 0.17275398969650269, -0.008143561892211437, 0.020283853635191917, -0.009439644403755665, 0.033812034875154495, 0.027624737471342087, 0.056601881980895996, -0.14923083782196045, -0.006714758928865194, 0.1399754136800766, -0.15482869744300842, 0.1337120682001114, -0.06269875913858414, 0.0602492019534111, 0.115780308842659, 0.032704878598451614, -0.07862111181020737, 0.00984890479594469, 0.03331487998366356, -0.14621281623840332, 0.007105122320353985, 0.11422983556985855, -0.0006670271977782249, 0.10681087523698807, -0.008327779360115528, 0.00910256989300251, 0.0036678980104625225, -0.03108031488955021, -0.07446702569723129, -0.04669875651597977, -0.04496454447507858, -0.07364756613969803, 0.05493738502264023, 0.08127860724925995, -0.06051435321569443, 0.016520269215106964, -0.07439526170492172, 0.05308951064944267, -0.01569235697388649, 0.002226901240646839, -0.014914946630597115, -0.15060342848300934, -0.06821057200431824, 0.18529585003852844, 0.001016997266560793, -0.13526467978954315, 0.01428663358092308, -0.04253356531262398, -0.004823018796741962, -0.1473890244960785, 0.09230969846248627, 0.08911612629890442, -0.029703008010983467, -0.025219278410077095, 0.01485325489193201, -0.02434423193335533, 0.07367143034934998, -0.18095360696315765, -0.09485147893428802 ]
null
null
transformers
# BatteryBERT-cased for QA **Language model:** batterybert-cased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 16 n_epochs = 4 base_LM_model = "batterybert-cased" max_seq_len = 386 learning_rate = 2e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 81.54, "f1": 89.16, ``` Evaluated on the battery device dataset. ``` "precision": 70.74, "recall": 84.19, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batterybert-cased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batterybert-cased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatteryBERT-cased for QA Language model: batterybert-cased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-cased for QA \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatteryBERT-cased for QA \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 64, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatteryBERT-cased for QA \r\nLanguage model: batterybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.07348915934562683, 0.02780606970191002, -0.004550732672214508, 0.05600939318537712, 0.22088462114334106, 0.04112272337079048, 0.11025670915842056, 0.04638279229402542, -0.1021728366613388, -0.02056519314646721, 0.04531377553939819, -0.033153459429740906, 0.077457956969738, 0.061577584594488144, 0.008907143026590347, -0.08559592068195343, 0.03550928458571434, -0.02996065653860569, 0.006684481166303158, 0.14341962337493896, 0.11783464252948761, -0.09675288200378418, 0.12642501294612885, 0.016437826678156853, -0.09183673560619354, 0.07271864265203476, -0.0068285297602415085, -0.07543203234672546, 0.10607361793518066, -0.08078379184007645, 0.13215509057044983, 0.04941032826900482, 0.018205827102065086, -0.18699435889720917, 0.039825525134801865, 0.0249712523072958, 0.027661291882395744, 0.03745715320110321, 0.014685374684631824, -0.046549249440431595, 0.04560922458767891, 0.030519021674990654, 0.04790642112493515, 0.02067605033516884, -0.05850983411073685, -0.22918963432312012, -0.08073554933071136, -0.056387703865766525, 0.07777821272611618, 0.14086167514324188, -0.026008684188127518, 0.09580451250076294, -0.040598344057798386, 0.07190389186143875, 0.18667176365852356, -0.10833998769521713, -0.00043991924030706286, 0.09520997107028961, 0.05387645587325096, -0.006569480989128351, -0.027070531621575356, -0.07255398482084274, 0.006752267014235258, 0.056001048535108566, 0.052211396396160126, -0.05354128032922745, -0.12284650653600693, 0.11433903127908707, -0.09172409772872925, -0.017258070409297943, 0.21258074045181274, -0.03589781001210213, -0.07039609551429749, 0.010817830450832844, -0.11042748391628265, -0.03713330626487732, -0.017702188342809677, -0.02026190422475338, -0.03364210203289986, -0.0024433585349470377, -0.08518078923225403, 0.03468189388513565, -0.024290116503834724, -0.10586513578891754, -0.12633727490901947, 0.05967118963599205, 0.051262062042951584, 0.038677364587783813, -0.12048026919364929, 0.07882636040449142, 0.0009550885879434645, -0.15194442868232727, -0.03390265628695488, -0.05936320126056671, -0.07074375450611115, -0.00006788316386519, 0.02754049375653267, 0.00700342608615756, 0.12104611843824387, 0.08909592777490616, -0.028079485520720482, 0.012430230155587196, 0.014639775268733501, -0.06463661789894104, 0.011697052046656609, 0.1481056958436966, -0.017504647374153137, -0.15000252425670624, -0.012998081743717194, 0.024596909061074257, -0.01195805799216032, -0.01694120652973652, -0.059838421642780304, 0.004149885382503271, 0.06037113070487976, 0.062145162373781204, 0.031253717839717865, 0.02542020194232464, -0.056612882763147354, -0.024302203208208084, 0.04209301993250847, -0.10127467662096024, -0.037834782153367996, -0.02845681458711624, 0.024137213826179504, 0.004550626501441002, -0.052087560296058655, 0.025256212800741196, -0.025817375630140305, -0.011969296261668205, -0.08327014744281769, -0.10204202681779861, -0.1171434074640274, -0.08061453700065613, -0.028082003816962242, 0.12809494137763977, 0.06662236154079437, -0.11922658234834671, -0.31700804829597473, 0.08439885824918747, 0.0006552785634994507, 0.00013697471877094358, -0.036815039813518524, -0.030489642173051834, -0.052682481706142426, -0.02803669683635235, -0.029190784320235252, 0.11555243283510208, -0.06370282173156738, 0.06949491798877716, 0.1663465052843094, 0.024419479072093964, -0.03488190844655037, 0.039574068039655685, -0.14352527260780334, 0.01342430617660284, -0.1344693899154663, 0.03157627210021019, -0.1178315058350563, 0.006688274908810854, -0.12074943631887436, -0.07462015002965927, 0.11362828314304352, -0.021260591223835945, 0.07610282301902771, 0.1421036422252655, -0.07682599127292633, -0.1293756514787674, 0.03078800067305565, -0.13380907475948334, -0.16654574871063232, 0.04171931371092796, -0.017286134883761406, 0.01722383126616478, 0.0234694741666317, 0.038665156811475754, 0.1363268345594406, -0.18026326596736908, -0.18483485281467438, -0.03899161517620087, 0.03438505530357361, -0.010703731328248978, 0.090732641518116, 0.04941139742732048, -0.06773271411657333, 0.07158840447664261, -0.06138911470770836, 0.04774310812354088, -0.052409153431653976, -0.1074286699295044, -0.05902419239282608, -0.028842343017458916, 0.12204848974943161, 0.004102325066924095, -0.05919589847326279, 0.012067874893546104, -0.0966872051358223, 0.09819342195987701, 0.10665252059698105, -0.03444148972630501, -0.0029858173802495003, -0.19456908106803894, 0.11022020876407623, -0.038430649787187576, 0.012913485988974571, -0.10194934904575348, -0.16161178052425385, 0.07645922899246216, -0.01617586798965931, -0.07805129140615463, -0.04514139145612717, 0.024623366072773933, 0.022857733070850372, -0.11295583099126816, -0.029904739931225777, -0.1531280279159546, -0.020397741347551346, 0.013187102042138577, -0.15657790005207062, 0.045170847326517105, -0.02850205823779106, 0.08469577133655548, 0.019455719739198685, 0.015133864246308804, 0.13945087790489197, 0.04887249320745468, -0.01443935465067625, -0.012266728095710278, -0.03710057958960533, 0.05684901028871536, -0.041147343814373016, -0.03097747452557087, -0.03947442024946213, -0.025067990645766258, -0.0004521610972005874, -0.03283900022506714, -0.10127328336238861, 0.06868840008974075, 0.13898374140262604, 0.11673372983932495, 0.01506083458662033, -0.08788937330245972, -0.003093949519097805, -0.04244176298379898, -0.098824143409729, -0.03594896197319031, 0.20183689892292023, 0.005446895491331816, 0.01759088598191738, -0.028260866180062294, -0.0024989191442728043, 0.0020218826830387115, 0.039724886417388916, 0.054036516696214676, 0.08556608110666275, 0.11384427547454834, 0.05598113313317299, 0.0642072856426239, 0.08889755606651306, -0.050840724259614944, 0.11674017459154129, -0.0846620723605156, -0.040128227323293686, -0.04465983435511589, -0.10787926614284515, -0.020668577402830124, 0.2697892487049103, -0.13060614466667175, 0.06624726206064224, 0.06393320858478546, -0.0028710931073874235, 0.021778428927063942, -0.02813936211168766, -0.008174646645784378, -0.011746319010853767, -0.04746770113706589, -0.1660061627626419, 0.023576868698000908, 0.05726192146539688, 0.003687553107738495, -0.01459829043596983, -0.04379187524318695, 0.022352971136569977, 0.011722433380782604, -0.06907988339662552, 0.21754330396652222, -0.1063985750079155, -0.19913038611412048, -0.11926114559173584, -0.10595971345901489, -0.01188552938401699, -0.07544513791799545, 0.025766924023628235, -0.12840570509433746, -0.09589780867099762, 0.02428118884563446, 0.18654604256153107, 0.014456313103437424, 0.0778186023235321, 0.04535718262195587, 0.014165256172418594, 0.007727126590907574, -0.10664796084165573, 0.011370319873094559, 0.06703691929578781, -0.09541863203048706, 0.031957462430000305, 0.0587780736386776, 0.08278761804103851, 0.08682848513126373, 0.012263829819858074, 0.015537770465016365, -0.054664045572280884, 0.231400728225708, -0.10792899876832962, -0.03181038424372673, 0.27761322259902954, 0.11423636227846146, -0.004234977066516876, 0.19669818878173828, -0.04076005145907402, -0.013307701796293259, 0.09715192764997482, -0.004165381658822298, -0.04737413674592972, -0.2550515830516815, -0.07735259085893631, -0.035267528146505356, 0.07709464430809021, -0.06908458471298218, 0.008879046887159348, -0.04315178841352463, 0.08559410274028778, 0.020149538293480873, -0.024121612310409546, -0.10489080846309662, 0.094483382999897, 0.24053916335105896, 0.05103293061256409, 0.0960647314786911, -0.014398922212421894, 0.017418835312128067, 0.07239622622728348, 0.06523223966360092, 0.17268657684326172, -0.04299080744385719, 0.036095742136240005, 0.036707453429698944, 0.15766453742980957, 0.027008626610040665, 0.02892637439072132, -0.013411213643848896, -0.0033802371472120285, 0.018055206164717674, -0.03352845832705498, 0.0169354360550642, -0.022524572908878326, -0.024905163794755936, 0.011107933707535267, -0.011881865561008453, 0.15945130586624146, 0.04977662116289139, -0.022836217656731606, 0.02473742514848709, -0.1942245215177536, -0.03125826269388199, 0.04697173833847046, -0.056565552949905396, -0.09590146690607071, 0.02654331922531128, 0.028889713808894157, -0.12916846573352814, 0.030841104686260223, -0.03617909178137779, 0.0724499523639679, -0.06526296585798264, 0.04952109605073929, 0.06209652125835419, 0.09869702905416489, 0.01382235437631607, 0.12020813673734665, -0.3718373477458954, 0.10598522424697876, 0.04255788028240204, 0.07472526282072067, 0.026589034125208855, 0.062351420521736145, -0.06132795289158821, -0.04008616879582405, 0.05670977383852005, -0.0017044307896867394, 0.04548272490501404, 0.04243306443095207, -0.1522374153137207, 0.1377267837524414, 0.0029963895212858915, -0.012966182082891464, 0.07829928398132324, -0.03553002327680588, 0.03359532728791237, -0.012565266340970993, 0.005377695430070162, -0.11517225205898285, -0.09019150584936142, 0.05503324046730995, -0.06810547411441803, 0.11256185173988342, -0.1546492874622345, -0.08326210826635361, 0.0752047449350357, 0.060704126954078674, -0.09565968066453934, -0.09929632395505905, -0.14202256500720978, 0.017385154962539673, 0.12278297543525696, -0.1071714237332344, 0.07825323939323425, 0.016050094738602638, 0.06517037004232407, -0.01782546378672123, -0.09311836957931519, 0.04966816306114197, -0.08130183815956116, -0.1408497542142868, -0.03289985656738281, 0.04804685339331627, 0.09148316085338593, 0.05695085600018501, 0.0008928895113058388, -0.004327167756855488, -0.13343694806098938, -0.1138402447104454, -0.02067965641617775, -0.07430984824895859, 0.04067748039960861, 0.00735953263938427, 0.14606548845767975, -0.023390596732497215, -0.027962950989603996, -0.03726467862725258, 0.018859000876545906, 0.15426938235759735, -0.020494937896728516, 0.06142614036798477, 0.10346285253763199, -0.01656305231153965, -0.1961708515882492, -0.06612098962068558, 0.07671558111906052, 0.058731209486722946, -0.01581926830112934, -0.24506163597106934, 0.14566382765769958, -0.03047478385269642, -0.06630183011293411, -0.10555145889520645, -0.1243351548910141, -0.11188039928674698, 0.1660882979631424, 0.061639633029699326, 0.03662671521306038, -0.060831326991319656, -0.030545419082045555, -0.0506976954638958, -0.2815331816673279, 0.11472247540950775, -0.14378170669078827, 0.020969726145267487, -0.016537336632609367, 0.24545009434223175, -0.03310585394501686, -0.03217914327979088, 0.13923132419586182, 0.04136304557323456, -0.028080876916646957, 0.02483350783586502, -0.021729933097958565, 0.1346147358417511, 0.02965245395898819, 0.026246482506394386, -0.01011744700372219, 0.10914615541696548, -0.12198421359062195, -0.03388598561286926, -0.055142778903245926, 0.09230541437864304, -0.04428250715136528, -0.1580103188753128, -0.09869594871997833, 0.04357258975505829, -0.04078960791230202, -0.06397992372512817, 0.0952865481376648, 0.05367014929652214, -0.0018554205307736993, 0.08520596474409103, 0.15657450258731842, 0.0316547192633152, -0.07370823621749878, 0.02806319110095501, -0.012877551838755608, 0.053321223706007004, -0.14851540327072144, 0.04692631959915161, 0.10789567977190018, 0.023840777575969696, 0.021668724715709686, 0.04189283400774002, -0.08244433999061584, 0.007813611067831516, 0.08610867708921432, -0.15580862760543823, -0.1490936428308487, -0.06931020319461823, 0.1849566549062729, -0.21908582746982574, 0.05953822284936905, 0.1195761188864708, -0.006961534731090069, -0.07286947965621948, 0.027764473110437393, 0.04202686622738838, -0.020115816965699196, 0.17612296342849731, 0.08735030144453049, 0.13393867015838623, -0.12047337740659714, 0.054899442940950394, 0.02590302564203739, 0.019430948421359062, -0.03949898108839989, 0.008944759145379066, -0.09095964580774307, 0.01894523575901985, 0.02807627059519291, 0.1213250681757927, -0.1187916025519371, -0.0732543021440506, -0.16439427435398102, -0.12224873155355453, 0.009821798652410507, 0.06841965764760971, 0.07734338194131851, 0.08471181988716125, -0.027512723580002785, 0.024733703583478928, -0.0783219039440155, 0.11050520092248917, 0.006345832720398903, 0.02164766937494278, -0.12624379992485046, -0.0993051752448082, -0.03252348676323891, 0.1087200939655304, -0.051847200840711594, 0.006367829628288746, -0.0848679170012474, 0.024670248851180077, -0.14961861073970795, -0.07415363937616348, -0.018316470086574554, -0.024070195853710175, 0.02852979674935341, -0.08546231687068939, -0.04969851300120354, 0.064925417304039, -0.04292243719100952, 0.05373706668615341, 0.04413997754454613, 0.030450107529759407, -0.13071469962596893, 0.09775377810001373, 0.09788082540035248, -0.054397083818912506, 0.10934645682573318, 0.06480985879898071, 0.07545090466737747, 0.1183788925409317, -0.06476262956857681, 0.017197541892528534, -0.00214198837056756, 0.07684467732906342, 0.0778198167681694, -0.1600826531648636, -0.003456354374065995, -0.007541953586041927, 0.009232265874743462, 0.02948777750134468, 0.046346284449100494, -0.10132047533988953, 0.05172058194875717, -0.05806292966008186, 0.008782709017395973, -0.04756763577461243, 0.03623796999454498, 0.10589979588985443, 0.07136841118335724, 0.11215788871049881, -0.03676434978842735, -0.0007849882822483778, -0.12275377660989761, 0.01993493363261223, -0.0295563992112875, -0.021497420966625214, -0.03869178518652916, -0.039623335003852844, 0.060746580362319946, 0.024630147963762283, 0.13754071295261383, -0.027952438220381737, 0.04724271595478058, 0.022410286590456963, 0.1256483793258667, 0.0919988825917244, -0.007813755422830582, 0.07712715864181519, 0.09082873165607452, 0.08440791815519333, 0.1116088330745697, -0.0293380469083786, 0.05845208093523979, 0.10270775109529495, 0.11523442715406418, 0.07220937311649323, 0.06048489734530449, 0.03816492110490799, 0.056373775005340576, -0.10063428431749344, -0.08311809599399567, -0.0009988804813474417, -0.07863729447126389, 0.016538290306925774, 0.0037743474822491407, 0.19491833448410034, 0.09026239812374115, -0.1210288479924202, 0.07582925260066986, -0.014959041960537434, -0.10157857835292816, -0.08904625475406647, 0.09877566993236542, -0.044931117445230484, -0.22610598802566528, 0.034183721989393234, -0.15101729333400726, -0.11500193923711777, 0.17716310918331146, 0.002462969394400716, -0.0071482895873487, 0.14762242138385773, 0.1992340236902237, -0.034981418401002884, -0.044942937791347504, -0.00485585443675518, 0.00011230834934394807, 0.05185084789991379, 0.04054561257362366, -0.03617092967033386, 0.1086745336651802, 0.04356657713651657, 0.009037205018103123, -0.07625783979892731, 0.01611870713531971, -0.011402871459722519, -0.04089188948273659, -0.0204171109944582, -0.0008904829737730324, -0.08718091249465942, 0.17952392995357513, -0.014885940589010715, 0.09673020988702774, 0.016590163111686707, 0.3216872215270996, -0.056134164333343506, -0.14509572088718414, -0.144014373421669, 0.03839128836989403, -0.052942387759685516, 0.05955219268798828, 0.051378052681684494, -0.08552373945713043, -0.11166549474000931, 0.14794979989528656, 0.02223495952785015, -0.003228677436709404, -0.0506315603852272, 0.042117469012737274, -0.01576714776456356, -0.05866299569606781, 0.016237681731581688, 0.08140028268098831, 0.18153248727321625, -0.06632892787456512, -0.022752033546566963, -0.05764070898294449, -0.07642991095781326, 0.019244134426116943, 0.025347821414470673, 0.013561172410845757, -0.026154285296797752, -0.024061448872089386, 0.05259990319609642, -0.09112685173749924, -0.1204790323972702, -0.09909506142139435, -0.043197304010391235, -0.11671120673418045, -0.057160068303346634, 0.233273446559906, 0.008437753655016422, 0.003451408352702856, -0.03092990815639496, 0.07368253171443939, 0.04630301892757416, 0.04782703518867493, -0.09670370072126389, -0.07200469821691513, 0.1963815838098526, -0.14998310804367065, 0.15002793073654175, -0.05339144915342331, 0.062364861369132996, 0.11947312951087952, -0.02951081097126007, -0.121434286236763, 0.037112992256879807, 0.06370815634727478, -0.17864082753658295, -0.015290869399905205, 0.1388372927904129, 0.026657987385988235, 0.12402664870023727, -0.0015926461201161146, 0.09023138135671616, -0.025831809267401695, -0.008816936984658241, -0.028638651594519615, -0.04469230771064758, -0.01677251048386097, -0.07047327607870102, 0.047128044068813324, 0.07885589450597763, -0.0687258169054985, 0.07496923953294754, -0.08277799934148788, 0.09563351422548294, -0.029719386249780655, 0.07909122109413147, 0.023543020710349083, -0.13416196405887604, -0.06503204256296158, 0.14595918357372284, -0.06988386064767838, -0.11105107516050339, 0.005292181856930256, -0.0006155687151476741, 0.0056783780455589294, -0.09694325923919678, 0.14299124479293823, 0.07721637189388275, -0.008216661401093006, -0.014017103239893913, 0.029631922021508217, -0.056085243821144104, 0.06219560652971268, -0.16663935780525208, -0.0829014852643013 ]
null
null
transformers
# BatteryBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the [bert-base-cased](https://huggingface.co/bert-base-cased) weights. It was introduced in [this paper](paper_link) and first released in [this repository](https://github.com/ShuHuang/batterybert). This model is case-sensitive: it makes a difference between english and English. ## Model description BatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the [bert-base-cased](https://huggingface.co/bert-base-cased) weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the [bert-base-cased](https://huggingface.co/bert-base-cased) weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at [Github](https://github.com/ShuHuang/batterybert/blob/main/corpus.txt). ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 28,996. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=batterybert) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='batterydata/batterybert-cased') >>> unmasker("Hello I'm a <mask> model.") ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batterybert-cased') model = BertModel.from_pretrained('batterydata/batterybert-cased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batterybert-cased') model = TFBertModel.from_pretrained('batterydata/batterybert-cased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Evaluation results Final loss: 0.9609. ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": ["exbert"], "datasets": ["batterypapers"]}
fill-mask
batterydata/batterybert-cased
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "exbert", "en", "dataset:batterypapers", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-cased weights. It was introduced in this paper and first released in this repository. This model is case-sensitive: it makes a difference between english and English. ## Model description BatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-cased weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github. ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 28,996. The inputs of the model are then of the form: The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by '[MASK]'. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Evaluation results Final loss: 0.9609. ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-uncased model\r\n\r\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-cased weights. It was introduced in\r\nthis paper and first released in\r\nthis repository. This model is case-sensitive: it makes a difference between english and English.", "## Model description\r\n\r\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-cased weights. This means\r\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\r\npublicly available data) with an automatic process to generate inputs and labels from those texts. \r\n\r\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\r\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\r\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\r\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\r\nlearn a bidirectional representation of the sentence.\r\n\r\nThis way, the model learns an inner representation of the English language that can then be used to extract features\r\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\r\nclassifier using the features produced by the BERT model as inputs.", "## Training data\r\n\r\nThe BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\r\n\r\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 28,996. The inputs of the model are\r\nthen of the form:\r\n\r\n\r\n\r\nThe details of the masking procedure for each sentence are the following:\r\n- 15% of the tokens are masked.\r\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\r\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\r\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\r\n\r\n\r\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\r\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\r\n\r\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\r\nSee the model hub to look for fine-tuned versions on a task that\r\ninterests you.\r\n\r\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\r\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\r\ngeneration you should look at model like GPT2.", "### How to use\r\n\r\nYou can use this model directly with a pipeline for masked language modeling:\r\n\r\n\r\n\r\nHere is how to use this model to get the features of a given text in PyTorch:\r\n\r\n\r\n\r\nand in TensorFlow:", "## Evaluation results\r\n\r\nFinal loss: 0.9609.", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryBERT-uncased model\r\n\r\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-cased weights. It was introduced in\r\nthis paper and first released in\r\nthis repository. This model is case-sensitive: it makes a difference between english and English.", "## Model description\r\n\r\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-cased weights. This means\r\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\r\npublicly available data) with an automatic process to generate inputs and labels from those texts. \r\n\r\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\r\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\r\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\r\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\r\nlearn a bidirectional representation of the sentence.\r\n\r\nThis way, the model learns an inner representation of the English language that can then be used to extract features\r\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\r\nclassifier using the features produced by the BERT model as inputs.", "## Training data\r\n\r\nThe BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\r\n\r\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 28,996. The inputs of the model are\r\nthen of the form:\r\n\r\n\r\n\r\nThe details of the masking procedure for each sentence are the following:\r\n- 15% of the tokens are masked.\r\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\r\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\r\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\r\n\r\n\r\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\r\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\r\n\r\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\r\nSee the model hub to look for fine-tuned versions on a task that\r\ninterests you.\r\n\r\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\r\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\r\ngeneration you should look at model like GPT2.", "### How to use\r\n\r\nYou can use this model directly with a pipeline for masked language modeling:\r\n\r\n\r\n\r\nHere is how to use this model to get the features of a given text in PyTorch:\r\n\r\n\r\n\r\nand in TensorFlow:", "## Evaluation results\r\n\r\nFinal loss: 0.9609.", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 61, 82, 282, 96, 3, 135, 113, 128, 49, 11, 50 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryBERT-uncased model\r\n\r\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-cased weights. It was introduced in\r\nthis paper and first released in\r\nthis repository. This model is case-sensitive: it makes a difference between english and English.## Model description\r\n\r\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-cased weights. This means\r\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\r\npublicly available data) with an automatic process to generate inputs and labels from those texts. \r\n\r\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\r\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\r\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\r\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\r\nlearn a bidirectional representation of the sentence.\r\n\r\nThis way, the model learns an inner representation of the English language that can then be used to extract features\r\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\r\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.055391427129507065, 0.09050572663545609, -0.005469315219670534, 0.04962697997689247, 0.08549407869577408, 0.0047820089384913445, 0.14589671790599823, -0.003752774326130748, -0.026874249801039696, 0.039737846702337265, 0.03029802069067955, 0.0019992354791611433, 0.0649605393409729, 0.07026061415672302, 0.09658321738243103, -0.26704859733581543, 0.05833619460463524, -0.09446095675230026, 0.08288810402154922, 0.06383085995912552, 0.07149132341146469, -0.09972473978996277, 0.021886220201849937, 0.055596284568309784, 0.0523337796330452, -0.007386418990790844, -0.0169509407132864, -0.09314937889575958, 0.09157712757587433, 0.039142757654190063, 0.07783614844083786, 0.021785665303468704, 0.03991491720080376, -0.029968474060297012, 0.026490185409784317, 0.032259806990623474, 0.015349474735558033, 0.07298078387975693, 0.05321718379855156, -0.005947045516222715, 0.07568895071744919, 0.05476292595267296, 0.08164455741643906, 0.03411953151226044, -0.07994698733091354, -0.15087801218032837, -0.010803391225636005, -0.0007314266404137015, -0.017456183210015297, 0.09757795184850693, -0.02692185342311859, 0.10327215492725372, -0.030606700107455254, 0.05956219509243965, 0.1882375031709671, -0.10082835704088211, -0.008400077000260353, 0.013117587193846703, 0.011729378253221512, -0.003196735167875886, -0.0154267568141222, -0.043332893401384354, -0.021855143830180168, 0.03527853265404701, 0.088079534471035, 0.008391921408474445, 0.05118780955672264, -0.037232547998428345, -0.15990813076496124, -0.02155381627380848, 0.11843752861022949, -0.038933902978897095, -0.0983375608921051, -0.13306857645511627, -0.08443677425384521, 0.12694300711154938, -0.0010181404650211334, 0.04052334278821945, -0.030198339372873306, 0.015283661894500256, 0.021790949627757072, -0.04490570351481438, -0.05847792699933052, -0.08064360171556473, -0.09757015854120255, 0.18458932638168335, 0.032322581857442856, 0.0483367033302784, -0.05914352461695671, 0.07507152855396271, -0.10051467269659042, -0.07532329857349396, -0.054508671164512634, -0.04463556781411171, -0.058674611151218414, 0.006439610384404659, -0.017704972997307777, -0.22330713272094727, -0.027974829077720642, 0.10814723372459412, 0.0043951598927378654, -0.0038364436477422714, -0.018076373264193535, 0.013344946317374706, 0.12257445603609085, 0.06544538587331772, -0.11833938211202621, 0.11629794538021088, 0.007456011138856411, -0.031834524124860764, 0.02190212905406952, -0.05701408535242081, -0.01869267039000988, 0.019319454208016396, -0.018066195771098137, 0.007154147606343031, 0.03750098496675491, 0.06531745195388794, -0.025933396071195602, -0.05555391684174538, 0.0957537367939949, -0.11395128816366196, -0.05157754570245743, 0.013287881389260292, -0.05535797402262688, 0.0012151839910075068, 0.05544217303395271, -0.07749977707862854, -0.061043452471494675, 0.04859793186187744, -0.07261373102664948, -0.09088883548974991, -0.11988097429275513, -0.16191725432872772, -0.041219063103199005, 0.027719661593437195, -0.05889048054814339, -0.08084841817617416, -0.27142614126205444, -0.02070467919111252, 0.030408233404159546, 0.009075131267309189, 0.0015655035385861993, -0.030447080731391907, 0.008197898976504803, -0.006594375241547823, -0.006180030293762684, -0.014490307308733463, -0.014538571238517761, 0.007628452964127064, -0.03184263035655022, 0.08134705573320389, 0.030833523720502853, 0.009615596383810043, -0.07456079870462418, -0.004275121726095676, -0.20682990550994873, 0.1061224490404129, -0.08311132341623306, -0.038664404302835464, -0.026797857135534286, -0.01734577678143978, 0.021488690748810768, 0.00694306381046772, 0.0374014750123024, 0.12660443782806396, -0.11885163187980652, -0.09520713239908218, 0.2089562714099884, -0.17417216300964355, 0.05691448599100113, 0.09218481928110123, -0.07160940766334534, 0.06544021517038345, 0.11507117003202438, -0.0184489618986845, 0.13662318885326385, -0.11945246905088425, -0.05686569958925247, 0.06479937583208084, -0.0565885454416275, 0.0869070515036583, 0.09843431413173676, -0.0006731805624440312, -0.1468827873468399, 0.04353677108883858, 0.018221402540802956, 0.03933718055486679, -0.027691787108778954, -0.017408328130841255, 0.04288412258028984, -0.021046895533800125, 0.08971770852804184, -0.010735313408076763, 0.005080737639218569, 0.019980616867542267, -0.09116730839014053, 0.10612968355417252, 0.1220642477273941, -0.05884014815092087, -0.027096180245280266, -0.12625223398208618, 0.06526640057563782, -0.03872211277484894, -0.030550789088010788, -0.16661185026168823, -0.01260379608720541, 0.05048099905252457, -0.1420951783657074, 0.09087483584880829, 0.019956834614276886, 0.0378485806286335, 0.13042280077934265, -0.035803426057100296, 0.018244052305817604, -0.030207065865397453, -0.027145013213157654, -0.013037222437560558, -0.13545648753643036, 0.0013964944519102573, -0.04593101143836975, 0.022008683532476425, 0.042365822941064835, 0.039001625031232834, -0.042972858995199203, -0.04037519916892052, 0.006759320385754108, -0.05768340826034546, 0.04818601533770561, -0.029941346496343613, -0.015492888167500496, -0.0022243880666792393, -0.02946736291050911, 0.03840062394738197, -0.05791641026735306, 0.02553391270339489, -0.15627695620059967, -0.10850077867507935, 0.05918717384338379, 0.07373301684856415, -0.0950048416852951, 0.05302150920033455, -0.027272483333945274, 0.006289685610681772, -0.12228231877088547, -0.03853149339556694, 0.23238694667816162, 0.011906985193490982, 0.05561260133981705, -0.08454879373311996, 0.03756529465317726, 0.03587725758552551, -0.024125266820192337, -0.003008487867191434, 0.06846302002668381, 0.058421432971954346, -0.05133691802620888, 0.012927527539432049, -0.11622199416160583, -0.011716167442500591, 0.16805650293827057, 0.040925126522779465, -0.0467730276286602, -0.06433836370706558, -0.04223224148154259, 0.022728675976395607, 0.05332065001130104, -0.07547836005687714, 0.04185975715517998, 0.055656395852565765, -0.006052608136087656, 0.027286382392048836, -0.0348067581653595, 0.046086668968200684, 0.058988507837057114, 0.010980082675814629, -0.07785025238990784, -0.07588919997215271, -0.01698504574596882, 0.07374757528305054, 0.058439891785383224, -0.021859727799892426, -0.00543128652498126, -0.02937757410109043, -0.09990973025560379, 0.17211520671844482, -0.0989927276968956, -0.2409912645816803, -0.158975750207901, -0.03192446008324623, -0.010758783668279648, 0.03396022692322731, -0.012311374768614769, -0.008096456527709961, -0.08461353182792664, -0.08175862580537796, 0.14044322073459625, -0.025136327371001244, 0.017421705648303032, -0.005340058822184801, -0.032642994076013565, -0.04274769127368927, -0.11921220272779465, 0.027796851471066475, -0.026479730382561684, -0.08898911625146866, 0.016200248152017593, -0.03901843726634979, 0.043106477707624435, 0.10285316407680511, 0.018554070964455605, -0.04429828003048897, -0.0368921123445034, 0.19541756808757782, -0.012462125159800053, 0.06697359681129456, 0.15220260620117188, 0.006584447808563709, 0.049709469079971313, 0.05794639140367508, 0.02053188718855381, -0.015321474522352219, 0.06193939968943596, -0.011737404391169548, -0.0897696241736412, -0.17096266150474548, -0.08801399171352386, -0.03323662653565407, 0.03816939890384674, 0.06864220649003983, 0.0072359307669103146, 0.01732669584453106, 0.07710562646389008, -0.03284749388694763, -0.031165242195129395, 0.06870424002408981, 0.06503532081842422, -0.009446260519325733, -0.03246929123997688, 0.09431938081979752, -0.03583713248372078, -0.029906583949923515, 0.11986477673053741, -0.14033924043178558, 0.14671854674816132, -0.08788648247718811, 0.1230604276061058, 0.054948847740888596, 0.030293989926576614, 0.08450368046760559, 0.15912555158138275, -0.06987851858139038, 0.05036139115691185, -0.04653065651655197, -0.07870565354824066, -0.0117096658796072, 0.036925703287124634, 0.010877544060349464, 0.02211185358464718, -0.04505060240626335, -0.010886591859161854, -0.0026152976788580418, 0.1370769888162613, 0.0174662284553051, -0.14856688678264618, -0.13037823140621185, -0.004256784450262785, -0.007154576946049929, -0.08163896948099136, 0.029014170169830322, 0.08872370421886444, -0.09271799027919769, 0.07518913596868515, 0.001805676962248981, 0.05699029937386513, -0.0850067138671875, 0.017114341259002686, -0.09370595961809158, 0.10790558904409409, -0.05791327729821205, 0.11926168203353882, -0.1716027557849884, 0.037951819598674774, 0.023475132882595062, 0.08544870465993881, -0.12029702961444855, 0.027315547689795494, 0.005161425098776817, -0.07103104144334793, 0.13222530484199524, 0.005022626835852861, -0.012703534215688705, -0.00134926731698215, -0.11777059733867645, 0.008085839450359344, 0.021104449406266212, -0.061157967895269394, 0.054376307874917984, 0.034219689667224884, 0.015807339921593666, 0.010386102832853794, 0.021952642127871513, -0.051169976592063904, -0.1200636550784111, 0.0635814517736435, -0.0402679368853569, -0.0643293485045433, -0.08021406829357147, -0.08235859870910645, -0.042645495384931564, 0.11771363019943237, -0.13466070592403412, -0.10035156458616257, -0.1183471605181694, -0.0074888779781758785, 0.11125408113002777, -0.08113806694746017, 0.031794704496860504, -0.03304629400372505, 0.1602429747581482, -0.08788350969552994, -0.11207928508520126, 0.0637281984090805, -0.06459327787160873, -0.14444948732852936, -0.042512908577919006, 0.026744602248072624, 0.21765641868114471, 0.05887049436569214, -0.03717799112200737, 0.05117747187614441, -0.04129042103886604, -0.11769605427980423, -0.027219504117965698, 0.14296340942382812, -0.01597013883292675, 0.10260625183582306, -0.00328839966095984, -0.1142716184258461, -0.013222363777458668, 0.014718228951096535, 0.07260438054800034, 0.15187419950962067, -0.04680030420422554, 0.15268577635288239, 0.28629207611083984, -0.06826506555080414, -0.24575206637382507, -0.045149363577365875, 0.06759081780910492, 0.060793839395046234, -0.003875655820593238, -0.29886433482170105, 0.06944484263658524, 0.010474681854248047, -0.04185103625059128, -0.10082156211137772, -0.18716014921665192, -0.13341732323169708, 0.16716688871383667, 0.038924187421798706, 0.1350041627883911, 0.012283079326152802, -0.013235189020633698, -0.03745618835091591, 0.06771045923233032, 0.17183418571949005, -0.0537688173353672, 0.08209824562072754, 0.025028619915246964, 0.01454953569918871, 0.041543148458004, -0.07300128787755966, 0.07365961372852325, -0.01885273866355419, 0.04082968458533287, -0.05045953020453453, -0.0010563755640760064, 0.12092878669500351, -0.010496307164430618, 0.11370544135570526, 0.10567930340766907, 0.0875626727938652, -0.019591400399804115, -0.12302915751934052, -0.06519273668527603, 0.09302998334169388, -0.005735427141189575, -0.10142343491315842, -0.09910127520561218, 0.08069479465484619, 0.04422331601381302, -0.022452443838119507, -0.004539355169981718, -0.026245085522532463, 0.054778486490249634, 0.14217165112495422, 0.1294817477464676, 0.04013044387102127, -0.05242794007062912, 0.019209392368793488, -0.014020988717675209, 0.08670118451118469, -0.002571985824033618, 0.03141673281788826, 0.0664396584033966, 0.05918816477060318, 0.0828602984547615, 0.05330939590930939, -0.18773676455020905, 0.003055968089029193, 0.015873759984970093, -0.16229583323001862, -0.1609853059053421, -0.03192511200904846, 0.11019670218229294, -0.12890267372131348, -0.025464244186878204, 0.11626634746789932, -0.03172513097524643, -0.021295683458447456, -0.01633983664214611, 0.032184209674596786, 0.031218623742461205, 0.1393904983997345, 0.023930998519062996, 0.04840405657887459, -0.07463623583316803, 0.06427977979183197, 0.057927269488573074, 0.005067833699285984, 0.03390013799071312, 0.0554141141474247, -0.08757595717906952, -0.02278432622551918, -0.016353828832507133, 0.15040352940559387, -0.006260219495743513, -0.033853646367788315, -0.046450186520814896, -0.09057356417179108, 0.07123427838087082, 0.14450863003730774, 0.02410917356610298, 0.040235184133052826, -0.08782432228326797, 0.03869619965553284, -0.047493208199739456, 0.04316865652799606, -0.033422306180000305, -0.03725915402173996, 0.035471703857183456, 0.10365933924913406, 0.023417074233293533, 0.024718446657061577, -0.06477165222167969, -0.09855180233716965, -0.0928078442811966, 0.003986907657235861, -0.020558806136250496, 0.029309086501598358, -0.08090455085039139, -0.009447772987186909, -0.014290978200733662, 0.03654257208108902, 0.04633302614092827, 0.0070027816109359264, -0.03380855172872543, 0.02684490941464901, -0.04028773307800293, -0.002584561938419938, -0.11756625771522522, 0.0342755950987339, 0.04651990532875061, -0.038811419159173965, 0.10303553193807602, 0.04442070797085762, -0.022638706490397453, 0.014037497341632843, -0.08439361304044724, 0.014123807661235332, -0.03859812766313553, 0.05540641397237778, -0.025243373587727547, -0.08357708901166916, 0.024457430467009544, -0.01862286403775215, -0.029085004702210426, -0.01719805784523487, 0.005359676666557789, -0.0693327784538269, 0.14392180740833282, 0.11428671330213547, 0.0020565118174999952, -0.09705264866352081, 0.04628969356417656, 0.004799689166247845, 0.027153491973876953, 0.05298905074596405, -0.0333443358540535, 0.0526861771941185, -0.09721025079488754, -0.007228665053844452, 0.008908349089324474, 0.0029599934350699186, 0.02760816551744938, -0.031351812183856964, 0.06743092834949493, 0.002075391821563244, 0.16209663450717926, 0.04600819572806358, -0.06615156680345535, -0.009894369170069695, 0.06269393116235733, 0.058510880917310715, -0.010900521650910378, 0.020148541778326035, 0.03424358740448952, -0.01561792939901352, 0.059347160160541534, 0.02018025889992714, -0.02362685464322567, 0.07517670840024948, 0.11503452062606812, -0.01697309873998165, 0.16155019402503967, -0.04084214195609093, 0.048355795443058014, 0.0027435317169874907, -0.06737449020147324, -0.04749467596411705, -0.004580949433147907, 0.04800385236740112, -0.004355132579803467, 0.022494234144687653, 0.04448125883936882, -0.04965030029416084, 0.1559126079082489, 0.06958822906017303, -0.05547940358519554, -0.12715689837932587, -0.22919991612434387, -0.011539552360773087, 0.04336749389767647, 0.0057928357273340225, -0.1221604272723198, -0.04761191084980965, 0.05208740383386612, 0.029390599578619003, 0.005243917927145958, 0.0685826763510704, -0.08284302055835724, -0.052896756678819656, 0.03077547997236252, -0.030929701402783394, -0.013636141084134579, -0.0043008411303162575, -0.008103551343083382, 0.05473901703953743, 0.13267405331134796, 0.03143464773893356, 0.05539427325129509, 0.03772996366024017, 0.025042474269866943, 0.0013368650106713176, -0.07380238175392151, -0.01301605999469757, -0.042878199368715286, -0.028868885710835457, 0.13557866215705872, -0.0057166810147464275, -0.03638720512390137, -0.006870925892144442, 0.13678529858589172, -0.04299592226743698, -0.08564411103725433, -0.11468332260847092, 0.2363787591457367, -0.0674295648932457, 0.022483430802822113, 0.0491187684237957, -0.06652772426605225, -0.046859607100486755, 0.21548517048358917, 0.16604122519493103, 0.023957883939146996, 0.0018380312249064445, 0.020452020689845085, -0.009896695613861084, 0.015283667482435703, 0.09877533465623856, -0.015384132973849773, 0.35976284742355347, -0.03598666936159134, 0.17160283029079437, -0.03655346855521202, -0.003051962936297059, -0.008831608109176159, 0.030181601643562317, -0.01356573961675167, 0.03746049106121063, -0.06460534781217575, 0.0611284039914608, -0.1164746806025505, -0.2741925120353699, -0.049054186791181564, 0.0461917519569397, -0.06950663775205612, -0.00812467746436596, 0.006120762787759304, 0.04784156382083893, 0.07451951503753662, -0.0030125596094876528, 0.057390861213207245, 0.03017103485763073, 0.02791275642812252, -0.11578424274921417, -0.07031851261854172, 0.08548642694950104, -0.0146557642146945, 0.18109801411628723, 0.019038617610931396, 0.031000936403870583, 0.09561991691589355, -0.03490208834409714, -0.10967543721199036, 0.07677435129880905, -0.008350781165063381, -0.08596828579902649, 0.003567721927538514, 0.130524680018425, -0.02877998538315296, 0.1284164935350418, 0.004499577451497316, -0.03273644298315048, 0.04344039782881737, -0.06663267314434052, -0.060271333903074265, -0.08365316689014435, 0.07303790003061295, -0.07108812779188156, 0.10810135304927826, 0.11032398790121078, -0.017565958201885223, -0.03830549493432045, -0.03762573003768921, 0.037935204803943634, -0.08841890096664429, 0.00019594126206357032, -0.03148941695690155, -0.13130131363868713, -0.05524709075689316, -0.021087097004055977, 0.05615117400884628, -0.27503806352615356, -0.007763793226331472, 0.008002863265573978, -0.05236395075917244, 0.02028181217610836, 0.01773218996822834, -0.012755981646478176, 0.03773616626858711, -0.02173706702888012, 0.023529263213276863, -0.01811157912015915, 0.039036985486745834, -0.13437625765800476, -0.06482874602079391 ]
null
null
transformers
# BatteryBERT-uncased for Battery Abstract Classification **Language model:** batterybert-uncased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 11 base_LM_model = "batterybert-uncased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 97.10, "Test accuracy": 96.94, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batterybert-uncased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batterybert-uncased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryBERT-uncased for Battery Abstract Classification Language model: batterybert-uncased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-uncased for Battery Abstract Classification \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryBERT-uncased for Battery Abstract Classification \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 71, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryBERT-uncased for Battery Abstract Classification \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.06678469479084015, 0.06125020235776901, -0.003756333375349641, 0.05189472436904907, 0.2590707540512085, 0.029930997639894485, 0.1844843327999115, 0.033447377383708954, -0.09817646443843842, -0.036899976432323456, 0.04206230491399765, 0.040243640542030334, 0.049198321998119354, 0.0659877136349678, -0.04858899861574173, -0.16907064616680145, 0.03671753779053688, -0.00459680613130331, 0.0334855280816555, 0.11696502566337585, 0.13804613053798676, -0.08840112388134003, 0.11608561873435974, 0.010011457838118076, -0.10004810988903046, 0.02469039335846901, 0.02264196053147316, -0.1392793506383896, 0.12721185386180878, -0.05916463956236839, 0.1563534140586853, 0.04089329391717911, 0.035778697580099106, -0.15169364213943481, 0.030458131805062294, 0.01595950499176979, 0.001306156744249165, 0.0707654133439064, 0.026306811720132828, -0.09133203327655792, 0.09713561832904816, -0.01580626145005226, 0.07457257062196732, 0.026601390913128853, -0.09023796766996384, -0.1668490618467331, -0.032025061547756195, 0.018926439806818962, 0.09237794578075409, 0.12133613228797913, -0.011906396597623825, 0.0768391489982605, -0.04461991786956787, 0.06992150098085403, 0.20922353863716125, -0.11712729185819626, 0.013072780333459377, 0.10815881937742233, 0.009550895541906357, -0.032466206699609756, -0.057464126497507095, -0.02840948849916458, 0.01207408495247364, 0.04669320583343506, 0.06828958541154861, -0.029934635385870934, -0.16335009038448334, 0.10098663717508316, -0.09982188045978546, 0.008802663534879684, 0.23199385404586792, -0.07017980515956879, -0.03729483485221863, -0.038907490670681, -0.094101682305336, -0.07336930185556412, -0.044009242206811905, 0.017244067043066025, -0.032032325863838196, 0.017818115651607513, -0.02212083712220192, 0.05510212853550911, -0.046282924711704254, -0.06501321494579315, -0.11982141435146332, 0.1068367213010788, 0.007849745452404022, 0.005421074107289314, -0.1037212535738945, 0.1167479082942009, 0.05772039666771889, -0.15916064381599426, 0.012647047638893127, -0.06267042458057404, -0.01893186755478382, -0.03275847062468529, 0.01491233054548502, 0.0005801229854114354, 0.08778811991214752, 0.10049056261777878, -0.017048416659235954, 0.0015399287221953273, 0.007725722622126341, -0.015541321597993374, 0.0005179482977837324, 0.13852915167808533, -0.03577417880296707, -0.05677361041307449, 0.032116785645484924, -0.0025204254779964685, 0.031308867037296295, -0.0384046770632267, -0.12717664241790771, -0.009722577407956123, 0.08200918883085251, 0.06351031363010406, -0.0314280241727829, 0.06496476382017136, -0.07587940990924835, -0.06269232928752899, 0.078933946788311, -0.09954468905925751, -0.020792335271835327, 0.007747492752969265, 0.035135481506586075, 0.05367529019713402, -0.02009836956858635, 0.024084482342004776, -0.02446514554321766, 0.07288320362567902, -0.08652900904417038, -0.05115966126322746, -0.10137680917978287, -0.09531748294830322, -0.010292432270944118, 0.07824614644050598, 0.0683230385184288, -0.12485180795192719, -0.2721388041973114, 0.07345369458198547, 0.0016165809938684106, -0.0011376857291907072, -0.06281311064958572, -0.06410828232765198, -0.07125267386436462, 0.015282326377928257, -0.009707720950245857, 0.040143273770809174, -0.07319923490285873, 0.046161673963069916, 0.10553395748138428, 0.03216614946722984, -0.06548662483692169, 0.04058777913451195, -0.13271111249923706, -0.009174986742436886, -0.19168560206890106, 0.018573110923171043, -0.10115642100572586, 0.0847032442688942, -0.07095689326524734, -0.05826452001929283, 0.1371564269065857, -0.024107975885272026, 0.02287885919213295, 0.14575079083442688, -0.07251349836587906, -0.14901123940944672, 0.03821674361824989, -0.13348224759101868, -0.09739923477172852, 0.0459350161254406, -0.0381716713309288, 0.018002036958932877, 0.06333027780056, 0.044092658907175064, 0.06514418125152588, -0.1236237958073616, -0.11272954940795898, 0.013857117854058743, 0.010431025177240372, -0.005630322732031345, 0.09379220753908157, 0.04810049757361412, -0.06212474778294563, 0.06800315529108047, -0.031724244356155396, 0.06362608075141907, -0.041099533438682556, -0.09188303351402283, -0.02391832321882248, -0.025087498128414154, 0.13645151257514954, 0.02565261535346508, 0.0064838966354727745, -0.0011967557948082685, -0.09000475704669952, 0.15828837454319, 0.09302416443824768, -0.04718705266714096, 0.005533413030207157, -0.14361554384231567, 0.08760818839073181, 0.036767590790987015, 0.010978414677083492, -0.10808011889457703, -0.1028338149189949, 0.03980884701013565, 0.06444457918405533, -0.04412972927093506, -0.07710248976945877, 0.005149762146174908, 0.03581712022423744, -0.10996896773576736, -0.01247668731957674, -0.07984788715839386, 0.00517765199765563, 0.00018504963372834027, -0.22883380949497223, 0.06860040873289108, -0.02962586283683777, 0.13589085638523102, -0.09901633858680725, 0.05122201889753342, 0.14031629264354706, 0.07409752905368805, -0.02122170478105545, 0.005416102707386017, -0.012179816141724586, 0.040001098066568375, -0.07716535776853561, -0.04460279643535614, 0.004941132385283709, -0.001841540914028883, -0.055335693061351776, 0.012486959807574749, -0.15036541223526, 0.11720705777406693, 0.17945560812950134, -0.015765516087412834, -0.04269276559352875, -0.03863048553466797, 0.0027901886496692896, -0.02010602504014969, -0.09641440957784653, -0.0007401816546916962, 0.23004892468452454, -0.02211577072739601, 0.07815069705247879, -0.03392677381634712, -0.011612039059400558, -0.0005620798328891397, 0.0352601557970047, 0.04471480846405029, 0.11151330918073654, 0.11057400703430176, -0.03606913238763809, 0.1048404723405838, 0.04135624319314957, -0.07126466929912567, 0.14944325387477875, -0.0654357373714447, -0.01452695857733488, 0.0005262125050649047, -0.10600150376558304, -0.0014368427218869328, 0.21115794777870178, -0.18707455694675446, 0.01628628745675087, 0.0627758651971817, -0.033959366381168365, 0.013177243992686272, -0.07775618135929108, -0.018603768199682236, -0.01132552232593298, -0.015584548935294151, -0.10441560298204422, 0.010552629828453064, 0.04728106036782265, 0.05866052210330963, -0.005810541566461325, -0.0964888408780098, 0.07355080544948578, 0.019037684425711632, -0.10881206393241882, 0.23176169395446777, -0.12341716885566711, -0.24085280299186707, -0.1250457614660263, -0.09015841782093048, -0.020080434158444405, -0.008406316861510277, 0.023785866796970367, -0.11194276064634323, -0.10954468697309494, 0.003982033580541611, 0.11381565034389496, -0.02256481535732746, 0.08105973154306412, 0.08161178976297379, 0.02355865389108658, -0.02043675258755684, -0.10486363619565964, 0.003779478371143341, 0.04386896267533302, -0.04806151241064072, 0.0401875376701355, 0.0447196364402771, 0.0384463332593441, 0.183669313788414, 0.0068895346485078335, 0.013589730486273766, -0.05487490072846413, 0.14751742780208588, -0.10423699021339417, -0.042369868606328964, 0.21222124993801117, 0.04157024621963501, -0.005315906368196011, 0.1693563610315323, -0.0011587778571993113, -0.022174235433340073, 0.07550236582756042, -0.02380208484828472, -0.05380527675151825, -0.2618788480758667, -0.13220994174480438, -0.06112215295433998, 0.10432477295398712, -0.03954749554395676, 0.03056326135993004, 0.030331594869494438, 0.057052381336688995, -0.00604660902172327, 0.07062460482120514, -0.03496577963232994, 0.1325589418411255, 0.2347661703824997, 0.06129522621631622, 0.09379183501005173, -0.01675551012158394, -0.03507905825972557, 0.06382953375577927, -0.05334550887346268, 0.12990620732307434, -0.0035442463122308254, 0.06176183372735977, 0.005894436500966549, 0.020289339125156403, 0.06413047015666962, 0.04608896002173424, 0.0025086388923227787, -0.013512810692191124, 0.022682197391986847, -0.042517367750406265, 0.05228535830974579, -0.01315292064100504, -0.10364917665719986, -0.01212941762059927, -0.04951150715351105, 0.0839749276638031, 0.08012501895427704, -0.046899136155843735, 0.03003428503870964, -0.23969712853431702, -0.06242915242910385, 0.050151385366916656, -0.05509209632873535, -0.10056042671203613, 0.029976028949022293, -0.02706618793308735, -0.1271217167377472, 0.07947945594787598, -0.016491740942001343, 0.07493545860052109, -0.10644835978746414, 0.06141621246933937, 0.03914741799235344, 0.06561041623353958, -0.00642621424049139, 0.12198232859373093, -0.3546450734138489, 0.13082264363765717, 0.047457002103328705, 0.06054095923900604, -0.022327877581119537, 0.03218875080347061, -0.022278646007180214, 0.056026481091976166, 0.02585013583302498, -0.0009746128926053643, 0.049073196947574615, 0.015736868605017662, -0.12435907870531082, 0.05533786118030548, 0.017625248059630394, -0.03242511674761772, 0.03375954180955887, -0.021698780357837677, 0.013153727166354656, 0.03714112937450409, -0.07041428238153458, -0.12159119546413422, -0.11784880608320236, 0.08392277359962463, -0.02526424080133438, 0.12668612599372864, -0.09769396483898163, -0.10891589522361755, -0.013290096074342728, 0.1319848597049713, -0.12689149379730225, -0.11140204221010208, -0.14650267362594604, 0.011267884634435177, 0.0487438403069973, -0.07379715889692307, 0.077765554189682, 0.0009356244117952883, 0.08572782576084137, -0.03744970262050629, -0.11483708769083023, 0.1112411767244339, -0.07309646159410477, -0.18732747435569763, -0.0472056120634079, 0.09601270407438278, 0.10226552933454514, 0.05821699649095535, -0.021192507818341255, -0.0062507810071110725, -0.08870705962181091, -0.12975698709487915, -0.006687909364700317, -0.0019004523055627942, 0.0714334100484848, 0.02486296556890011, 0.07089819014072418, -0.09396369755268097, -0.025890884920954704, -0.0656680092215538, 0.07398847490549088, 0.18079212307929993, -0.0278043020516634, 0.10737203061580658, 0.07387445867061615, -0.048431891947984695, -0.2426794171333313, -0.04583870619535446, 0.013067738153040409, 0.03314674645662308, -0.042759284377098083, -0.2594587206840515, 0.14173024892807007, -0.04043594375252724, -0.0658600777387619, -0.0336991511285305, -0.12372247874736786, -0.13717800378799438, 0.21086741983890533, 0.02049805037677288, 0.12322773039340973, -0.08314158022403717, -0.02103765867650509, -0.09267424046993256, -0.19986563920974731, 0.21728459000587463, -0.148204043507576, 0.04794364795088768, -0.016705222427845, 0.1843140721321106, -0.013101167045533657, -0.03718232363462448, 0.12835213541984558, 0.037234313786029816, 0.008035353384912014, -0.011293938383460045, -0.09168899059295654, 0.21861031651496887, 0.04038289561867714, 0.005880861543118954, -0.043414004147052765, 0.05241410434246063, -0.09530246257781982, -0.052157364785671234, -0.02621733769774437, 0.11848670989274979, -0.03107999823987484, -0.15496625006198883, -0.10286291688680649, 0.024080868810415268, -0.023828895762562752, -0.04905524477362633, 0.20837806165218353, 0.07399659603834152, 0.027537716552615166, 0.02880425937473774, 0.2182164490222931, -0.012662981636822224, 0.05059671774506569, 0.009715761989355087, -0.025722810998558998, 0.0604274719953537, -0.1651124209165573, 0.033773455768823624, 0.09386558085680008, 0.0388810820877552, 0.03660453110933304, 0.10108748823404312, -0.055615101009607315, -0.025141943246126175, 0.09788140654563904, -0.22300301492214203, -0.030466504395008087, -0.09983367472887039, 0.14403554797172546, -0.13301433622837067, 0.08597000688314438, 0.11147196590900421, -0.038064636290073395, -0.06424621492624283, 0.049016859382390976, 0.03250863403081894, -0.033215541392564774, 0.12053567916154861, 0.0501285158097744, 0.10935929417610168, -0.1389063000679016, 0.0574871301651001, 0.06764053553342819, 0.03437396511435509, -0.02246135286986828, 0.07001908123493195, -0.13927234709262848, -0.012125824578106403, 0.02374575287103653, 0.1863114833831787, -0.1265190988779068, -0.048298709094524384, -0.18970809876918793, -0.1486063152551651, 0.028567839413881302, 0.1330489069223404, 0.10811937600374222, 0.09022976458072662, -0.047991055995225906, 0.012868191115558147, -0.09913253784179688, 0.08877499401569366, 0.004086649045348167, 0.023350613191723824, -0.15475884079933167, 0.02311728149652481, -0.014801562763750553, 0.04576095566153526, -0.06461073458194733, 0.0031449422240257263, -0.15401676297187805, 0.01704397425055504, -0.08520637452602386, -0.061787091195583344, -0.035141583532094955, -0.0221212487667799, 0.011792470701038837, -0.039568882435560226, -0.01937374286353588, 0.041806407272815704, -0.043691493570804596, 0.05817818269133568, 0.030181940644979477, 0.05494141951203346, -0.10217393189668655, 0.06547006964683533, 0.06525438278913498, -0.06029442697763443, 0.12817324697971344, 0.09435556828975677, 0.07714437693357468, 0.09917529672384262, -0.16891290247440338, 0.014494800008833408, 0.05305369570851326, 0.04736955463886261, 0.05111837387084961, -0.12541180849075317, -0.007850215770304203, -0.00280403601936996, 0.02451247349381447, 0.04956476390361786, 0.05845135450363159, -0.10222853720188141, 0.05929737538099289, -0.04717133939266205, -0.008173009380698204, -0.05102737620472908, 0.024752173572778702, 0.1249765008687973, 0.03540429845452309, 0.11874572187662125, -0.0653967335820198, 0.00034382136072963476, -0.07658832520246506, 0.029295925050973892, -0.030389832332730293, -0.06560242921113968, -0.0859934538602829, -0.034055158495903015, 0.05416587367653847, 0.009465481154620647, 0.20793591439723969, 0.01967485062777996, -0.016445113345980644, 0.05448891967535019, 0.14947296679019928, 0.06849805265665054, 0.03554977476596832, 0.12372785061597824, 0.10423147678375244, 0.051085811108350754, 0.07898871600627899, -0.0009815628873184323, 0.13015758991241455, 0.09615661203861237, 0.1511172205209732, 0.024141725152730942, -0.02439763769507408, 0.035536494106054306, 0.01010204292833805, -0.1108509749174118, -0.1672830879688263, 0.019895462319254875, -0.06710320711135864, 0.0429486446082592, -0.04204298555850983, 0.18267430365085602, 0.11608611047267914, -0.09219998866319656, 0.05421429127454758, -0.027393925935029984, -0.0997845008969307, -0.1063569039106369, -0.00039481415296904743, -0.03909897431731224, -0.19228173792362213, 0.034538496285676956, -0.1342100203037262, -0.08711022883653641, 0.1262275129556656, -0.005139065440744162, -0.033865299075841904, 0.08501962572336197, 0.17309443652629852, -0.020213481038808823, 0.027720924466848373, -0.010344339534640312, -0.024008333683013916, -0.016842538490891457, -0.0010238474933430552, -0.07795222848653793, 0.0612126961350441, 0.007051529362797737, -0.015196891501545906, -0.03105901926755905, 0.012776672840118408, -0.00749788386747241, -0.06213890016078949, 0.01192648708820343, -0.001636965898796916, -0.07870007306337357, 0.17709806561470032, -0.015462001785635948, 0.06327852606773376, 0.0445692203938961, 0.2835147976875305, -0.06388521939516068, -0.15383769571781158, -0.09725861251354218, 0.12374186515808105, -0.011849314905703068, 0.03867686539888382, 0.0530422106385231, -0.09328432381153107, -0.0978483334183693, 0.17484839260578156, 0.10375960171222687, 0.004727010149508715, -0.01078533660620451, -0.008598336949944496, -0.013297916390001774, -0.004503740929067135, 0.019577927887439728, 0.07537618279457092, 0.1725028157234192, -0.06810899078845978, -0.06108403205871582, -0.048097338527441025, -0.046090319752693176, -0.06052590161561966, 0.05533863604068756, 0.03889405354857445, -0.04034209996461868, -0.018942518159747124, 0.04804360121488571, -0.17073732614517212, -0.10945667326450348, -0.0989794209599495, -0.10119412839412689, -0.0895688459277153, -0.06156056746840477, 0.15197934210300446, -0.004777140915393829, 0.011448130011558533, -0.011415054090321064, 0.036030903458595276, 0.02401263825595379, 0.04909684136509895, -0.14243288338184357, 0.0006101575563661754, 0.1311328113079071, -0.1559237539768219, 0.13572059571743011, -0.06162552163004875, 0.06557735055685043, 0.11810020357370377, 0.030381502583622932, -0.07415543496608734, 0.015747543424367905, 0.03302120789885521, -0.14013740420341492, 0.012056016363203526, 0.11805500090122223, -0.005858732387423515, 0.09862574189901352, -0.006288883276283741, 0.01411551795899868, -0.005232658702880144, -0.008925506845116615, -0.08175937831401825, -0.046020686626434326, -0.039414193481206894, -0.07547380030155182, 0.05598810687661171, 0.0756131187081337, -0.0636674091219902, 0.014185146428644657, -0.07626502215862274, 0.051549192517995834, -0.017119377851486206, -0.0015378898242488503, -0.017741234973073006, -0.14869171380996704, -0.05635412409901619, 0.19297784566879272, 0.011150997132062912, -0.14684024453163147, 0.0224141888320446, -0.046751540154218674, -0.0000847880364744924, -0.15142861008644104, 0.09088630229234695, 0.07793725281953812, -0.035462308675050735, -0.02752518467605114, 0.018942104652523994, -0.027658691629767418, 0.07179068773984909, -0.17427299916744232, -0.09412035346031189 ]
null
null
transformers
# BatteryBERT-uncased for QA **Language model:** batterybert-uncased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 3 base_LM_model = "batterybert-uncased" max_seq_len = 386 learning_rate = 3e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 81.08, "f1": 88.41, ``` Evaluated on the battery device dataset. ``` "precision": 68.27, "recall": 80.88, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batterybert-uncased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batterybert-uncased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatteryBERT-uncased for QA Language model: batterybert-uncased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-uncased for QA \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatteryBERT-uncased for QA \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 66, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatteryBERT-uncased for QA \r\nLanguage model: batterybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.08698596805334091, 0.03932749852538109, -0.004285785369575024, 0.05968068167567253, 0.23035301268100739, 0.03959587216377258, 0.11446557939052582, 0.04404348507523537, -0.10159076005220413, -0.02527792379260063, 0.036759987473487854, -0.032385025173425674, 0.07063262164592743, 0.07091840356588364, 0.006257476285099983, -0.09634119272232056, 0.04323998838663101, -0.034888606518507004, 0.02366548776626587, 0.14080674946308136, 0.11539973318576813, -0.09064141660928726, 0.11492826789617538, 0.011028829030692577, -0.09012522548437119, 0.061866819858551025, -0.006155235227197409, -0.07663898915052414, 0.10015404224395752, -0.08250439912080765, 0.12265461683273315, 0.05563902482390404, 0.016290398314595222, -0.17460963129997253, 0.03748232498764992, 0.02474411018192768, 0.026879984885454178, 0.030905529856681824, 0.01759100891649723, -0.03707865625619888, 0.04441438615322113, 0.024758687242865562, 0.03600586950778961, 0.024131428450345993, -0.05659226328134537, -0.23020128905773163, -0.07170940190553665, -0.05041535571217537, 0.06732499599456787, 0.14082343876361847, -0.02970399148762226, 0.09142252057790756, -0.04341396316885948, 0.06880179047584534, 0.16883401572704315, -0.11200181394815445, 0.00341883790679276, 0.10052333027124405, 0.04294520616531372, -0.0008404811378568411, -0.01422133855521679, -0.0773918405175209, -0.002452380955219269, 0.055236779153347015, 0.040663883090019226, -0.055111806839704514, -0.15070584416389465, 0.11081019043922424, -0.09863951057195663, -0.0063990261405706406, 0.19468244910240173, -0.04626400023698807, -0.07692231237888336, 0.005333242937922478, -0.12012762576341629, -0.03920476511120796, -0.023426387459039688, -0.016691608354449272, -0.029297009110450745, -0.002081978367641568, -0.07801024615764618, 0.029185624793171883, -0.011957977898418903, -0.10424548387527466, -0.12951470911502838, 0.08677752315998077, 0.055702585726976395, 0.03774087503552437, -0.11838459223508835, 0.07793234288692474, -0.005740609485656023, -0.1526750922203064, -0.03532590717077255, -0.06483085453510284, -0.07362047582864761, 0.010972385294735432, 0.01839928701519966, 0.011713211424648762, 0.12482316046953201, 0.09091371297836304, -0.030984750017523766, 0.012232465669512749, 0.01464757975190878, -0.06847461313009262, 0.0016461571212857962, 0.14805036783218384, -0.029288994148373604, -0.1323276311159134, -0.008921937085688114, 0.03526972979307175, -0.007763272151350975, -0.022154754027724266, -0.06488299369812012, -0.009122396819293499, 0.06463376432657242, 0.06450191885232925, 0.01959524117410183, 0.03526417911052704, -0.05289573594927788, -0.037038739770650864, 0.049494419246912, -0.10514700412750244, -0.0492822490632534, -0.02461130917072296, 0.01641966961324215, 0.01069937739521265, -0.05031566321849823, 0.0232847910374403, -0.03548835590481758, -0.025358375161886215, -0.07940305024385452, -0.09127701818943024, -0.11879423260688782, -0.0836484357714653, -0.03900882601737976, 0.12757372856140137, 0.06016397103667259, -0.1236957311630249, -0.3323834538459778, 0.089286208152771, -0.0056642163544893265, 0.009893600828945637, -0.037403132766485214, -0.03318687155842781, -0.04161316528916359, -0.030300172045826912, -0.041450679302215576, 0.09775931388139725, -0.06292294710874557, 0.0686909407377243, 0.17068062722682953, 0.0203277375549078, -0.05136873573064804, 0.04474798962473869, -0.1478475034236908, 0.007304334547370672, -0.13660143315792084, 0.044344231486320496, -0.12082643061876297, -0.007725197356194258, -0.10853676497936249, -0.07564470916986465, 0.12411177158355713, -0.019338887184858322, 0.08262497186660767, 0.14458197355270386, -0.08334718644618988, -0.12011250853538513, 0.03223128989338875, -0.13744276762008667, -0.15048812329769135, 0.04691881313920021, -0.013736225664615631, 0.017339907586574554, 0.0171369481831789, 0.05168277025222778, 0.14219233393669128, -0.16592244803905487, -0.18187552690505981, -0.032819777727127075, 0.04475044086575508, -0.011120264418423176, 0.08370335400104523, 0.03936569020152092, -0.07421799749135971, 0.07878788560628891, -0.06841646879911423, 0.052648186683654785, -0.05216467007994652, -0.1092076525092125, -0.06081699579954147, -0.03716502711176872, 0.13844002783298492, 0.0038209366612136364, -0.05603726953268051, 0.027626052498817444, -0.09280828386545181, 0.11184750497341156, 0.10952215641736984, -0.04455213621258736, -0.003058790694922209, -0.19373372197151184, 0.09863302856683731, -0.03418074548244476, 0.016955498605966568, -0.11340481787919998, -0.16876018047332764, 0.08392664790153503, -0.032316114753484726, -0.07031570374965668, -0.04749813675880432, 0.017222318798303604, 0.026354100555181503, -0.1107836589217186, -0.026492485776543617, -0.15185002982616425, -0.02200530469417572, 0.015129299834370613, -0.13505421578884125, 0.044854458421468735, -0.026394503191113472, 0.08975453674793243, 0.025789977982640266, 0.021252982318401337, 0.12311417609453201, 0.03206268325448036, -0.027124105021357536, -0.01911776140332222, -0.026290275156497955, 0.0645386278629303, -0.032047826796770096, -0.03827331215143204, -0.030730193480849266, -0.02049975097179413, -0.01027995441108942, -0.02963055670261383, -0.11945856362581253, 0.07855924218893051, 0.14148473739624023, 0.11914893239736557, 0.01737268455326557, -0.08936400711536407, -0.002426520921289921, -0.04189978539943695, -0.09364351630210876, -0.020359471440315247, 0.22195307910442352, 0.0004948139539919794, 0.02309233322739601, -0.02940911427140236, 0.0170110072940588, -0.0003946272481698543, 0.045833881944417953, 0.06764804571866989, 0.07966713607311249, 0.10230230540037155, 0.06351348012685776, 0.05621163174510002, 0.07034675031900406, -0.05746865272521973, 0.12528178095817566, -0.08787710219621658, -0.03544633463025093, -0.03979581594467163, -0.10409623384475708, -0.021833986043930054, 0.26712366938591003, -0.13136768341064453, 0.06971180438995361, 0.06694669276475906, -0.001894654124043882, 0.029981303960084915, -0.02813059836626053, -0.011667419224977493, -0.018659794703125954, -0.055595289915800095, -0.1734096109867096, 0.033035147935152054, 0.05743382126092911, 0.004851077683269978, -0.02093781717121601, -0.04750775918364525, 0.03043777495622635, 0.015371226705610752, -0.06793743371963501, 0.22183439135551453, -0.10839208215475082, -0.20677998661994934, -0.10804806649684906, -0.09465934336185455, -0.039139263331890106, -0.07655668258666992, 0.021749624982476234, -0.12392190098762512, -0.09478213638067245, 0.02221049927175045, 0.1837887465953827, 0.016719777137041092, 0.07313095033168793, 0.03397804871201515, -0.0026423779781907797, 0.014720620587468147, -0.10828371345996857, 0.014539127238094807, 0.05885687097907066, -0.10962630808353424, 0.03441787138581276, 0.0604155994951725, 0.08302050083875656, 0.099287249147892, 0.007212545722723007, 0.016764042899012566, -0.05044535547494888, 0.2304670512676239, -0.09844520688056946, -0.0432867594063282, 0.29166674613952637, 0.12361876666545868, -0.002164755715057254, 0.19386786222457886, -0.03519422933459282, -0.01467779278755188, 0.096133291721344, -0.010962807573378086, -0.049032632261514664, -0.2477571964263916, -0.08821520954370499, -0.03343302384018898, 0.06907321512699127, -0.07113642245531082, 0.009940260089933872, -0.04641497880220413, 0.09294673055410385, 0.019853899255394936, -0.018675075843930244, -0.10853241384029388, 0.0916176587343216, 0.2479788213968277, 0.058613602072000504, 0.10064523667097092, -0.006494247354567051, 0.02305304817855358, 0.06637547165155411, 0.05527045205235481, 0.16625888645648956, -0.055141985416412354, 0.0371810719370842, 0.032838817685842514, 0.16233427822589874, 0.031872399151325226, 0.030656687915325165, -0.006403244566172361, 0.0009456411935389042, 0.01952621527016163, -0.038589444011449814, 0.00836225040256977, -0.030231140553951263, -0.034054454416036606, 0.012213614769279957, -0.021443532779812813, 0.15260091423988342, 0.04934132844209671, -0.02608577162027359, 0.018008286133408546, -0.1925753653049469, -0.03557923063635826, 0.03439252823591232, -0.04316762089729309, -0.09961452335119247, 0.03263380378484726, 0.02016756683588028, -0.12455260753631592, 0.034636106342077255, -0.03832809999585152, 0.06885799020528793, -0.06230288743972778, 0.05857554450631142, 0.06222029775381088, 0.11137466877698898, 0.019099783152341843, 0.12258555740118027, -0.39314478635787964, 0.09339063614606857, 0.045352112501859665, 0.0761498510837555, 0.034764111042022705, 0.053940966725349426, -0.05369042232632637, -0.04158548265695572, 0.05118879675865173, 0.0008308853139169514, 0.06894969195127487, 0.04335653409361839, -0.1577875018119812, 0.14812125265598297, 0.009704072028398514, -0.0009916264098137617, 0.07426048815250397, -0.02685743011534214, 0.0403263084590435, -0.014625625684857368, -0.0032941484823822975, -0.13842988014221191, -0.0889165997505188, 0.06378912925720215, -0.05895564705133438, 0.11812348663806915, -0.1493757665157318, -0.08436544984579086, 0.05729648470878601, 0.05877010151743889, -0.07578590512275696, -0.11227767914533615, -0.14493624866008759, 0.007366240490227938, 0.129490926861763, -0.10934089124202728, 0.08042427152395248, 0.013540112413465977, 0.056167494505643845, -0.019995804876089096, -0.10359068214893341, 0.05028791353106499, -0.0772324800491333, -0.14978298544883728, -0.02299431525170803, 0.0297502800822258, 0.09585902839899063, 0.052759651094675064, -0.002568146213889122, -0.013584611937403679, -0.1352282166481018, -0.12374641001224518, -0.03703032061457634, -0.06308725476264954, 0.0320073738694191, 0.023051586002111435, 0.14337630569934845, -0.025939805433154106, -0.02900220826268196, -0.027283066883683205, 0.006640880834311247, 0.15455518662929535, -0.021344514563679695, 0.0611402653157711, 0.11227129399776459, -0.014896371401846409, -0.20720353722572327, -0.06827419251203537, 0.07292649149894714, 0.0544796958565712, -0.02705458365380764, -0.25202223658561707, 0.13386355340480804, -0.02592969313263893, -0.06721067428588867, -0.10194863379001617, -0.12838229537010193, -0.11690802127122879, 0.18782749772071838, 0.06349927932024002, 0.05382997915148735, -0.06730454415082932, -0.025070859119296074, -0.04990571737289429, -0.264492392539978, 0.12358757108449936, -0.1524866819381714, 0.027943644672632217, -0.022825438529253006, 0.24719572067260742, -0.02818625420331955, -0.03038826212286949, 0.14280472695827484, 0.0619012787938118, -0.0288398340344429, 0.030677270144224167, -0.0055918884463608265, 0.1298472136259079, 0.029972312971949577, 0.0375572070479393, 0.0000436607951996848, 0.12128525972366333, -0.11735929548740387, -0.035989824682474136, -0.05872226133942604, 0.09109312295913696, -0.04136187583208084, -0.16319389641284943, -0.0937151238322258, 0.04255323112010956, -0.038101062178611755, -0.06006617844104767, 0.08243302255868912, 0.04199046641588211, -0.005221089813858271, 0.09615623950958252, 0.1548653095960617, 0.02964603714644909, -0.07180555909872055, 0.033282794058322906, -0.013052757829427719, 0.05833256244659424, -0.1371697634458542, 0.041173625737428665, 0.10909974575042725, 0.027813415974378586, 0.011230183765292168, 0.04614455997943878, -0.07783074676990509, 0.006980773061513901, 0.07283822447061539, -0.15507036447525024, -0.1447363644838333, -0.07305427640676498, 0.18074268102645874, -0.21221354603767395, 0.06549554318189621, 0.11786500364542007, -0.010854693129658699, -0.07180758565664291, 0.027348587289452553, 0.03289935365319252, -0.012754545547068119, 0.17075397074222565, 0.08952146768569946, 0.13542960584163666, -0.12893255054950714, 0.058445900678634644, 0.030565308406949043, 0.021114937961101532, -0.043086446821689606, 0.003984631504863501, -0.0982753112912178, 0.012328775599598885, 0.010940379463136196, 0.11579132825136185, -0.13357087969779968, -0.07433725893497467, -0.16407209634780884, -0.11962708830833435, 0.002744807628914714, 0.06957200169563293, 0.08246643096208572, 0.07798963040113449, -0.033148881047964096, 0.016851259395480156, -0.0798344612121582, 0.09984748065471649, 0.007235702592879534, 0.023541487753391266, -0.11911360174417496, -0.08255663514137268, -0.031712375581264496, 0.11876311898231506, -0.05529652163386345, 0.00203006062656641, -0.08781365305185318, 0.023884685710072517, -0.162776380777359, -0.0745471715927124, -0.018020352348685265, -0.019316470250487328, 0.021601179614663124, -0.07333677262067795, -0.039016272872686386, 0.06486556679010391, -0.04290728271007538, 0.054583750665187836, 0.047368742525577545, 0.018191616982221603, -0.12490936368703842, 0.09923875331878662, 0.08140486478805542, -0.051124535501003265, 0.11921630799770355, 0.07137075066566467, 0.06920953840017319, 0.1161423921585083, -0.07338370382785797, 0.01645137369632721, -0.0061124409548938274, 0.07691352814435959, 0.07633010298013687, -0.13659171760082245, -0.00028410041704773903, -0.007178917992860079, 0.01208765059709549, 0.02704722248017788, 0.06356003880500793, -0.1037060096859932, 0.061479806900024414, -0.05948967859148979, 0.03537724167108536, -0.05201642960309982, 0.040428195148706436, 0.09517255425453186, 0.08698487281799316, 0.10009602457284927, -0.0425884909927845, -0.005824178457260132, -0.11722954362630844, 0.011926427483558655, -0.03027569316327572, -0.013558567501604557, -0.0377051904797554, -0.03829330578446388, 0.06836549937725067, 0.03250677138566971, 0.15202362835407257, -0.02364511601626873, 0.056205444037914276, 0.010740796104073524, 0.11389505863189697, 0.09162016957998276, -0.012625610455870628, 0.09082269668579102, 0.09897138923406601, 0.09301779419183731, 0.12129779905080795, -0.02341420017182827, 0.05558715760707855, 0.11987171322107315, 0.11778515577316284, 0.06252948939800262, 0.044027283787727356, 0.04242672771215439, 0.04510071501135826, -0.10940343886613846, -0.08359487354755402, -0.0037531547714024782, -0.06929388642311096, 0.02121257409453392, -0.0052666752599179745, 0.180522158741951, 0.07430262118577957, -0.11092469096183777, 0.08480433374643326, -0.0052962349727749825, -0.10674316436052322, -0.09677024930715561, 0.10782507061958313, -0.04213139787316322, -0.2246483564376831, 0.03432169184088707, -0.14946776628494263, -0.1166754812002182, 0.18881912529468536, -0.002653449075296521, -0.006798182614147663, 0.14263470470905304, 0.19781504571437836, -0.03477797284722328, -0.03793199360370636, -0.015618684701621532, 0.004997978452593088, 0.06021542102098465, 0.030641108751296997, -0.032617561519145966, 0.11261028051376343, 0.04905642569065094, 0.006583981215953827, -0.07153654098510742, 0.018017882481217384, -0.007891369052231312, -0.03398912772536278, -0.018350383266806602, 0.006180742289870977, -0.0759461298584938, 0.19158649444580078, -0.009391996078193188, 0.08622321486473083, 0.010645102709531784, 0.31417420506477356, -0.05105236545205116, -0.15925882756710052, -0.13095615804195404, 0.06665048003196716, -0.05182858183979988, 0.053978174924850464, 0.05275724083185196, -0.08300826698541641, -0.1180279329419136, 0.14360876381397247, 0.020006397739052773, -0.0263507142663002, -0.0503988154232502, 0.03881409391760826, -0.018170064315199852, -0.05552028492093086, 0.034980811178684235, 0.08202419430017471, 0.1964358687400818, -0.06492284685373306, -0.02394562028348446, -0.052392907440662384, -0.07636861503124237, 0.00002898256207117811, 0.030833112075924873, 0.009737838990986347, -0.0219753198325634, -0.02480044588446617, 0.04371153190732002, -0.11133505403995514, -0.09713180363178253, -0.11165972799062729, -0.03251514956355095, -0.11617191880941391, -0.06618695706129074, 0.20961885154247284, 0.0072639803402125835, 0.0006241013179533184, -0.04144478961825371, 0.08369404822587967, 0.04774120822548866, 0.03757697716355324, -0.09034355729818344, -0.057971272617578506, 0.19230756163597107, -0.1346958875656128, 0.14735355973243713, -0.04601577669382095, 0.07500005513429642, 0.11014778912067413, -0.024795986711978912, -0.11661956459283829, 0.05516171082854271, 0.05457956716418266, -0.17519864439964294, -0.008534434251487255, 0.1366322636604309, 0.02484210394322872, 0.11273666471242905, -0.009046497754752636, 0.08290249109268188, -0.020100776106119156, 0.001324741868302226, -0.034012991935014725, -0.04852183163166046, -0.028006399050354958, -0.06942161917686462, 0.046020571142435074, 0.07462666183710098, -0.06048688665032387, 0.07197467982769012, -0.08204131573438644, 0.09824317693710327, -0.038142185658216476, 0.08586987107992172, 0.02080049179494381, -0.1184590607881546, -0.0642705038189888, 0.15767262876033783, -0.06942521035671234, -0.1189495101571083, 0.015808630734682083, -0.00282580079510808, 0.005742460954934359, -0.08856360614299774, 0.14512532949447632, 0.06001986190676689, -0.008105543442070484, -0.02072284370660782, 0.03970736265182495, -0.05787619948387146, 0.05889512225985527, -0.17032894492149353, -0.09099191427230835 ]
null
null
transformers
# BatteryBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the [bert-base-uncased](https://huggingface.co/bert-base-uncased) weights. It was introduced in [this paper](paper_link) and first released in [this repository](https://github.com/ShuHuang/batterybert). This model is uncased: it does not make a difference between english and English. ## Model description BatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the [bert-base-uncased](https://huggingface.co/bert-base-uncased) weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the [bert-base-uncased](https://huggingface.co/bert-base-uncased) weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at [Github](https://github.com/ShuHuang/batterybert/blob/main/corpus.txt). ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 30,522. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=batterybert) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='batterydata/batterybert-uncased') >>> unmasker("Hello I'm a <mask> model.") ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batterybert-uncased') model = BertModel.from_pretrained('batterydata/batterybert-uncased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batterybert-uncased') model = TFBertModel.from_pretrained('batterydata/batterybert-uncased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Evaluation results Final loss: 1.0317. ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": ["exbert"], "datasets": ["batterypapers"]}
fill-mask
batterydata/batterybert-uncased
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "exbert", "en", "dataset:batterypapers", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-uncased weights. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. ## Model description BatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-uncased weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github. ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 30,522. The inputs of the model are then of the form: The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by '[MASK]'. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Evaluation results Final loss: 1.0317. ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.", "## Model description\n\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 30,522. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.0317.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.", "## Model description\n\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatteryBERT model was pretrained on the full text of battery papers only, after initialized from the bert-base-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 30,522. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.0317.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 61, 84, 283, 97, 3, 136, 113, 128, 49, 11, 50 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the bert-base-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.## Model description\n\nBatteryBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the bert-base-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.05132001265883446, 0.08077171444892883, -0.006066843867301941, 0.04470290616154671, 0.09435369819402695, 0.01168136764317751, 0.13184002041816711, 0.006085169035941362, -0.02031603828072548, 0.043706148862838745, 0.04241371527314186, 0.0047059440985322, 0.05085068568587303, 0.05189827084541321, 0.10268843173980713, -0.25415903329849243, 0.0418178029358387, -0.10190104693174362, 0.09251967072486877, 0.0729428231716156, 0.05511574447154999, -0.09618941694498062, 0.02680739015340805, 0.03830705210566521, 0.04058275744318962, -0.01000748760998249, -0.020798856392502785, -0.0807870402932167, 0.09385265409946442, 0.02769286371767521, 0.08151248097419739, 0.0005065630539320409, 0.02381017804145813, -0.03980039432644844, 0.02397853136062622, 0.04829318821430206, 0.0027915802784264088, 0.06865672022104263, 0.05453873425722122, -0.018247565254569054, 0.11617518961429596, 0.052171491086483, 0.07680836319923401, 0.04673750326037407, -0.09731627255678177, -0.13870778679847717, -0.018208997324109077, -0.0011315499432384968, -0.00037695918581448495, 0.10071530193090439, -0.03497585281729698, 0.08258410543203354, -0.027518054470419884, 0.06941034644842148, 0.18500418961048126, -0.10673001408576965, -0.009394319728016853, 0.0166102796792984, 0.01928923837840557, -0.008117643184959888, -0.011283432133495808, -0.04018455743789673, -0.029987361282110214, 0.0499265119433403, 0.07342764735221863, -0.005283973179757595, 0.041739508509635925, -0.04641655460000038, -0.15394186973571777, -0.02235059067606926, 0.09810483455657959, -0.044910840690135956, -0.09394503384828568, -0.11934492737054825, -0.08518185466527939, 0.09888077527284622, -0.0008202158496715128, 0.05596320331096649, -0.029352782294154167, 0.01699872687458992, 0.01376846618950367, -0.049065280705690384, -0.06161831319332123, -0.07361678779125214, -0.10825321823358536, 0.18764039874076843, 0.028780421242117882, 0.04250122979283333, -0.0594753697514534, 0.07989799976348877, -0.1103513091802597, -0.07240137457847595, -0.05750960111618042, -0.045852016657590866, -0.07895201444625854, 0.004156418610364199, -0.022335125133395195, -0.20644162595272064, -0.027961453422904015, 0.12010577321052551, -0.01592230796813965, -0.0026155232917517424, -0.020324276760220528, 0.023838452994823456, 0.10963156074285507, 0.06760064512491226, -0.102493055164814, 0.08282692730426788, 0.010336332954466343, -0.04515521600842476, 0.047168027609586716, -0.053373340517282486, -0.013852176256477833, 0.027034908533096313, -0.05261136218905449, 0.022237448021769524, 0.04901799559593201, 0.06239418312907219, -0.021936707198619843, -0.04994455352425575, 0.06465278565883636, -0.11085376143455505, -0.04235246405005455, 0.0019339881837368011, -0.02324463240802288, -0.008911306969821453, 0.061499737203121185, -0.08501516282558441, -0.061349090188741684, 0.06392543762922287, -0.07281642407178879, -0.07221385091543198, -0.12073547393083572, -0.14775218069553375, -0.029223065823316574, 0.00947378110140562, -0.06386386603116989, -0.08291906118392944, -0.2548932433128357, -0.024263599887490273, 0.03056691773235798, 0.005655224435031414, 0.00451750447973609, -0.029942594468593597, 0.0071617173962295055, -0.00332687352783978, 0.013011857867240906, -0.009090560488402843, -0.005162422079592943, 0.0062151760794222355, -0.02080378122627735, 0.084470734000206, 0.03571973368525505, 0.012847241014242172, -0.08903619647026062, 0.002318240702152252, -0.21335716545581818, 0.10410274565219879, -0.07642805576324463, -0.020196283236145973, -0.0272495299577713, 0.0020795867312699556, -0.0077797966077923775, 0.01458959374576807, 0.0436413399875164, 0.13202859461307526, -0.10398346930742264, -0.0985228642821312, 0.2175031453371048, -0.16376855969429016, 0.054057806730270386, 0.09528154879808426, -0.05684869736433029, 0.06672278791666031, 0.11444459855556488, -0.020507847890257835, 0.09951579570770264, -0.11060137301683426, -0.05682292580604553, 0.06455212086439133, -0.06401797384023666, 0.09141390025615692, 0.09690713882446289, -0.029533492401242256, -0.13370202481746674, 0.035377632826566696, 0.024810437113046646, 0.03772791102528572, -0.0272959116846323, -0.02602737955749035, 0.03883188217878342, -0.022448880597949028, 0.07447436451911926, 0.0013898895122110844, -0.0036595501005649567, 0.025365127250552177, -0.09954888373613358, 0.07837292551994324, 0.12108052521944046, -0.0643785372376442, -0.004225669428706169, -0.10872456431388855, 0.05729898810386658, -0.050242550671100616, -0.0210296381264925, -0.1727277934551239, -0.01056588999927044, 0.03690182790160179, -0.11950498819351196, 0.09358453005552292, -0.022848382592201233, 0.03394371271133423, 0.12542405724525452, -0.05475572869181633, 0.021532641723752022, -0.029428834095597267, -0.03697393089532852, -0.02041487582027912, -0.13862325251102448, 0.007063152268528938, -0.05883999541401863, 0.022009996697306633, 0.04840117692947388, 0.035213518887758255, -0.045292098075151443, -0.054469455033540726, 0.00798074435442686, -0.05606938153505325, 0.04343162477016449, -0.011516283266246319, -0.015042201615869999, -0.009800001047551632, -0.017230505123734474, 0.04375012591481209, -0.07450977712869644, 0.01569533906877041, -0.16878391802310944, -0.14193496108055115, 0.06063767150044441, 0.033772945404052734, -0.10959456861019135, 0.07440073788166046, -0.014002425596117973, 0.0024195206351578236, -0.13339686393737793, -0.0342591218650341, 0.20928217470645905, 0.024055009707808495, 0.04896930232644081, -0.07369527220726013, 0.03070150315761566, 0.029775122180581093, -0.01420722994953394, 0.008803031407296658, 0.08325749635696411, 0.07811009138822556, -0.06319363415241241, 0.013309060595929623, -0.1042560264468193, -0.041646987199783325, 0.17519734799861908, 0.04368007555603981, -0.05056046321988106, -0.061619941145181656, -0.02739296667277813, 0.025151966139674187, 0.03710087761282921, -0.10404892265796661, 0.035770922899246216, 0.04667183384299278, -0.011148135177791119, 0.02367616444826126, -0.021048961207270622, 0.0391322560608387, 0.05391019210219383, 0.014543616212904453, -0.09919033944606781, -0.0684540793299675, -0.026552323251962662, 0.06498230248689651, 0.04894862696528435, -0.004284191876649857, 0.006073343101888895, -0.027627650648355484, -0.09823483228683472, 0.18052177131175995, -0.0871407613158226, -0.2442195564508438, -0.15109194815158844, -0.030646182596683502, -0.016409944742918015, 0.03735515847802162, 0.004864692222326994, -0.01239564549177885, -0.07693023979663849, -0.06940389424562454, 0.13589933514595032, -0.014414394274353981, 0.0035518240183591843, -0.017009051516652107, -0.03077201545238495, -0.05359572544693947, -0.11007589846849442, 0.02596050128340721, -0.024656791239976883, -0.0932232216000557, 0.008004764094948769, -0.02671741507947445, 0.04233039543032646, 0.11261464655399323, 0.025854656472802162, -0.036105070263147354, -0.03030131198465824, 0.18111301958560944, -0.034822508692741394, 0.06002756580710411, 0.14959155023097992, 0.00509171886369586, 0.04831352084875107, 0.052601054310798645, 0.022291209548711777, -0.011444809846580029, 0.05620395392179489, 0.010385517962276936, -0.10077567398548126, -0.1806393712759018, -0.08699460327625275, -0.051927678287029266, 0.049367599189281464, 0.060941990464925766, 0.005808337591588497, 0.011683447286486626, 0.06704714894294739, -0.05239409953355789, -0.04211511090397835, 0.04506490379571915, 0.07976440340280533, -0.012091484852135181, -0.04515071585774422, 0.1011556014418602, -0.04314336180686951, -0.031030477955937386, 0.10682766139507294, -0.13243898749351501, 0.16493874788284302, -0.07040122151374817, 0.12169945240020752, 0.060705821961164474, -0.0014277000445872545, 0.06728272885084152, 0.14529471099376678, -0.07357300072908401, 0.03485417366027832, -0.041989151388406754, -0.0818822905421257, -0.0009396672248840332, 0.042058903723955154, 0.0022135667968541384, 0.015349557623267174, -0.03735930472612381, 0.013741029426455498, 0.007513147778809071, 0.13591276109218597, 0.01954961195588112, -0.1223554015159607, -0.10850661993026733, -0.002863521920517087, -0.005552518181502819, -0.08923294395208359, 0.025030290707945824, 0.08227062970399857, -0.09019213914871216, 0.07563035190105438, 0.004057107027620077, 0.057496484369039536, -0.06495019048452377, 0.011676985770463943, -0.09187997132539749, 0.0971316397190094, -0.05814175680279732, 0.11608260869979858, -0.17831507325172424, 0.04878848418593407, 0.024405639618635178, 0.09852730482816696, -0.10355842113494873, 0.025003895163536072, 0.014665941707789898, -0.05078884959220886, 0.15508867800235748, 0.011106472462415695, -0.04272845387458801, 0.022036142647266388, -0.09747740626335144, 0.004078547470271587, 0.02736177109181881, -0.05672905966639519, 0.050445958971977234, 0.031807612627744675, 0.006963795050978661, 0.010005274787545204, 0.014533419162034988, -0.03890276700258255, -0.124277263879776, 0.05958475545048714, -0.05104231834411621, -0.07160026580095291, -0.06100979819893837, -0.06590423732995987, -0.016101622954010963, 0.1305113583803177, -0.14720234274864197, -0.10049782693386078, -0.11964252591133118, -0.024003727361559868, 0.10547789186239243, -0.09166234731674194, 0.04865262284874916, -0.029092172160744667, 0.13276778161525726, -0.08946050703525543, -0.10828842967748642, 0.06554540246725082, -0.06724763661623001, -0.13081830739974976, -0.03359057381749153, 0.041140031069517136, 0.20249153673648834, 0.06204359233379364, -0.036799415946006775, 0.0547972247004509, -0.04808105155825615, -0.11782242357730865, -0.019534679129719734, 0.1198984906077385, 0.0012566642835736275, 0.10828424245119095, -0.01297740451991558, -0.08661884069442749, -0.01309987809509039, -0.004464687779545784, 0.09625126421451569, 0.1809232085943222, -0.04955585300922394, 0.14955741167068481, 0.28696003556251526, -0.07805965840816498, -0.25277990102767944, -0.06055450066924095, 0.052139442414045334, 0.05735703185200691, 0.005893060937523842, -0.2977624833583832, 0.08073422312736511, 0.01915004476904869, -0.036246925592422485, -0.08117877691984177, -0.17632664740085602, -0.13113634288311005, 0.17582212388515472, 0.030222758650779724, 0.1329612135887146, 0.0006997290765866637, -0.008744081482291222, -0.050312407314777374, 0.09979984909296036, 0.1684071272611618, -0.07063134759664536, 0.1044294610619545, 0.024652794003486633, 0.009348851628601551, 0.03654055297374725, -0.07054038345813751, 0.0711693987250328, -0.013755569234490395, 0.04079451039433479, -0.03979004919528961, -0.024719396606087685, 0.12884213030338287, -0.008961346931755543, 0.10173194855451584, 0.09362185001373291, 0.07743597775697708, -0.003940315917134285, -0.10890010744333267, -0.07668083161115646, 0.09591103345155716, -0.01597605086863041, -0.09738104790449142, -0.08680856227874756, 0.07751326262950897, 0.04741166532039642, -0.018738240003585815, 0.016447382047772408, -0.047983862459659576, 0.060337699949741364, 0.14004898071289062, 0.12917691469192505, 0.03486227989196777, -0.04457375034689903, 0.008981144987046719, -0.028413375839591026, 0.0849924236536026, -0.015838736668229103, 0.03404967114329338, 0.05082168057560921, 0.07409090548753738, 0.07245751470327377, 0.05698130652308464, -0.1787763088941574, -0.005766712594777346, 0.030302492901682854, -0.15946334600448608, -0.13924340903759003, -0.03918210789561272, 0.07652872800827026, -0.13001467287540436, -0.026100723072886467, 0.10588619858026505, -0.018771007657051086, -0.023226013407111168, -0.0038813867140561342, 0.04070909321308136, 0.015078231692314148, 0.13103815913200378, 0.02498818375170231, 0.041878703981637955, -0.0745275542140007, 0.06398006528615952, 0.05071480572223663, 0.006927287671715021, 0.02970188483595848, 0.08655447512865067, -0.07525128871202469, -0.03678203001618385, -0.028562886640429497, 0.14060522615909576, -0.026847034692764282, -0.030930159613490105, -0.03890955075621605, -0.07768626511096954, 0.06954342126846313, 0.1452733874320984, 0.023649465292692184, 0.04542991518974304, -0.0852130800485611, 0.041199374943971634, -0.05300578102469444, 0.054158780723810196, -0.035314567387104034, -0.029274070635437965, 0.0349908210337162, 0.12545475363731384, 0.02035018615424633, 0.029320938512682915, -0.05904781445860863, -0.0937981903553009, -0.09991686046123505, 0.014854039065539837, 0.006887138355523348, 0.01764032617211342, -0.06894005835056305, -0.014801754616200924, -0.010644211433827877, 0.043183378875255585, 0.03329715505242348, -0.004291912075132132, -0.032440122216939926, 0.025267671793699265, -0.041962798684835434, -0.014032491482794285, -0.11561702191829681, 0.03796403110027313, 0.04086726903915405, -0.036053046584129333, 0.09707479178905487, 0.04154856503009796, -0.007186516188085079, 0.005851229187101126, -0.0845542922616005, 0.022142961621284485, -0.03827450051903725, 0.036843691021203995, -0.04126459360122681, -0.0989946722984314, 0.017524998635053635, -0.02949672006070614, -0.03245556727051735, -0.007270145695656538, 0.00007586886931676418, -0.08092867583036423, 0.14983022212982178, 0.11156632751226425, -0.017495404928922653, -0.0959262102842331, 0.04857492074370384, -0.012105048634111881, 0.02425384148955345, 0.05478968471288681, -0.030618174001574516, 0.05799969658255577, -0.098190076649189, 0.0020688252989202738, 0.010197482071816921, 0.009128988720476627, 0.030627161264419556, -0.03645975887775421, 0.06357252597808838, 0.0050742123275995255, 0.13817882537841797, 0.03422749042510986, -0.059523917734622955, -0.008811418898403645, 0.07454913854598999, 0.04235571250319481, -0.02358178049325943, 0.014158508740365505, 0.026231685653328896, -0.031015953049063683, 0.04648711532354355, 0.02331114374101162, -0.011362600140273571, 0.09463973343372345, 0.12850403785705566, -0.022256189957261086, 0.17875754833221436, -0.03194108605384827, 0.041500285267829895, 0.0006142378551885486, -0.08191855996847153, -0.049776285886764526, 0.01399728562682867, 0.03342057019472122, -0.007161213085055351, 0.00849036406725645, 0.06683517247438431, -0.05280018597841263, 0.1681218147277832, 0.06344233453273773, -0.052095383405685425, -0.1301753968000412, -0.24312810599803925, -0.0023490358144044876, 0.05153677985072136, 0.00490125548094511, -0.11386720091104507, -0.04068312421441078, 0.06425848603248596, 0.025012489408254623, -0.0039442237466573715, 0.09146373718976974, -0.09383128583431244, -0.05791383236646652, 0.03181390464305878, -0.02572205290198326, 0.001160075655207038, 0.01888362318277359, 0.001967174932360649, 0.049367595463991165, 0.12693199515342712, 0.039630599319934845, 0.05644778534770012, 0.030941849574446678, 0.035166122019290924, -0.011094653978943825, -0.06980092078447342, -0.017258668318390846, -0.02998059056699276, -0.009741338901221752, 0.16370297968387604, -0.014678320847451687, -0.03276640176773071, 0.004653771407902241, 0.1531863957643509, -0.06094338744878769, -0.09924476593732834, -0.11723891645669937, 0.2530730664730072, -0.0446065254509449, 0.017880979925394058, 0.026957077905535698, -0.0743885338306427, -0.03162362426519394, 0.20443560183048248, 0.15930147469043732, 0.0345965214073658, 0.001307331258431077, 0.040723953396081924, -0.00733640743419528, 0.019192662090063095, 0.08174854516983032, -0.016583554446697235, 0.37553274631500244, -0.04931042343378067, 0.17839139699935913, -0.025541014969348907, -0.01023365929722786, -0.02895469404757023, 0.043153222650289536, -0.004124048165977001, 0.02514338679611683, -0.04886613041162491, 0.05962146818637848, -0.1378832310438156, -0.23707516491413116, -0.04147425666451454, 0.04474238306283951, -0.07559000700712204, -0.007956334389746189, 0.01249692589044571, 0.05158059671521187, 0.07513456791639328, -0.0013136687921360135, 0.04037271812558174, 0.03401024267077446, 0.02806840091943741, -0.10294285416603088, -0.09319716691970825, 0.07713596522808075, -0.013132636435329914, 0.1720978319644928, 0.015772147104144096, 0.021768594160676003, 0.0961354449391365, -0.03803187608718872, -0.09329082816839218, 0.07189395278692245, -0.012524417601525784, -0.08738195151090622, 0.019784782081842422, 0.14727534353733063, -0.03577510640025139, 0.11232082545757294, -0.0003589367843233049, -0.012962671928107738, 0.04602893814444542, -0.04850052297115326, -0.05340701341629028, -0.07877537608146667, 0.07953377813100815, -0.07791012525558472, 0.11655378341674805, 0.12281021475791931, -0.009178160689771175, -0.03289821371436119, -0.0466119721531868, 0.03468344360589981, -0.06718844920396805, -0.002851415891200304, -0.0313941054046154, -0.11522241681814194, -0.032086580991744995, 0.012841509655117989, 0.05113575607538223, -0.26900115609169006, -0.01397332176566124, 0.01667039841413498, -0.0528959296643734, -0.006377618294209242, 0.006853350438177586, -0.006843717768788338, 0.023712998256087303, -0.018845297396183014, -0.012274110689759254, -0.005112147890031338, 0.02985217049717903, -0.13684765994548798, -0.07118408381938934 ]
null
null
transformers
# BatteryOnlyBERT-cased for Battery Abstract Classification **Language model:** batteryonlybert-cased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 14 base_LM_model = "batteryonlybert-cased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 97.33, "Test accuracy": 97.34, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batteryonlybert-cased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batteryonlybert-cased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryOnlyBERT-cased for Battery Abstract Classification Language model: batteryonlybert-cased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryOnlyBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryOnlyBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 73, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryOnlyBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.06501241773366928, 0.08365874737501144, -0.004337509628385305, 0.05261564254760742, 0.2405385822057724, 0.031636666506528854, 0.19674797356128693, 0.04219766706228256, -0.09023874253034592, -0.03372187167406082, 0.04984490945935249, 0.03683934360742569, 0.053219545632600784, 0.05905504152178764, -0.046300314366817474, -0.17478050291538239, 0.030460666865110397, -0.011528155766427517, 0.028184853494167328, 0.12500402331352234, 0.13912920653820038, -0.08421700447797775, 0.11795789003372192, 0.007920478470623493, -0.10449990630149841, 0.02368328534066677, 0.013185950927436352, -0.1288214772939682, 0.12453014403581619, -0.05916745215654373, 0.1511755734682083, 0.04152105376124382, 0.034539882093667984, -0.17286622524261475, 0.03241981193423271, 0.014070657081902027, -0.003350213635712862, 0.07174980640411377, 0.028537293896079063, -0.09952326864004135, 0.11713239550590515, -0.028456764295697212, 0.07145947962999344, 0.029708340764045715, -0.0982276126742363, -0.1307641863822937, -0.04179852828383446, 0.013306993991136551, 0.08614864945411682, 0.11660299450159073, -0.015798846259713173, 0.06794735044240952, -0.05364178121089935, 0.07664470374584198, 0.20587541162967682, -0.11592570692300797, -0.00045222873450256884, 0.12180505692958832, 0.016035296022892, -0.014527713879942894, -0.06532522290945053, -0.012188747525215149, 0.009869161993265152, 0.04395496845245361, 0.06449349969625473, -0.032072361558675766, -0.15019133687019348, 0.09345897287130356, -0.09176276624202728, -0.006690573878586292, 0.24837909638881683, -0.07023420184850693, -0.033854126930236816, -0.046877916902303696, -0.09029106050729752, -0.0759124606847763, -0.03474665433168411, 0.025761280208826065, -0.02923799306154251, 0.022273603826761246, -0.005117699038237333, 0.04955977946519852, -0.05043009668588638, -0.06683320552110672, -0.11047687381505966, 0.08317881077528, 0.0001171964468085207, 0.0029094091150909662, -0.10188572853803635, 0.12878161668777466, 0.07258376479148865, -0.16251730918884277, 0.01118061225861311, -0.061779748648405075, -0.02463182434439659, -0.027509842067956924, 0.019247520714998245, -0.01712856814265251, 0.08915632218122482, 0.07799879461526871, -0.012563307769596577, 0.008180641569197178, 0.000286536494968459, -0.010439116507768631, -0.0014287137892097235, 0.13181206583976746, -0.016128486022353172, -0.05922592431306839, 0.02850032038986683, -0.007428917568176985, 0.03092687390744686, -0.036339666694402695, -0.12371664494276047, -0.009338470175862312, 0.07934632897377014, 0.0682397335767746, -0.029007496312260628, 0.059362057596445084, -0.08096064627170563, -0.05811931565403938, 0.0802873894572258, -0.09692376106977463, -0.017584288492798805, 0.01058017648756504, 0.02620919607579708, 0.044793177396059036, -0.01096936222165823, 0.025488078594207764, -0.02947959490120411, 0.0766303688287735, -0.08264496922492981, -0.04691699892282486, -0.08949851989746094, -0.09066816419363022, 0.0035431350115686655, 0.06009799242019653, 0.06246734783053398, -0.11876857280731201, -0.29272156953811646, 0.05944618582725525, 0.002446593949571252, -0.006705012638121843, -0.07103592902421951, -0.072744220495224, -0.07343286275863647, 0.02093241550028324, -0.0013799271546304226, 0.0460171215236187, -0.07062818109989166, 0.04714023321866989, 0.09201335161924362, 0.037089768797159195, -0.05782737582921982, 0.03745214268565178, -0.12261432409286499, -0.006111808121204376, -0.1903921216726303, 0.033896129578351974, -0.09202848374843597, 0.09307502210140228, -0.07814491540193558, -0.05020155757665634, 0.1167188435792923, -0.02098122239112854, 0.010330837219953537, 0.15889762341976166, -0.09267095476388931, -0.1443946212530136, 0.06069629639387131, -0.12410115450620651, -0.09382577985525131, 0.05503888428211212, -0.043657779693603516, 0.03030509315431118, 0.07808800041675568, 0.05642241984605789, 0.07818213850259781, -0.11707215756177902, -0.10264820605516434, 0.0023360878694802523, 0.01799284666776657, -0.004684061277657747, 0.09823642671108246, 0.04449018836021423, -0.0415017306804657, 0.0613749623298645, -0.027606580406427383, 0.05938512831926346, -0.038466617465019226, -0.08825765550136566, -0.02210179716348648, -0.027799906209111214, 0.1274227797985077, 0.025335248559713364, 0.014352587051689625, -0.010822533629834652, -0.0867450088262558, 0.13319800794124603, 0.0895732119679451, -0.04295748844742775, 0.007090195547789335, -0.13309642672538757, 0.07378305494785309, 0.04270307719707489, 0.013350256718695164, -0.10861455649137497, -0.09043896943330765, 0.037942074239254, 0.06025763973593712, -0.035464756190776825, -0.06965937465429306, 0.006377389654517174, 0.032349843531847, -0.10601738840341568, -0.01065972913056612, -0.06759794801473618, -0.0031756225507706404, -0.0039087701588869095, -0.21992965042591095, 0.07967808097600937, -0.03670363500714302, 0.1710166186094284, -0.12858562171459198, 0.061871252954006195, 0.14957810938358307, 0.07551447302103043, -0.01769198104739189, -0.0005028418963775039, 0.005011247005313635, 0.03081868216395378, -0.06840242445468903, -0.04529501497745514, 0.0031738970428705215, -0.001682301051914692, -0.06481283158063889, 0.017437899485230446, -0.13731415569782257, 0.09868309646844864, 0.16801555454730988, -0.009217808023095131, -0.03329191356897354, -0.04816646873950958, 0.00394129054620862, -0.019765090197324753, -0.10099004954099655, -0.0028616050258278847, 0.22206911444664001, -0.017659926787018776, 0.07912835478782654, -0.04181808605790138, -0.016937298700213432, -0.003898712107911706, 0.03120887465775013, 0.04498232528567314, 0.12999725341796875, 0.0896068811416626, -0.05256140977144241, 0.10135895013809204, 0.05270165577530861, -0.062474772334098816, 0.1534179002046585, -0.06479667872190475, -0.016943736001849174, 0.0027433570940047503, -0.09624132513999939, -0.002475936431437731, 0.19415557384490967, -0.18560215830802917, 0.012422140687704086, 0.055327463895082474, -0.027660224586725235, 0.009002016857266426, -0.07732043415307999, -0.022237051278352737, -0.004907062277197838, -0.02311665005981922, -0.10782072693109512, 0.011268600821495056, 0.04297017306089401, 0.06545762717723846, -0.01380095910280943, -0.08052650839090347, 0.06615521013736725, 0.006492134649306536, -0.11149349063634872, 0.23008199036121368, -0.11730296164751053, -0.24271516501903534, -0.12722373008728027, -0.09374130517244339, -0.018327457830309868, -0.008040356449782848, 0.030238216742873192, -0.11615060269832611, -0.0946149080991745, -0.010155749507248402, 0.1044355109333992, -0.021120944991707802, 0.07110308855772018, 0.07840197533369064, 0.0255033690482378, -0.028727078810334206, -0.10672414302825928, -0.00024454068625345826, 0.04079733416438103, -0.04648243263363838, 0.045542653650045395, 0.02268212102353573, 0.04161718487739563, 0.18139120936393738, 0.009477773681282997, 0.01802777871489525, -0.0531272292137146, 0.1437254399061203, -0.11173353344202042, -0.03605370596051216, 0.20477429032325745, 0.036777034401893616, -0.002721658442169428, 0.16620570421218872, -0.00010769470100058243, -0.02522510103881359, 0.0771094262599945, -0.02140726149082184, -0.05572439357638359, -0.2775493860244751, -0.13025422394275665, -0.06498926877975464, 0.09958260506391525, -0.032833486795425415, 0.0376281701028347, 0.0384630486369133, 0.04908738285303116, -0.003070431761443615, 0.06536572426557541, -0.017765123397111893, 0.1246507540345192, 0.22890064120292664, 0.06178761273622513, 0.08472073823213577, -0.028881974518299103, -0.03495185449719429, 0.06595293432474136, -0.04389483109116554, 0.12788215279579163, 0.0019747649785131216, 0.08582988381385803, 0.012105681002140045, 0.006726919673383236, 0.06916981935501099, 0.04872139170765877, 0.008017357438802719, -0.007672073785215616, 0.020746717229485512, -0.04306647926568985, 0.04284452274441719, -0.015600060112774372, -0.09497509151697159, -0.011381175369024277, -0.04209316894412041, 0.0871507078409195, 0.08065430074930191, -0.031779054552316666, 0.031115243211388588, -0.23727941513061523, -0.05311214178800583, 0.06609338521957397, -0.05709310248494148, -0.09342537820339203, 0.02542734518647194, -0.01150144636631012, -0.13211007416248322, 0.08539135009050369, -0.012708621099591255, 0.08406493067741394, -0.12163059413433075, 0.05283050611615181, 0.0256962850689888, 0.04889125004410744, -0.013298802077770233, 0.11898624151945114, -0.35172969102859497, 0.14281249046325684, 0.045435741543769836, 0.054622963070869446, -0.031055105850100517, 0.03394275903701782, -0.013777107000350952, 0.06261333078145981, 0.039406877011060715, -0.00042642539483495057, 0.03705527260899544, 0.03968275710940361, -0.13188926875591278, 0.05224330723285675, 0.012855114415287971, -0.04827435687184334, 0.029678579419851303, -0.012555951252579689, 0.013647623360157013, 0.028130829334259033, -0.07391399145126343, -0.1187521368265152, -0.12439538538455963, 0.07656911760568619, -0.029431341215968132, 0.11346055567264557, -0.0969395786523819, -0.10071784257888794, -0.0019939495250582695, 0.12344019114971161, -0.1402939260005951, -0.10016690939664841, -0.14187367260456085, 0.003310640575364232, 0.04396609961986542, -0.0698319524526596, 0.06448467075824738, 0.0016856389120221138, 0.0857597291469574, -0.04259667173027992, -0.09764961898326874, 0.1013651043176651, -0.06522975862026215, -0.18293456733226776, -0.05604014918208122, 0.11007720977067947, 0.10379894822835922, 0.0608929768204689, -0.01407789345830679, -0.0017529603792354465, -0.0869765505194664, -0.12792465090751648, 0.0008756871684454381, 0.008871659636497498, 0.06286471337080002, 0.0329475961625576, 0.06685356050729752, -0.09294378012418747, -0.031604960560798645, -0.07067914307117462, 0.08946558833122253, 0.18185588717460632, -0.023503677919507027, 0.1099642962217331, 0.07171039283275604, -0.06283774226903915, -0.2294645458459854, -0.03779773786664009, 0.005778544116765261, 0.033702969551086426, -0.04799268767237663, -0.25812268257141113, 0.13997863233089447, -0.03256012126803398, -0.05982019752264023, -0.023066824302077293, -0.15006031095981598, -0.13880373537540436, 0.2005118578672409, 0.015346527099609375, 0.10807210206985474, -0.09337710589170456, -0.03764783963561058, -0.10045251995325089, -0.2120295614004135, 0.20875215530395508, -0.16202464699745178, 0.05033290013670921, -0.017732081934809685, 0.17729346454143524, -0.01463077962398529, -0.032755739986896515, 0.1322031468153, 0.04455626755952835, 0.0005026116850785911, -0.02375400997698307, -0.0885404497385025, 0.2129177749156952, 0.028750885277986526, 0.012033108621835709, -0.06703503429889679, 0.04489016905426979, -0.12564696371555328, -0.04566207900643349, -0.02268644981086254, 0.11462071537971497, -0.03556735813617706, -0.13669122755527496, -0.09919167309999466, 0.01936335489153862, -0.01821419596672058, -0.04492264986038208, 0.19725768268108368, 0.05837525427341461, 0.025583747774362564, 0.04063195362687111, 0.22749006748199463, -0.0007042623474262655, 0.06266085803508759, 0.0014452851610258222, -0.017830858007073402, 0.0518207810819149, -0.18759147822856903, 0.03183523938059807, 0.09609802067279816, 0.03641366586089134, 0.050382960587739944, 0.09127853810787201, -0.056403081864118576, -0.02115478552877903, 0.09844373911619186, -0.22670680284500122, -0.04265623912215233, -0.09326844662427902, 0.14324869215488434, -0.1311689168214798, 0.07851238548755646, 0.11462119966745377, -0.03723858296871185, -0.05919725447893143, 0.04949668422341347, 0.04452221840620041, -0.0342918299138546, 0.12748390436172485, 0.047269098460674286, 0.10843862593173981, -0.1369447112083435, 0.06698920577764511, 0.07256045192480087, 0.03784555196762085, -0.0295167975127697, 0.09393589198589325, -0.13717226684093475, -0.01519556250423193, 0.012705342844128609, 0.19645202159881592, -0.10405365377664566, -0.04651237651705742, -0.1776255965232849, -0.15206502377986908, 0.028083544224500656, 0.14625053107738495, 0.10648289322853088, 0.09252536296844482, -0.05717359855771065, 0.00655137374997139, -0.09456972032785416, 0.09010729193687439, 0.01695146970450878, 0.027835382148623466, -0.14531342685222626, 0.013185188174247742, -0.01382906548678875, 0.041751928627491, -0.05821466073393822, -0.001868239138275385, -0.14234967529773712, 0.021714283153414726, -0.07138494402170181, -0.04946235567331314, -0.0351237878203392, -0.020660586655139923, 0.0016756175318732858, -0.0434090718626976, -0.022530091926455498, 0.028714369982481003, -0.04483318328857422, 0.05566506087779999, 0.020652098581194878, 0.06924498081207275, -0.10962370038032532, 0.05851041525602341, 0.0704394280910492, -0.06224290281534195, 0.13511045277118683, 0.08509320765733719, 0.060804229229688644, 0.09545696526765823, -0.17548951506614685, 0.022322455421090126, 0.06084717810153961, 0.0391644649207592, 0.03883344680070877, -0.12875060737133026, -0.009210160933434963, -0.0012636992614716291, 0.029618440195918083, 0.05504591390490532, 0.04672447219491005, -0.0945296585559845, 0.062059734016656876, -0.04858645051717758, -0.017896700650453568, -0.05472723767161369, 0.035606034100055695, 0.13898909091949463, 0.03319508209824562, 0.12392206490039825, -0.061321016401052475, 0.005093637388199568, -0.08334166556596756, 0.02897103875875473, -0.01764594577252865, -0.07768043875694275, -0.09659647196531296, -0.040762998163700104, 0.0562414787709713, 0.001244748244062066, 0.20571361482143402, 0.011740488931536674, -0.037916600704193115, 0.054660286754369736, 0.17052890360355377, 0.06077021360397339, 0.03528015315532684, 0.11387567967176437, 0.10008592903614044, 0.038035739213228226, 0.07847928255796432, -0.012433155439794064, 0.12243558466434479, 0.08442767709493637, 0.15631312131881714, 0.041421759873628616, -0.009251300245523453, 0.0435672365128994, 0.014736943878233433, -0.0828554630279541, -0.14622993767261505, 0.02628643810749054, -0.0658411905169487, 0.040979091078042984, -0.04755868762731552, 0.18726854026317596, 0.1166568174958229, -0.09950462728738785, 0.04735076054930687, -0.025419434532523155, -0.10662813484668732, -0.10484501719474792, -0.013743575662374496, -0.04191472753882408, -0.17279188334941864, 0.02740040421485901, -0.1396256685256958, -0.06771225482225418, 0.13094839453697205, -0.00046136067248880863, -0.043535858392715454, 0.08554913848638535, 0.17515580356121063, -0.020696336403489113, 0.0467308945953846, -0.013883049599826336, -0.015717236325144768, -0.028536343947052956, 0.007802074775099754, -0.07747898250818253, 0.06667428463697433, 0.011438424699008465, -0.007018644828349352, -0.023928964510560036, 0.006724682170897722, -0.023113105446100235, -0.06457746028900146, 0.010403918102383614, 0.004679516423493624, -0.08277428895235062, 0.17236682772636414, -0.007042078301310539, 0.05172480270266533, 0.05193672701716423, 0.2806524634361267, -0.059840127825737, -0.14688602089881897, -0.0953894853591919, 0.09551454335451126, -0.015125514939427376, 0.03490113466978073, 0.048504408448934555, -0.0930115208029747, -0.08286205679178238, 0.16899777948856354, 0.12180694192647934, -0.0031824971083551645, -0.008518394082784653, -0.012296807020902634, -0.010159965604543686, 0.009608468972146511, 0.013331261463463306, 0.06857160478830338, 0.15728901326656342, -0.06832437217235565, -0.03965689241886139, -0.04407094046473503, -0.05732812359929085, -0.05839289352297783, 0.059074852615594864, 0.03846052289009094, -0.040598757565021515, -0.02231309376657009, 0.06928874552249908, -0.1812136471271515, -0.1058061346411705, -0.09863799810409546, -0.11456675082445145, -0.0958336815237999, -0.05866264924407005, 0.14592665433883667, 0.0012911694357171655, 0.010438065975904465, -0.01084001362323761, 0.02684042789041996, 0.018742535263299942, 0.04738948866724968, -0.14463385939598083, -0.005629054736346006, 0.12592527270317078, -0.17251531779766083, 0.13931682705879211, -0.05933979153633118, 0.050334252417087555, 0.11602021008729935, 0.019916517660021782, -0.07784684747457504, 0.010925709269940853, 0.03685368224978447, -0.12074092030525208, 0.012177356518805027, 0.1204194575548172, -0.005265932064503431, 0.0882653295993805, 0.012298076413571835, 0.02457420714199543, 0.0026938607916235924, -0.013008484616875648, -0.07269542664289474, -0.05722183361649513, -0.03522856906056404, -0.07554122060537338, 0.0550796203315258, 0.08292235434055328, -0.0620645172894001, 0.023428918793797493, -0.07604265213012695, 0.03808757662773132, -0.019980857148766518, -0.005230209790170193, -0.013927682302892208, -0.14331884682178497, -0.05797819048166275, 0.19749577343463898, 0.02257613092660904, -0.16501843929290771, 0.013788540847599506, -0.059680573642253876, -0.0009448782075196505, -0.14359667897224426, 0.09683133661746979, 0.08960016071796417, -0.03067745268344879, -0.0232603270560503, -0.0031418856233358383, -0.02218201942741871, 0.0856507271528244, -0.16613361239433289, -0.08888314664363861 ]
null
null
transformers
# BatteryOnlyBERT-cased for QA **Language model:** batteryonlybert-cased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 16 n_epochs = 3 base_LM_model = "batteryonlybert-cased" max_seq_len = 386 learning_rate = 2e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 79.61, "f1": 87.30, ``` Evaluated on the battery device dataset. ``` "precision": 64.28, "recall": 82.72, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batteryonlybert-cased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batteryonlybert-cased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatteryOnlyBERT-cased for QA Language model: batteryonlybert-cased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryOnlyBERT-cased for QA \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatteryOnlyBERT-cased for QA \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 68, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatteryOnlyBERT-cased for QA \r\nLanguage model: batteryonlybert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.08079832792282104, 0.032594382762908936, -0.0042159222066402435, 0.06078297272324562, 0.23093898594379425, 0.0355365164577961, 0.13760748505592346, 0.04351189360022545, -0.10658019781112671, -0.027120541781187057, 0.03539686277508736, -0.03221694007515907, 0.07133567333221436, 0.06533680111169815, 0.012910266406834126, -0.11539989709854126, 0.03759542852640152, -0.0379268117249012, 0.019749969244003296, 0.1463412493467331, 0.11707369983196259, -0.0851849764585495, 0.11293014138936996, 0.019384296610951424, -0.10665177553892136, 0.06460337340831757, -0.01533759105950594, -0.07954446226358414, 0.09784703701734543, -0.08263854682445526, 0.13242046535015106, 0.047301989048719406, 0.02964753471314907, -0.18714186549186707, 0.03735654428601265, 0.0306814294308424, 0.027095675468444824, 0.023025518283247948, 0.025998154655098915, -0.05170462280511856, 0.066691555082798, 0.021908648312091827, 0.04614482820034027, 0.020460540428757668, -0.05650586634874344, -0.2268374115228653, -0.05308942496776581, -0.05929781496524811, 0.08425483852624893, 0.1494562029838562, -0.03258039057254791, 0.08675552904605865, -0.0475301668047905, 0.07977107167243958, 0.16037458181381226, -0.10508308559656143, -0.0049460469745099545, 0.11248504370450974, 0.03895236551761627, -0.01433508563786745, -0.01532731018960476, -0.0802343413233757, -0.000274815276497975, 0.05506310611963272, 0.031095098704099655, -0.057953596115112305, -0.15440979599952698, 0.09692317992448807, -0.10497783869504929, -0.004525382071733475, 0.20683176815509796, -0.04831629619002342, -0.06784767657518387, -0.0030876800883561373, -0.12090664356946945, -0.02612810954451561, -0.019105050712823868, -0.009885524399578571, -0.02756435051560402, 0.0007443083450198174, -0.06325465440750122, 0.03066410683095455, -0.026638081297278404, -0.10176216065883636, -0.12652882933616638, 0.10276971012353897, 0.056452471762895584, 0.04214414209127426, -0.13448014855384827, 0.08198656141757965, 0.009651192463934422, -0.16329796612262726, -0.030714739114046097, -0.06350613385438919, -0.08054466545581818, 0.0044156527146697044, 0.0170882698148489, 0.00881289504468441, 0.12452861666679382, 0.08153615891933441, -0.03358064964413643, 0.01709078997373581, 0.021867280825972557, -0.06581823527812958, 0.01325095072388649, 0.14779694378376007, -0.02579174004495144, -0.13497795164585114, -0.011199094355106354, 0.030298318713903427, -0.0003181239007972181, -0.01311465259641409, -0.07242085784673691, -0.0034625015687197447, 0.05685891956090927, 0.06481598317623138, 0.01846739836037159, 0.030023418366909027, -0.056601621210575104, -0.0389319583773613, 0.03500058129429817, -0.1070956289768219, -0.04874262586236, -0.025186609476804733, 0.008690589107573032, 0.01449532713741064, -0.044212277978658676, 0.02162718027830124, -0.03961319103837013, -0.019482441246509552, -0.07246758788824081, -0.08444658666849136, -0.12832111120224, -0.08253360539674759, -0.036432623863220215, 0.13505825400352478, 0.0732189267873764, -0.11293298751115799, -0.3383442461490631, 0.07580526173114777, 0.005851006601005793, 0.0020896028727293015, -0.046371906995773315, -0.04198373481631279, -0.04512563347816467, -0.030580302700400352, -0.03321515768766403, 0.10240358859300613, -0.0640447810292244, 0.06021995469927788, 0.16314491629600525, 0.027552342042326927, -0.031446672976017, 0.04206860810518265, -0.14798450469970703, 0.003844690043479204, -0.13164857029914856, 0.04845106601715088, -0.1125897765159607, 0.0023374331649392843, -0.11580181866884232, -0.08375634998083115, 0.11937467008829117, -0.01230275072157383, 0.07767832279205322, 0.14008261263370514, -0.08903022110462189, -0.11607292294502258, 0.04632853716611862, -0.13571496307849884, -0.14253683388233185, 0.049382176250219345, -0.017439713701605797, 0.029389379546046257, 0.016971686854958534, 0.04972022771835327, 0.1347394436597824, -0.16925938427448273, -0.15315870940685272, -0.02082516811788082, 0.040212828665971756, -0.009352886117994785, 0.0832701101899147, 0.0373968705534935, -0.07201588153839111, 0.07285979390144348, -0.0636201798915863, 0.0502752885222435, -0.046370986849069595, -0.10651830583810806, -0.05507112666964531, -0.03341268002986908, 0.10437005013227463, 0.014805328100919724, -0.05406607314944267, 0.015918973833322525, -0.09670570492744446, 0.1304740309715271, 0.12405125796794891, -0.04814843460917473, -0.009350385516881943, -0.18601660430431366, 0.0863443985581398, -0.027847733348608017, 0.008318926207721233, -0.11694955080747604, -0.1722659170627594, 0.08174961805343628, -0.030531348660588264, -0.06440038233995438, -0.03679537773132324, 0.02148161642253399, 0.03263330087065697, -0.10121851414442062, -0.03403802961111069, -0.14210960268974304, -0.0218606349080801, 0.014943767338991165, -0.15894372761249542, 0.05097225680947304, -0.03359612077474594, 0.11403331160545349, 0.022190488874912262, 0.022577397525310516, 0.10974734276533127, 0.03173770383000374, -0.036686010658741, -0.012547914870083332, -0.015218342654407024, 0.0581774078309536, -0.03367162495851517, -0.03955908864736557, -0.030182356014847755, -0.014581733383238316, 0.005207172594964504, -0.04003360867500305, -0.12558111548423767, 0.059448983520269394, 0.14656880497932434, 0.1136440858244896, 0.009759584441781044, -0.09062782675027847, -0.0024159676395356655, -0.042100049555301666, -0.09666256606578827, -0.005692068487405777, 0.22466658055782318, -0.004308177623897791, 0.021082596853375435, -0.03605539724230766, 0.010046688839793205, -0.005104182753711939, 0.03671324998140335, 0.06096711382269859, 0.08892826735973358, 0.10724673420190811, 0.03870663791894913, 0.0647069439291954, 0.05630655214190483, -0.07142998278141022, 0.1338968127965927, -0.08327872306108475, -0.032247185707092285, -0.03992658853530884, -0.09741611033678055, -0.028694814071059227, 0.2653932571411133, -0.14537523686885834, 0.0658307820558548, 0.061967432498931885, -0.013649399392306805, 0.03338553011417389, -0.037293680012226105, -0.012343657203018665, -0.015851741656661034, -0.05428426340222359, -0.14742210507392883, 0.02511242963373661, 0.05536968633532524, 0.010547208599746227, -0.023618245497345924, -0.06436403840780258, 0.02023603580892086, 0.007767379749566317, -0.07597478479146957, 0.2337609976530075, -0.11558908224105835, -0.19947311282157898, -0.10426857322454453, -0.08470165729522705, -0.034664858132600784, -0.07503843307495117, 0.02129323035478592, -0.1267852634191513, -0.10010319203138351, 0.01846882700920105, 0.1753310263156891, 0.003057273803278804, 0.06582583487033844, 0.0423104465007782, 0.007985793985426426, 0.002066369168460369, -0.1144995465874672, 0.012564274482429028, 0.051118720322847366, -0.10289427638053894, 0.03459036350250244, 0.04997815936803818, 0.07507877796888351, 0.11748136579990387, 0.016201041638851166, 0.02044108882546425, -0.05081896856427193, 0.23376832902431488, -0.10234733670949936, -0.03442440181970596, 0.2916809916496277, 0.12642012536525726, -0.0002474411448929459, 0.20714536309242249, -0.03226645290851593, -0.018483199179172516, 0.10392570495605469, -0.008979320526123047, -0.051749520003795624, -0.25018030405044556, -0.08980153501033783, -0.03411375358700752, 0.05539553984999657, -0.06316763907670975, 0.013063405640423298, -0.038107745349407196, 0.08799491822719574, 0.012232379987835884, -0.02846408821642399, -0.09005258977413177, 0.0981193333864212, 0.23944616317749023, 0.05911460518836975, 0.0993240550160408, -0.0036623713094741106, 0.017191149294376373, 0.06357725709676743, 0.054353002458810806, 0.165835902094841, -0.0573110468685627, 0.03658246248960495, 0.03830937668681145, 0.14540709555149078, 0.03399444743990898, 0.02838371880352497, -0.011857684701681137, 0.010972507297992706, 0.021096549928188324, -0.04151267558336258, 0.01616997830569744, -0.024806911125779152, -0.03886384144425392, 0.01750815659761429, -0.00655366899445653, 0.16697585582733154, 0.04746263846755028, -0.01324431411921978, 0.020504357293248177, -0.20233379304409027, -0.027887960895895958, 0.04568798467516899, -0.035595741122961044, -0.10023748129606247, 0.03289346769452095, 0.014360289089381695, -0.12944211065769196, 0.04180008918046951, -0.02931041084229946, 0.06863526254892349, -0.06553041934967041, 0.048276450484991074, 0.06726309657096863, 0.12288596481084824, 0.015470224432647228, 0.1269751936197281, -0.3967536985874176, 0.10082846134901047, 0.04158995300531387, 0.08234966546297073, 0.030851105228066444, 0.0653144046664238, -0.05431403964757919, -0.04133213683962822, 0.0607638843357563, -0.0017841920489445329, 0.06654742360115051, 0.051276132464408875, -0.15409888327121735, 0.13401411473751068, 0.003125593066215515, 0.004254631698131561, 0.06727004051208496, -0.024824315682053566, 0.030364084988832474, -0.006071466486901045, -0.029444994404911995, -0.13759593665599823, -0.09557053446769714, 0.05240808427333832, -0.07132063060998917, 0.0956406518816948, -0.14502619206905365, -0.08506479114294052, 0.06429727375507355, 0.06824792176485062, -0.07356275618076324, -0.12468767911195755, -0.14294792711734772, 0.0034440357703715563, 0.11933700740337372, -0.1117333248257637, 0.0760037750005722, 0.0018976161954924464, 0.05549515038728714, -0.0261341854929924, -0.10801296681165695, 0.05149498209357262, -0.08984324336051941, -0.1412370502948761, -0.029128707945346832, 0.036941446363925934, 0.10416816920042038, 0.05216536298394203, 0.003756928024813533, -0.015420543029904366, -0.1423698216676712, -0.12380252033472061, -0.04021361842751503, -0.0469004325568676, 0.0366317518055439, 0.012856398709118366, 0.13401785492897034, -0.0319170206785202, -0.028258629143238068, -0.026041952893137932, 0.009008930064737797, 0.1436595320701599, -0.018159741535782814, 0.05359867215156555, 0.1247285008430481, -0.020956460386514664, -0.1948443204164505, -0.059099651873111725, 0.08181203156709671, 0.05386390537023544, -0.04074224457144737, -0.25878745317459106, 0.144019216299057, -0.024925552308559418, -0.066483274102211, -0.11063766479492188, -0.1449241191148758, -0.12149671465158463, 0.20142194628715515, 0.07077323645353317, 0.04922086000442505, -0.07190972566604614, -0.020911231637001038, -0.048562757670879364, -0.25042209029197693, 0.12776802480220795, -0.1516042947769165, 0.021062558516860008, -0.015312268398702145, 0.2335684895515442, -0.027750352397561073, -0.030407218262553215, 0.14143219590187073, 0.06754034012556076, -0.02282319776713848, 0.021803487092256546, 0.0015499709406867623, 0.12670259177684784, 0.032578837126493454, 0.048753783106803894, 0.004215157590806484, 0.1135249212384224, -0.1253320574760437, -0.042966704815626144, -0.057151395827531815, 0.09573708474636078, -0.050594571977853775, -0.16627167165279388, -0.09025583416223526, 0.04372143745422363, -0.03665262460708618, -0.06148229166865349, 0.07373202592134476, 0.04739926755428314, -0.01558538619428873, 0.10777931660413742, 0.16789725422859192, 0.044855307787656784, -0.06989837437868118, 0.03834351524710655, -0.009319382719695568, 0.057045526802539825, -0.14004522562026978, 0.025669243186712265, 0.10642336308956146, 0.029907258227467537, 0.017679724842309952, 0.04133428633213043, -0.08830884844064713, -0.005181066691875458, 0.07539393752813339, -0.15311746299266815, -0.15214058756828308, -0.07484232634305954, 0.17030493915081024, -0.19440264999866486, 0.05242626368999481, 0.10544019192457199, -0.007623409386724234, -0.07065687328577042, 0.02736532874405384, 0.04223370552062988, -0.009805819019675255, 0.17173267900943756, 0.08983146399259567, 0.13429582118988037, -0.13088752329349518, 0.056972213089466095, 0.039606865495443344, 0.034272920340299606, -0.03906606510281563, 0.030436620116233826, -0.09778415411710739, 0.01054996158927679, 0.029813706874847412, 0.12432511150836945, -0.13493098318576813, -0.08012329041957855, -0.15863396227359772, -0.1330799013376236, -0.004106585402041674, 0.074086032807827, 0.09248446673154831, 0.06592553853988647, -0.030644768849015236, 0.019045112654566765, -0.08427228778600693, 0.10721039026975632, 0.011834765784442425, 0.022708458825945854, -0.10814417153596878, -0.08740828931331635, -0.021881485357880592, 0.11539691686630249, -0.053284868597984314, -0.007938484661281109, -0.09265005588531494, 0.029003623872995377, -0.15287941694259644, -0.06395447254180908, -0.0335841067135334, -0.0157958772033453, 0.024567298591136932, -0.07458283007144928, -0.034987710416316986, 0.06919696182012558, -0.040420617908239365, 0.06010998785495758, 0.04246928170323372, 0.0216531939804554, -0.12978430092334747, 0.08935471624135971, 0.07418109476566315, -0.0573982410132885, 0.1172177642583847, 0.07257728278636932, 0.07057838141918182, 0.10881386697292328, -0.07202395051717758, 0.01928703673183918, 0.0025687520392239094, 0.08114410191774368, 0.07677510380744934, -0.14890287816524506, 0.0030335148330777884, -0.00850317720323801, 0.012839464470744133, 0.019171057268977165, 0.059905532747507095, -0.10405108332633972, 0.07196080684661865, -0.05237139016389847, 0.022019045427441597, -0.055838342756032944, 0.040064290165901184, 0.10123640298843384, 0.0788455531001091, 0.10663368552923203, -0.041202086955308914, -0.0008662017062306404, -0.1168544664978981, 0.009293470531702042, -0.027556877583265305, -0.01009521447122097, -0.042275406420230865, -0.04069266840815544, 0.07114720344543457, 0.03292621299624443, 0.14357943832874298, -0.007916372269392014, 0.061984527856111526, 0.01273760013282299, 0.12452498078346252, 0.09639617055654526, -0.005160865373909473, 0.0700279027223587, 0.08808089047670364, 0.0965600460767746, 0.12368548661470413, -0.028888355940580368, 0.06343398988246918, 0.1101451963186264, 0.11544614285230637, 0.06022394821047783, 0.05055898800492287, 0.04041652753949165, 0.05275730788707733, -0.10741817951202393, -0.0917084813117981, -0.006325290072709322, -0.07034683972597122, 0.014852326363325119, -0.007101766299456358, 0.1758139282464981, 0.08995579183101654, -0.11521881818771362, 0.08227045089006424, -0.0027805217541754246, -0.11475298553705215, -0.10720910876989365, 0.08857717365026474, -0.04414873570203781, -0.20821334421634674, 0.0342906191945076, -0.15816257894039154, -0.11414436250925064, 0.1806655377149582, 0.007774733938276768, -0.011549780145287514, 0.13474617898464203, 0.20982816815376282, -0.03602835536003113, -0.04220084473490715, -0.007876847870647907, 0.019595516845583916, 0.05335557088255882, 0.03401777148246765, -0.047316424548625946, 0.1045086607336998, 0.045306287705898285, 0.008511287160217762, -0.07480438798666, 0.012407304719090462, -0.004559525288641453, -0.027347031980752945, -0.01753895729780197, 0.004522108472883701, -0.07462240010499954, 0.1955731362104416, -0.0007433398859575391, 0.08297265321016312, 0.016621971502900124, 0.30614912509918213, -0.05107215791940689, -0.17817893624305725, -0.12875092029571533, 0.07495755702257156, -0.05951179563999176, 0.05601447448134422, 0.05923190712928772, -0.08328920602798462, -0.12701831758022308, 0.15549826622009277, 0.03441370278596878, -0.018483975902199745, -0.0458146408200264, 0.03429383039474487, -0.020003948360681534, -0.052904579788446426, 0.028207065537571907, 0.08170434087514877, 0.21016338467597961, -0.07398536056280136, -0.013562675565481186, -0.05466877296566963, -0.08194182813167572, -0.014344029128551483, 0.02322283759713173, 0.014900580048561096, -0.029651811346411705, -0.014193015173077583, 0.05060935392975807, -0.12016192078590393, -0.09642543643712997, -0.09794963151216507, -0.03154731169342995, -0.12224570661783218, -0.06431161612272263, 0.21647627651691437, 0.014310888014733791, 0.006114372052252293, -0.04451129585504532, 0.06953247636556625, 0.051492031663656235, 0.038633689284324646, -0.10471737384796143, -0.06385737657546997, 0.19038280844688416, -0.1522582471370697, 0.1414840966463089, -0.04870770126581192, 0.08581038564443588, 0.1137348935008049, -0.02360410802066326, -0.11870315670967102, 0.05607743188738823, 0.06337819993495941, -0.17054350674152374, 0.003488274058327079, 0.13163602352142334, 0.015695063397288322, 0.11118444800376892, -0.007936256006360054, 0.07699928432703018, -0.017150014638900757, 0.0016771518858149648, -0.03132211044430733, -0.05699727311730385, -0.018766416236758232, -0.07729919999837875, 0.04491473734378815, 0.07549454271793365, -0.06024898216128349, 0.06385073065757751, -0.07844389230012894, 0.0949530154466629, -0.03920212760567665, 0.08595450967550278, 0.021203622221946716, -0.11644932627677917, -0.06277406215667725, 0.1379457712173462, -0.05049370229244232, -0.11260455846786499, 0.012963601388037205, -0.001088994788005948, -0.0012889059726148844, -0.09567853808403015, 0.1406078189611435, 0.0688788965344429, -0.0029706712812185287, -0.02190021239221096, 0.04854986071586609, -0.058231696486473083, 0.06362414360046387, -0.1760292649269104, -0.09575653076171875 ]
null
null
transformers
# BatteryOnlyBERT-uncased for Battery Abstract Classification **Language model:** batteryonlybert-uncased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 16 n_epochs = 13 base_LM_model = "batteryonlybert-uncased" learning_rate = 3e-5 ``` ## Performance ``` "Validation accuracy": 97.18, "Test accuracy": 97.08, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batteryonlybert-uncased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batteryonlybert-uncased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatteryOnlyBERT-uncased for Battery Abstract Classification Language model: batteryonlybert-uncased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryOnlyBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatteryOnlyBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 75, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatteryOnlyBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.07742828875780106, 0.10324563086032867, -0.004303965251892805, 0.06040967255830765, 0.23968106508255005, 0.04231724143028259, 0.19550156593322754, 0.04532476142048836, -0.0928277000784874, -0.03630809858441353, 0.0477912463247776, 0.04150407761335373, 0.050924938172101974, 0.05722937360405922, -0.04931613802909851, -0.1593005359172821, 0.03393811732530594, -0.020234057679772377, 0.03483179956674576, 0.11443459242582321, 0.13344796001911163, -0.07543669641017914, 0.11243012547492981, -0.0020331775303930044, -0.10889507830142975, 0.023807981982827187, 0.012860985472798347, -0.12885993719100952, 0.11887551099061966, -0.06193253770470619, 0.15412302315235138, 0.0430913046002388, 0.025872882455587387, -0.15816538035869598, 0.03345577046275139, 0.015735451132059097, -0.006728474982082844, 0.07143739610910416, 0.023071931675076485, -0.08873412013053894, 0.10589144378900528, -0.02089519612491131, 0.07488304376602173, 0.026673490181565285, -0.0976005345582962, -0.1406627595424652, -0.041916925460100174, 0.0025973301380872726, 0.08126939833164215, 0.11356621980667114, -0.01419435441493988, 0.06420384347438812, -0.05805201455950737, 0.07228117436170578, 0.17512808740139008, -0.10152952373027802, 0.008174908347427845, 0.12260273098945618, -0.008146114647388458, -0.02478770539164543, -0.0598883256316185, -0.022920887917280197, 0.004467982333153486, 0.03969825804233551, 0.04488520696759224, -0.029340941458940506, -0.1622522622346878, 0.08833508938550949, -0.08981232345104218, -0.003964718896895647, 0.24842819571495056, -0.06938768923282623, -0.04042889550328255, -0.04566238448023796, -0.0937851220369339, -0.07421740144491196, -0.02980753779411316, 0.032867830246686935, -0.03552556410431862, 0.021421311423182487, -0.01484366413205862, 0.04618270322680473, -0.04762033373117447, -0.07351215183734894, -0.11494394391775131, 0.10171109437942505, 0.0044415914453566074, 0.0033435621298849583, -0.09579703211784363, 0.13534323871135712, 0.07117049396038055, -0.16463254392147064, 0.010867556557059288, -0.06363280117511749, -0.032875847071409225, -0.023650599643588066, 0.004444505553692579, -0.016253596171736717, 0.09713172912597656, 0.08713957667350769, -0.014826434664428234, -0.0017080380348488688, 0.005179545842111111, -0.01064522098749876, -0.000051883733249269426, 0.12199177592992783, -0.020800186321139336, -0.042036525905132294, 0.03572012484073639, 0.008432377129793167, 0.03912733122706413, -0.04560157656669617, -0.12643776834011078, -0.0065993331372737885, 0.08217720687389374, 0.07045096904039383, -0.02430603839457035, 0.05835307016968727, -0.08901619166135788, -0.06486058980226517, 0.07812952250242233, -0.1071191355586052, -0.022035513073205948, 0.013336967676877975, 0.020525053143501282, 0.029508549720048904, -0.003807969158515334, 0.024933427572250366, -0.025279389694333076, 0.05801628157496452, -0.08477472513914108, -0.05359030142426491, -0.09486673772335052, -0.0918254479765892, -0.0034402692690491676, 0.05848966911435127, 0.06540066003799438, -0.12018733471632004, -0.300601989030838, 0.060803551226854324, 0.007010300178080797, 0.0016310263890773058, -0.06849867850542068, -0.0779380351305008, -0.058908045291900635, 0.01756792515516281, 0.001008366234600544, 0.024049172177910805, -0.0640145093202591, 0.051946092396974564, 0.09790027141571045, 0.03707132861018181, -0.07066591084003448, 0.0423140749335289, -0.13549616932868958, -0.006109650246798992, -0.1772659420967102, 0.03378131985664368, -0.09609808027744293, 0.08713313192129135, -0.06388509273529053, -0.03729059174656868, 0.12073537707328796, -0.0272525567561388, 0.022549550980329514, 0.1578570157289505, -0.09499788284301758, -0.1454019695520401, 0.03892643004655838, -0.1324259340763092, -0.08951839059591293, 0.04984533414244652, -0.041418615728616714, 0.03420058265328407, 0.0649416372179985, 0.06415099650621414, 0.08918505162000656, -0.10517271608114243, -0.11332929134368896, 0.007593845948576927, 0.016750607639551163, 0.0005590187502093613, 0.09699123352766037, 0.035020679235458374, -0.04487817361950874, 0.06604393571615219, -0.01840684749186039, 0.06058786064386368, -0.035332709550857544, -0.0985196977853775, -0.02606181800365448, -0.032312821596860886, 0.13775722682476044, 0.024735715240240097, 0.007808479480445385, 0.0019605886191129684, -0.08938515186309814, 0.13842549920082092, 0.10479017347097397, -0.045058008283376694, 0.009666383266448975, -0.13484559953212738, 0.06332825124263763, 0.03764672204852104, 0.016467517241835594, -0.11180666089057922, -0.09877432882785797, 0.04376237839460373, 0.030989930033683777, -0.03517068922519684, -0.0744229108095169, 0.0009487670031376183, 0.040160469710826874, -0.0967043936252594, -0.020181044936180115, -0.08250979334115982, 0.00006759011739632115, -0.0007236259989440441, -0.2238125205039978, 0.08120770007371902, -0.03597404435276985, 0.1778215765953064, -0.10990817844867706, 0.06652273237705231, 0.1497916728258133, 0.07392662018537521, -0.017830779775977135, -0.008144229650497437, 0.014074327424168587, 0.03171480819582939, -0.06774059683084488, -0.04971317574381828, 0.004886846523731947, 0.0013202955015003681, -0.07368321716785431, 0.0016915961168706417, -0.14774107933044434, 0.09841347485780716, 0.16636380553245544, 0.009160919114947319, -0.03345245122909546, -0.03758580610156059, 0.0073429979383945465, -0.02444470301270485, -0.09409657120704651, 0.0024350362364202738, 0.21951143443584442, -0.01819070801138878, 0.07873362302780151, -0.03338810056447983, -0.009790973737835884, -0.00358374509960413, 0.030297676101326942, 0.04450037330389023, 0.11990740895271301, 0.08665947616100311, -0.05313878133893013, 0.09524814039468765, 0.052156560122966766, -0.07228671759366989, 0.1591194123029709, -0.06484556943178177, -0.01619006134569645, -0.00701409624889493, -0.09989361464977264, 0.0018562276381999254, 0.2073979526758194, -0.17791609466075897, 0.02108367718756199, 0.056533340364694595, -0.025915777310729027, 0.0178457573056221, -0.07257543504238129, -0.024271171540021896, -0.007102354429662228, -0.020493529736995697, -0.11337056010961533, 0.010443250648677349, 0.03859097510576248, 0.05987334996461868, -0.01500266045331955, -0.08346651494503021, 0.07715798914432526, 0.009620719589293003, -0.10117003321647644, 0.2319374680519104, -0.12391116470098495, -0.24701814353466034, -0.12869864702224731, -0.07273482531309128, -0.032792288810014725, -0.00874626450240612, 0.03153980150818825, -0.10897742211818695, -0.09346380084753036, -0.0059547643177211285, 0.0937524363398552, -0.01998664252460003, 0.07295046001672745, 0.07049816101789474, 0.01836707815527916, -0.022618573158979416, -0.10860297083854675, 0.008205350488424301, 0.037528883665800095, -0.05681682005524635, 0.04885551705956459, 0.03399426490068436, 0.04452410340309143, 0.18401333689689636, 0.01084554847329855, 0.01868365705013275, -0.05004536360502243, 0.13701434433460236, -0.10737544298171997, -0.04060080274939537, 0.19941313564777374, 0.04011710360646248, 0.002558936597779393, 0.16453701257705688, 0.005084905307739973, -0.02769923396408558, 0.0797516331076622, -0.025914711877703667, -0.04959281161427498, -0.2684744596481323, -0.12414159625768661, -0.06036056950688362, 0.10016721487045288, -0.031889140605926514, 0.034097276628017426, 0.042111653834581375, 0.0531601719558239, -0.0025927708484232426, 0.07140082120895386, -0.03203006833791733, 0.12128227204084396, 0.2308507263660431, 0.062214549630880356, 0.08013121783733368, -0.026114201173186302, -0.02517610602080822, 0.06727135181427002, -0.04539759084582329, 0.1299537569284439, -0.006865158211439848, 0.08210156857967377, 0.009620899334549904, 0.00974938552826643, 0.06457912921905518, 0.04454470053315163, 0.006163329817354679, -0.002356762532144785, 0.02335502579808235, -0.04483373835682869, 0.03930382803082466, -0.02028285712003708, -0.09720945358276367, -0.004094834439456463, -0.05174441635608673, 0.09296847134828568, 0.08073285967111588, -0.05209621787071228, 0.03531567007303238, -0.2245289534330368, -0.056634753942489624, 0.05196910351514816, -0.044019587337970734, -0.09471035748720169, 0.021425146609544754, -0.014833367429673672, -0.12018357217311859, 0.07271962612867355, -0.012529382482171059, 0.07226419448852539, -0.12422453612089157, 0.056907180696725845, 0.0273381769657135, 0.06272516399621964, -0.01422381866723299, 0.12051449716091156, -0.3547531068325043, 0.12494765222072601, 0.05266130343079567, 0.06067907065153122, -0.01845044456422329, 0.03695562854409218, -0.00483182817697525, 0.05849263817071915, 0.028929555788636208, 0.0005984784220345318, 0.04895024001598358, 0.03282925486564636, -0.14042846858501434, 0.055148862302303314, 0.014447938650846481, -0.05497949197888374, 0.025027301162481308, -0.00807284377515316, 0.02195797860622406, 0.027619294822216034, -0.06739505380392075, -0.1348826289176941, -0.1274966299533844, 0.09065208584070206, -0.025795690715312958, 0.11371206492185593, -0.10255295038223267, -0.09779521077871323, 0.003272890578955412, 0.12461178004741669, -0.14766110479831696, -0.10474582761526108, -0.14068515598773956, -0.00010238102549919859, 0.04728217422962189, -0.07235340029001236, 0.06435775756835938, -0.0017090466571971774, 0.07317036390304565, -0.0478396899998188, -0.10213612020015717, 0.1024542823433876, -0.0594441182911396, -0.1791636049747467, -0.05717318877577782, 0.0852792039513588, 0.0998181700706482, 0.05757240206003189, -0.01704183965921402, -0.004400061909109354, -0.0925794392824173, -0.1244317814707756, 0.0008375702309422195, 0.01041122991591692, 0.050415292382240295, 0.04787570238113403, 0.06859865039587021, -0.09247086942195892, -0.025351447984576225, -0.06923480331897736, 0.07829825580120087, 0.1880590170621872, -0.01942496933043003, 0.10607773065567017, 0.08598724752664566, -0.06125722452998161, -0.23729974031448364, -0.04488563910126686, 0.0034464120399206877, 0.032430488616228104, -0.053924914449453354, -0.27624231576919556, 0.12382345646619797, -0.019732553511857986, -0.06330577284097672, -0.017827942967414856, -0.145025834441185, -0.135559543967247, 0.21417446434497833, 0.01308252103626728, 0.10633775591850281, -0.08856787532567978, -0.038914017379283905, -0.09345299005508423, -0.18305474519729614, 0.21703264117240906, -0.16738364100456238, 0.05871986597776413, -0.01883973367512226, 0.17274615168571472, -0.009847289882600307, -0.03109191544353962, 0.1359236240386963, 0.04776555299758911, 0.011359869502484798, -0.017331520095467567, -0.09033562988042831, 0.2090655118227005, 0.023676609620451927, 0.01538318395614624, -0.05109883099794388, 0.05538744851946831, -0.11643673479557037, -0.03953376039862633, -0.016371918842196465, 0.11984343081712723, -0.03513520956039429, -0.14866773784160614, -0.09491162747144699, 0.022484609857201576, -0.026746977120637894, -0.04098546877503395, 0.19436094164848328, 0.06712186336517334, 0.02175271324813366, 0.043166808784008026, 0.21338988840579987, 0.010725448839366436, 0.06196490302681923, 0.0026741335168480873, -0.020227063447237015, 0.0616469569504261, -0.18737557530403137, 0.026204368099570274, 0.08812981843948364, 0.04457690194249153, 0.04440607130527496, 0.08837894350290298, -0.06118982657790184, -0.019938891753554344, 0.10059955716133118, -0.2232953906059265, -0.02072034776210785, -0.09962177276611328, 0.1486693024635315, -0.13692022860050201, 0.08682414144277573, 0.11441676318645477, -0.030527180060744286, -0.05416201055049896, 0.049789272248744965, 0.03925779089331627, -0.02919599786400795, 0.12442424893379211, 0.05858854576945305, 0.10800504684448242, -0.1335812658071518, 0.07978937029838562, 0.07921776920557022, 0.02945626899600029, -0.026335150003433228, 0.07964945584535599, -0.1350202113389969, -0.02160642296075821, -0.013402760960161686, 0.20335666835308075, -0.11319668591022491, -0.046294890344142914, -0.18578006327152252, -0.1305524706840515, 0.019211310893297195, 0.13174307346343994, 0.10796276479959488, 0.10792611539363861, -0.055892832577228546, 0.01079384796321392, -0.0969073623418808, 0.0890381932258606, 0.0035147694870829582, 0.02886619046330452, -0.14874106645584106, 0.011081214994192123, -0.013357103802263737, 0.041574493050575256, -0.05514593794941902, -0.008529066108167171, -0.14486613869667053, 0.014074532315135002, -0.05915527418255806, -0.054437682032585144, -0.026860179379582405, -0.01500800158828497, 0.008877890184521675, -0.03509177640080452, -0.01262037456035614, 0.03273199871182442, -0.03925542160868645, 0.0496332049369812, 0.02796243317425251, 0.06315743923187256, -0.10595093667507172, 0.05845663323998451, 0.060565799474716187, -0.06538049876689911, 0.1382027417421341, 0.09120133519172668, 0.07048200070858002, 0.0857912078499794, -0.16859912872314453, 0.027703048661351204, 0.04592255502939224, 0.03963613137602806, 0.03713817149400711, -0.12018413096666336, -0.010123138315975666, -0.011315117590129375, 0.021684441715478897, 0.053154777735471725, 0.06159503385424614, -0.10328636318445206, 0.07314082235097885, -0.04153832793235779, -0.005559384822845459, -0.0637403354048729, 0.036282747983932495, 0.1314515620470047, 0.040304526686668396, 0.12392407655715942, -0.06320027261972427, 0.002790858270600438, -0.07948140054941177, 0.026365526020526886, -0.01982930488884449, -0.06783347576856613, -0.09361320734024048, -0.03324144333600998, 0.052453190088272095, 0.005081431474536657, 0.198076069355011, 0.009850706905126572, -0.037896350026130676, 0.044977542012929916, 0.16464759409427643, 0.057809703052043915, 0.029350878670811653, 0.11684196442365646, 0.1058303713798523, 0.04793296754360199, 0.07839125394821167, -0.008977401070296764, 0.11999362707138062, 0.07994993031024933, 0.14060868322849274, 0.024467285722494125, -0.0095113730058074, 0.04033634811639786, 0.01464287843555212, -0.09352853149175644, -0.13926298916339874, 0.04400321841239929, -0.05284704640507698, 0.039703369140625, -0.041494086384773254, 0.15578697621822357, 0.11297686398029327, -0.08577308803796768, 0.050010230392217636, -0.011937631294131279, -0.10700672119855881, -0.10351834446191788, -0.006982486695051193, -0.03630686178803444, -0.1705695390701294, 0.026544881984591484, -0.13364212214946747, -0.07593736052513123, 0.13099798560142517, -0.01141299493610859, -0.03942800313234329, 0.0933491587638855, 0.17119677364826202, -0.012962634675204754, 0.040928877890110016, -0.018456798046827316, -0.03213943913578987, -0.023138146847486496, -0.0009709727601148188, -0.06676524132490158, 0.07414525747299194, 0.016020571812987328, -0.013042937032878399, -0.01804879866540432, 0.010457797907292843, -0.016213420778512955, -0.06501642614603043, 0.010967413894832134, 0.0009125073556788266, -0.08372332155704498, 0.17182081937789917, -0.0023473750334233046, 0.051289789378643036, 0.049107227474451065, 0.2835136950016022, -0.06205352023243904, -0.15333832800388336, -0.09170160442590714, 0.10525859147310257, -0.014974440447986126, 0.036355145275592804, 0.05214385688304901, -0.10666593164205551, -0.08290129899978638, 0.14737169444561005, 0.1156521886587143, -0.007629877422004938, -0.007607945706695318, -0.0041091470047831535, -0.012703520245850086, -0.007816976867616177, 0.01309028547257185, 0.06039116531610489, 0.14731118083000183, -0.06031803414225578, -0.03873318061232567, -0.041262928396463394, -0.056613121181726456, -0.06190785393118858, 0.057504162192344666, 0.03960561752319336, -0.03425104543566704, -0.021112149581313133, 0.06172191724181175, -0.17759256064891815, -0.0976109653711319, -0.09409679472446442, -0.10306432098150253, -0.09341508150100708, -0.06775679439306259, 0.1551666259765625, -0.0042135692201554775, 0.013645111583173275, -0.01775306835770607, 0.032680630683898926, 0.015313546173274517, 0.04150504618883133, -0.1471063792705536, -0.00862723495811224, 0.11727510392665863, -0.16456907987594604, 0.1395336538553238, -0.05651415139436722, 0.06302065402269363, 0.11531201750040054, 0.01498119905591011, -0.08047393709421158, 0.02723648212850094, 0.03347290679812431, -0.11992653459310532, 0.014364805072546005, 0.12060929834842682, -0.0066644358448684216, 0.10291768610477448, 0.005068082828074694, 0.023191535845398903, 0.00022947753313928843, 0.007186071015894413, -0.07889655977487564, -0.05339137464761734, -0.035023175179958344, -0.06313513219356537, 0.05895208567380905, 0.07194880396127701, -0.062111761420965195, 0.019164247438311577, -0.0710490494966507, 0.03971773758530617, -0.01213068701326847, -0.008862276561558247, -0.017273567616939545, -0.12399646639823914, -0.052710775285959244, 0.18322505056858063, 0.023901687934994698, -0.16585223376750946, 0.01842532306909561, -0.050738733261823654, -0.0023110194597393274, -0.1425197422504425, 0.0919123962521553, 0.07792460918426514, -0.02891632728278637, -0.024804681539535522, 0.010814334265887737, -0.02516096457839012, 0.07098999619483948, -0.16435204446315765, -0.0983770489692688 ]
null
null
transformers
# BatteryOnlyBERT-uncased for QA **Language model:** batteryonlybert-uncased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 16 n_epochs = 2 base_LM_model = "batteryonlybert-uncased" max_seq_len = 386 learning_rate = 2e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 79.53, "f1": 87.22, ``` Evaluated on the battery device dataset. ``` "precision": 67.20, "recall": 83.82, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batteryonlybert-uncased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batteryonlybert-uncased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatteryOnlyBERT-uncased for QA Language model: batteryonlybert-uncased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatteryOnlyBERT-uncased for QA \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatteryOnlyBERT-uncased for QA \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 70, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatteryOnlyBERT-uncased for QA \r\nLanguage model: batteryonlybert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.07915658503770828, 0.04639317840337753, -0.004253112245351076, 0.0666617825627327, 0.22728046774864197, 0.04471142590045929, 0.14725103974342346, 0.043492503464221954, -0.09297074377536774, -0.027500653639435768, 0.03533560782670975, -0.040488943457603455, 0.07112124562263489, 0.06418238580226898, 0.015294237062335014, -0.10378307849168777, 0.03585332632064819, -0.040270086377859116, 0.05269744247198105, 0.13700202107429504, 0.11054771393537521, -0.08317040652036667, 0.10975746065378189, 0.014754255302250385, -0.09952733665704727, 0.07087180763483047, -0.030588248744606972, -0.06900492310523987, 0.09058455377817154, -0.10122814774513245, 0.14337152242660522, 0.04222875088453293, 0.034875281155109406, -0.18706129491329193, 0.03162911906838417, 0.025281978771090508, 0.03236456587910652, 0.020449569448828697, 0.02989167906343937, -0.05359511822462082, 0.03311951085925102, 0.027322890236973763, 0.053316425532102585, 0.009286236017942429, -0.05535442382097244, -0.21047766506671906, -0.053312983363866806, -0.06493043154478073, 0.09089891612529755, 0.1526564359664917, -0.032891612499952316, 0.08717784285545349, -0.04942221939563751, 0.08042214065790176, 0.16314971446990967, -0.11001113057136536, -0.004986349493265152, 0.11564488708972931, 0.02800811640918255, -0.03735772520303726, -0.012735195457935333, -0.09367428719997406, 0.004205271601676941, 0.05602116510272026, 0.0157866682857275, -0.051747675985097885, -0.15577661991119385, 0.08546262979507446, -0.10426917672157288, -0.0035683251917362213, 0.206675723195076, -0.048852887004613876, -0.07132409512996674, 0.004500322509557009, -0.11856932938098907, -0.03137785941362381, -0.016014764085412025, 0.00039175126585178077, -0.028204284608364105, 0.0029903091490268707, -0.06631150841712952, 0.03498581424355507, -0.026179153472185135, -0.10665840655565262, -0.11784399300813675, 0.09885392338037491, 0.06921644508838654, 0.04226279631257057, -0.12991921603679657, 0.0921003669500351, 0.008558433502912521, -0.1656872034072876, -0.027670981362462044, -0.058099742978811264, -0.07283952087163925, -0.0029753425624221563, 0.01708177477121353, 0.009255679324269295, 0.12045814096927643, 0.08554112911224365, -0.05228282883763313, 0.016520686447620392, 0.02860666625201702, -0.06296541541814804, 0.019009854644536972, 0.12934748828411102, -0.023938290774822235, -0.1355656087398529, -0.02044336125254631, 0.03158897906541824, 0.00037245743442326784, -0.005424713250249624, -0.06780003756284714, 0.0054898252710700035, 0.062222886830568314, 0.060806550085544586, 0.008048424497246742, 0.022001517936587334, -0.06478380411863327, -0.038642048835754395, 0.0224283579736948, -0.10524331033229828, -0.04159800335764885, -0.030665917322039604, 0.004477023147046566, 0.014426449313759804, -0.03523426875472069, 0.019108450040221214, -0.04896269366145134, -0.01104745827615261, -0.0682232603430748, -0.08490461856126785, -0.13430973887443542, -0.07930045574903488, -0.04497869685292244, 0.12129081040620804, 0.07567210495471954, -0.10786272585391998, -0.3570040762424469, 0.07665932178497314, 0.012916609644889832, -0.005153039004653692, -0.03672855347394943, -0.05617377161979675, -0.0379452221095562, -0.032054342329502106, -0.028987886384129524, 0.08246791362762451, -0.06314365565776825, 0.05177673324942589, 0.1561255306005478, 0.028836790472269058, -0.01970822922885418, 0.03874574974179268, -0.14764827489852905, 0.0033412990160286427, -0.11503587663173676, 0.041979383677244186, -0.10942453891038895, 0.007636824157088995, -0.11690644174814224, -0.08978119492530823, 0.10959915071725845, -0.012502326630055904, 0.08933817595243454, 0.1403101235628128, -0.08215449750423431, -0.10927405953407288, 0.05727892741560936, -0.13348881900310516, -0.13090269267559052, 0.045687850564718246, -0.017887776717543602, 0.04524245113134384, 0.017721697688102722, 0.058208730071783066, 0.1112714633345604, -0.1694357693195343, -0.15672999620437622, -0.021576903760433197, 0.050635527819395065, -0.004050364252179861, 0.07458891719579697, 0.03481295704841614, -0.0782650038599968, 0.07037726789712906, -0.04459108039736748, 0.04000740125775337, -0.05128242447972298, -0.10677304118871689, -0.05005470663309097, -0.041589368134737015, 0.10178559273481369, 0.01126850675791502, -0.04652765020728111, 0.00965800229460001, -0.09815887361764908, 0.13949398696422577, 0.12821047008037567, -0.05716969072818756, -0.009727461263537407, -0.18884487450122833, 0.07851362973451614, -0.02739490568637848, 0.004276107531040907, -0.10596515238285065, -0.16492950916290283, 0.09112567454576492, -0.0401206910610199, -0.07506439089775085, -0.03536810353398323, 0.02299930900335312, 0.03997061029076576, -0.10210013389587402, -0.0394914336502552, -0.12814712524414062, -0.02285894937813282, 0.015010681934654713, -0.16736769676208496, 0.05432707816362381, -0.03272650018334389, 0.10571041703224182, 0.02962113730609417, 0.0221379604190588, 0.08891202509403229, 0.03794834762811661, -0.048318758606910706, -0.018611567094922066, -0.009836369194090366, 0.053610660135746, -0.029912345111370087, -0.04190061241388321, -0.027098102495074272, -0.009077479131519794, 0.01606946438550949, -0.06099895015358925, -0.12282607704401016, 0.036485057324171066, 0.14350934326648712, 0.11160249263048172, -0.0034427817445248365, -0.06705246865749359, 0.007059767842292786, -0.049738746136426926, -0.10058493912220001, 0.0017586908070370555, 0.23423054814338684, -0.007328323554247618, 0.02254031039774418, -0.033250730484724045, 0.0047007277607917786, -0.01265134010463953, 0.02830057218670845, 0.06011662259697914, 0.08241312205791473, 0.08465145528316498, 0.022499995306134224, 0.05355582386255264, 0.04836297780275345, -0.08725279569625854, 0.1305539608001709, -0.06758218258619308, -0.03101804293692112, -0.04326440393924713, -0.09037125110626221, -0.03109545074403286, 0.2623918354511261, -0.15442059934139252, 0.07228979468345642, 0.054285816848278046, -0.01686297170817852, 0.03481842204928398, -0.03532779961824417, -0.014271591790020466, -0.015344507992267609, -0.04981459677219391, -0.15003907680511475, 0.015403962694108486, 0.04774293676018715, 0.005382854491472244, -0.021351415663957596, -0.0682038813829422, 0.02048470638692379, 0.006276516709476709, -0.07361141592264175, 0.23370836675167084, -0.11332473903894424, -0.18479028344154358, -0.09638381004333496, -0.08682417124509811, -0.02190619893372059, -0.07611662149429321, 0.02679450996220112, -0.1270313411951065, -0.10360302776098251, 0.014374911785125732, 0.15643152594566345, -0.003141437191516161, 0.06666497886180878, 0.06370440870523453, 0.0009375347290188074, 0.00462074251845479, -0.10744292289018631, 0.014638769440352917, 0.05441225692629814, -0.11639489978551865, 0.034677084535360336, 0.047486066818237305, 0.06420578062534332, 0.11153693497180939, 0.019038060680031776, 0.027119269594550133, -0.044226497411727905, 0.23767291009426117, -0.0888645127415657, -0.03582116216421127, 0.27765989303588867, 0.12557947635650635, 0.004202732816338539, 0.20602473616600037, -0.033068347722291946, -0.020519206300377846, 0.10209756344556808, -0.006833864375948906, -0.04634426534175873, -0.2508659064769745, -0.08200491964817047, -0.025985604152083397, 0.05231388285756111, -0.04076235368847847, 0.011964094825088978, -0.032335393130779266, 0.08753059804439545, 0.011904546990990639, -0.01968676783144474, -0.09184931963682175, 0.1067991629242897, 0.21632665395736694, 0.06255470216274261, 0.09444551169872284, -0.0000653246243018657, 0.023606695234775543, 0.062120821326971054, 0.05200986564159393, 0.16086164116859436, -0.05925920978188515, 0.029063578695058823, 0.03811714053153992, 0.1415213644504547, 0.022704798728227615, 0.031096119433641434, -0.02355007827281952, 0.016708746552467346, 0.024903589859604836, -0.03706490248441696, 0.02688569761812687, -0.030624186620116234, -0.047967541962862015, 0.022440139204263687, -0.003536858595907688, 0.165458545088768, 0.04028753563761711, -0.007057007402181625, 0.03722786158323288, -0.20588624477386475, -0.02723710797727108, 0.0457359217107296, -0.033646903932094574, -0.10331505537033081, 0.03403462469577789, 0.013268707320094109, -0.122409887611866, 0.04295884817838669, -0.016981085762381554, 0.06413809210062027, -0.072711281478405, 0.04854026064276695, 0.08308249711990356, 0.11907371878623962, 0.015938319265842438, 0.12255315482616425, -0.3849460184574127, 0.09311731159687042, 0.041242972016334534, 0.09696286171674728, 0.03915368765592575, 0.059788525104522705, -0.05794334411621094, -0.05815186724066734, 0.06913946568965912, -0.0002442162367515266, 0.0842939019203186, 0.06269848346710205, -0.16054074466228485, 0.1182263195514679, -0.005788030102849007, 0.006232115440070629, 0.06130396947264671, -0.022584741935133934, 0.029745126143097878, -0.004902092274278402, -0.04536935314536095, -0.1470848023891449, -0.0844525694847107, 0.05266815423965454, -0.06547290831804276, 0.09176905453205109, -0.13759702444076538, -0.08367283642292023, 0.06918291002511978, 0.07472886890172958, -0.08115241676568985, -0.1267244964838028, -0.13109782338142395, 0.003935558255761862, 0.11952459812164307, -0.10557380318641663, 0.07604483515024185, 0.002542180707678199, 0.045698586851358414, -0.02895711176097393, -0.1049911230802536, 0.04313638433814049, -0.09124244004487991, -0.1327347606420517, -0.03137531504034996, 0.029032183811068535, 0.11292214691638947, 0.05019070580601692, 0.00646218890324235, -0.01541023887693882, -0.14961516857147217, -0.11819271743297577, -0.03473581001162529, -0.05189267173409462, 0.049562256783246994, 0.014409085735678673, 0.12713289260864258, -0.024441827088594437, -0.02582060545682907, -0.025422951206564903, 0.0030929555650800467, 0.15285564959049225, -0.013140667229890823, 0.04325828701257706, 0.13872970640659332, -0.027420612052083015, -0.19120706617832184, -0.07110533863306046, 0.09557497501373291, 0.04655537009239197, -0.05473809316754341, -0.2684873640537262, 0.14491239190101624, -0.023725828155875206, -0.06234396621584892, -0.1181236058473587, -0.14199884235858917, -0.12065508216619492, 0.20947188138961792, 0.07641072571277618, 0.0701141208410263, -0.0652572363615036, -0.01768268644809723, -0.049548372626304626, -0.23302319645881653, 0.12441498041152954, -0.14175193011760712, 0.024611061438918114, -0.01449919305741787, 0.2432546317577362, -0.029424479231238365, -0.035906873643398285, 0.1374271810054779, 0.0657571330666542, -0.023240815848112106, 0.016687674447894096, 0.004149637650698423, 0.11961714923381805, 0.02498646453022957, 0.062305744737386703, 0.02126912586390972, 0.11430880427360535, -0.10872725397348404, -0.042978685349226, -0.05248222500085831, 0.0900566577911377, -0.054258737713098526, -0.1641688048839569, -0.07877562940120697, 0.04665021598339081, -0.029262131080031395, -0.05523112043738365, 0.07086829096078873, 0.06060389429330826, -0.023554787039756775, 0.08864979445934296, 0.1646157205104828, 0.05797455459833145, -0.06580439954996109, 0.04071946069598198, -0.011059162206947803, 0.0546131394803524, -0.12608802318572998, 0.02176031656563282, 0.10103266686201096, 0.029915468767285347, 0.012742307037115097, 0.04511045664548874, -0.08991726487874985, -0.014630910940468311, 0.0753411203622818, -0.15985293686389923, -0.13346326351165771, -0.06794912368059158, 0.18558940291404724, -0.18916644155979156, 0.050998203456401825, 0.1009582057595253, -0.005216788500547409, -0.06883060932159424, 0.025092581287026405, 0.04623371735215187, -0.013337030075490475, 0.1751277595758438, 0.07829660177230835, 0.13137461245059967, -0.12658433616161346, 0.06994446367025375, 0.046705376356840134, 0.055611371994018555, -0.027576539665460587, 0.030264761298894882, -0.09680234640836716, 0.0069877104833722115, 0.030264705419540405, 0.1339246779680252, -0.1326870322227478, -0.07688219100236893, -0.16300047934055328, -0.11448725312948227, -0.0036768715362995863, 0.06520383805036545, 0.0928969532251358, 0.06694389879703522, -0.027615299448370934, 0.0168549083173275, -0.08917170763015747, 0.10314849764108658, 0.02067807875573635, 0.02140604704618454, -0.1076730340719223, -0.09690312296152115, -0.015737328678369522, 0.10595421493053436, -0.0479143001139164, -0.00722341425716877, -0.09898960590362549, 0.03487570583820343, -0.14037470519542694, -0.05751214176416397, -0.034294869750738144, -0.021927746012806892, 0.027807161211967468, -0.0687728002667427, -0.02942151203751564, 0.0700104758143425, -0.03619864955544472, 0.06210298091173172, 0.04811748117208481, 0.018988383933901787, -0.1243976578116417, 0.08938661217689514, 0.06770817190408707, -0.06027993559837341, 0.10640385746955872, 0.08231101185083389, 0.07415367662906647, 0.09570810198783875, -0.0707656592130661, 0.013978630304336548, -0.009160136803984642, 0.07716399431228638, 0.07210841029882431, -0.15101312100887299, -0.004715184215456247, -0.010390078648924828, 0.009677404537796974, 0.01643441990017891, 0.06855038553476334, -0.09723953902721405, 0.08212458342313766, -0.04751184955239296, 0.023883577436208725, -0.06382269412279129, 0.03841023147106171, 0.11328393220901489, 0.08239386230707169, 0.10519835352897644, -0.04515695199370384, 0.001257779891602695, -0.10424686968326569, 0.005209712777286768, -0.018819669261574745, -0.005924045108258724, -0.04569217935204506, -0.040762096643447876, 0.06776029616594315, 0.0317503996193409, 0.13329772651195526, -0.017420675605535507, 0.05250343307852745, 0.0031172235030680895, 0.1143452525138855, 0.0865430161356926, -0.006037786602973938, 0.06664182245731354, 0.07556669414043427, 0.09637106955051422, 0.1148761585354805, -0.030608315020799637, 0.06723172217607498, 0.10685674101114273, 0.11300566792488098, 0.0556938573718071, 0.056597236543893814, 0.043464142829179764, 0.037177249789237976, -0.11798218637704849, -0.09776525944471359, 0.009198389016091824, -0.05984965339303017, 0.018107229843735695, -0.0026625932659953833, 0.17568987607955933, 0.07720425724983215, -0.11859457939863205, 0.07791153341531754, -0.0012475784169510007, -0.12343893200159073, -0.10713658481836319, 0.07584261149168015, -0.03997267782688141, -0.19558085501194, 0.039297912269830704, -0.15875054895877838, -0.10880681127309799, 0.1748051941394806, 0.015626806765794754, -0.005851869471371174, 0.14126336574554443, 0.20300665497779846, -0.021667372435331345, -0.04988855868577957, -0.009198972024023533, 0.018704691901803017, 0.05050502344965935, 0.030973833054304123, -0.05304782837629318, 0.11800781637430191, 0.049739640206098557, 0.003405733034014702, -0.07559618353843689, 0.005809694994240999, 0.013761406764388084, -0.02856583334505558, -0.014165951870381832, 0.00787016935646534, -0.07190655171871185, 0.1921728104352951, -0.0010997859062626958, 0.08139815181493759, 0.012658927589654922, 0.29955798387527466, -0.05567503347992897, -0.18455475568771362, -0.12587857246398926, 0.06505739688873291, -0.05097334831953049, 0.05165138840675354, 0.06041934713721275, -0.08698984235525131, -0.1257549226284027, 0.1524151712656021, 0.029650911688804626, -0.013772029429674149, -0.0520489402115345, 0.036446865648031235, -0.020865073427557945, -0.05321899801492691, 0.017104357481002808, 0.07768929749727249, 0.19285495579242706, -0.08050210028886795, -0.0015561311738565564, -0.05420440435409546, -0.08494096994400024, -0.02979617565870285, 0.010646125301718712, 0.014595849439501762, -0.0263934638351202, -0.019052185118198395, 0.05050719901919365, -0.11620030552148819, -0.10005676001310349, -0.08647029101848602, -0.03119480237364769, -0.12164685130119324, -0.05655593052506447, 0.2282014936208725, 0.005882392171770334, 0.004293609876185656, -0.052178144454956055, 0.0771031528711319, 0.051737602800130844, 0.03484755754470825, -0.09382162243127823, -0.07390882819890976, 0.17603963613510132, -0.14755550026893616, 0.15834569931030273, -0.045022204518318176, 0.08972525596618652, 0.11036854237318039, -0.030666230246424675, -0.11513353139162064, 0.04949752613902092, 0.061756353825330734, -0.17002493143081665, 0.002596004866063595, 0.14383023977279663, 0.001335250330157578, 0.12286621332168579, -0.01604459248483181, 0.07203031331300735, -0.024876611307263374, 0.018428228795528412, -0.03938203305006027, -0.05292502045631409, -0.020585941150784492, -0.07118278741836548, 0.05159679800271988, 0.07600083202123642, -0.06702297925949097, 0.05790898948907852, -0.07127480953931808, 0.1012267991900444, -0.038406480103731155, 0.07998587191104889, 0.02082209102809429, -0.11403408646583557, -0.056393709033727646, 0.12937496602535248, -0.03196026757359505, -0.10761979222297668, 0.016304628923535347, 0.0025568287819623947, -0.004987500142306089, -0.10679670423269272, 0.13635630905628204, 0.06400391459465027, 0.005393065977841616, -0.022326823323965073, 0.05248711258172989, -0.05891186371445656, 0.05300838500261307, -0.16905392706394196, -0.10612320899963379 ]
null
null
transformers
# BatterySciBERT-cased for Battery Abstract Classification **Language model:** batteryscibert-cased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 11 base_LM_model = "batteryscibert-cased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 97.06, "Test accuracy": 97.19, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batteryscibert-cased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batteryscibert-cased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatterySciBERT-cased for Battery Abstract Classification Language model: batteryscibert-cased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatterySciBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 72, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatterySciBERT-cased for Battery Abstract Classification \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.06472767144441605, 0.05003158003091812, -0.004300367087125778, 0.0496634878218174, 0.2479923516511917, 0.03231930360198021, 0.19530346989631653, 0.030132675543427467, -0.09084128588438034, -0.029783716425299644, 0.04104619845747948, 0.04150325804948807, 0.05602128803730011, 0.05864802747964859, -0.04158348962664604, -0.17187824845314026, 0.03024134784936905, -0.004030480980873108, 0.029881933704018593, 0.1242157444357872, 0.1365336924791336, -0.09049676358699799, 0.12120754271745682, 0.0136170145124197, -0.10470312088727951, 0.028160136193037033, 0.016344571486115456, -0.12657558917999268, 0.13060471415519714, -0.06167493015527725, 0.16009041666984558, 0.036823879927396774, 0.039229683578014374, -0.16632845997810364, 0.03289586678147316, 0.018522586673498154, -0.0020136951934546232, 0.074550561606884, 0.03188240900635719, -0.09878088533878326, 0.12299325317144394, -0.030804073438048363, 0.07085336744785309, 0.026982847601175308, -0.09576057642698288, -0.14121906459331512, -0.03981143236160278, 0.01917075365781784, 0.09161362051963806, 0.11691052466630936, -0.013372039422392845, 0.07249557226896286, -0.049557559192180634, 0.07326251268386841, 0.2233935296535492, -0.11410494148731232, 0.0017548148753121495, 0.1122177317738533, 0.0116808470338583, -0.017384734004735947, -0.06117355823516846, -0.017219239845871925, 0.0098867891356349, 0.0464223250746727, 0.07478402554988861, -0.03568461537361145, -0.15703314542770386, 0.0982166975736618, -0.09689465165138245, -0.0017949910834431648, 0.2291911393404007, -0.0724848210811615, -0.03622926026582718, -0.04113791510462761, -0.09550070017576218, -0.07640102505683899, -0.04039677232503891, 0.021560102701187134, -0.028298527002334595, 0.017115555703639984, -0.00507830735296011, 0.054369986057281494, -0.04884633049368858, -0.07144162803888321, -0.11628127843141556, 0.09758540987968445, 0.005969997029751539, 0.004496817942708731, -0.10336770117282867, 0.12003348022699356, 0.07448195666074753, -0.1592993140220642, 0.010160556994378567, -0.05493120849132538, -0.027169670909643173, -0.031159529462456703, 0.02449658513069153, 0.00442109489813447, 0.08340755850076675, 0.08673034608364105, -0.02027439884841442, 0.005313970614224672, 0.009743006899952888, -0.013241911306977272, -0.0006650349823758006, 0.13207373023033142, -0.020694447681307793, -0.057041313499212265, 0.02546292543411255, -0.017006145790219307, 0.03608011081814766, -0.035903092473745346, -0.12297196686267853, -0.010057128965854645, 0.07394244521856308, 0.06706161797046661, -0.034691136330366135, 0.05621521174907684, -0.08085454255342484, -0.05801916867494583, 0.09159279614686966, -0.093962661921978, -0.01616363599896431, 0.009359115734696388, 0.036184195429086685, 0.055783823132514954, -0.022074446082115173, 0.025174502283334732, -0.024258553981781006, 0.08602629601955414, -0.08698521554470062, -0.04623093828558922, -0.09418749809265137, -0.08716673403978348, -0.0027459722477942705, 0.060272861272096634, 0.06618855893611908, -0.12137343734502792, -0.27929481863975525, 0.070398710668087, -0.0031888550147414207, -0.004731074906885624, -0.06513594835996628, -0.06229998543858528, -0.07290175557136536, 0.01882343925535679, -0.0013377126306295395, 0.05620633438229561, -0.0730438232421875, 0.04167962074279785, 0.09810938686132431, 0.03385693207383156, -0.0577680729329586, 0.035484205931425095, -0.1247413381934166, -0.005083072930574417, -0.20082394778728485, 0.02227390743792057, -0.09022682905197144, 0.08497850596904755, -0.08005837351083755, -0.0571819469332695, 0.12086547911167145, -0.023106612265110016, 0.01391913928091526, 0.155000701546669, -0.0824909582734108, -0.15173493325710297, 0.05425579473376274, -0.13017551600933075, -0.09914419054985046, 0.050210967659950256, -0.041104961186647415, 0.012794584967195988, 0.0724141076207161, 0.04529716819524765, 0.06468714028596878, -0.12582285702228546, -0.10789607465267181, 0.005102233495563269, 0.01758757419884205, -0.008351767435669899, 0.1009722352027893, 0.046236101537942886, -0.045066189020872116, 0.05726886913180351, -0.028610756620764732, 0.05509303882718086, -0.04212341085076332, -0.08720792829990387, -0.024683039635419846, -0.022389855235815048, 0.118604876101017, 0.02586563676595688, 0.006347884889692068, -0.007340674288570881, -0.08553443849086761, 0.1426326483488083, 0.09255379438400269, -0.03929474949836731, 0.007815780118107796, -0.1358477771282196, 0.08480779081583023, 0.03943285346031189, 0.010430914349853992, -0.10666550695896149, -0.09843919426202774, 0.04029381275177002, 0.05934690311551094, -0.03984984755516052, -0.07846033573150635, 0.004994196817278862, 0.03822321072220802, -0.10758102685213089, -0.010638806037604809, -0.07450192421674728, -0.0030961206648498774, -0.0027052012737840414, -0.22998428344726562, 0.07029972970485687, -0.03023253194987774, 0.1514805555343628, -0.11554566770792007, 0.055588096380233765, 0.1481340527534485, 0.07790035009384155, -0.01882418990135193, 0.0003838555421680212, -0.009082437492907047, 0.03804789483547211, -0.07385119795799255, -0.04310503974556923, 0.00040135273593477905, -0.006228477694094181, -0.05501480773091316, 0.014905838295817375, -0.1409856677055359, 0.10385745763778687, 0.17176033556461334, -0.020153800025582314, -0.033900827169418335, -0.049435634166002274, 0.0017325192457064986, -0.01704198494553566, -0.10653931647539139, -0.004161580465734005, 0.22184249758720398, -0.021059714257717133, 0.07317566126585007, -0.041779790073633194, -0.016728967428207397, 0.0004861399647779763, 0.03533899784088135, 0.04771801084280014, 0.12557972967624664, 0.11136432737112045, -0.04525011405348778, 0.10271064192056656, 0.04581567645072937, -0.0610584020614624, 0.14382490515708923, -0.06411578506231308, -0.02077593468129635, 0.0033593426924198866, -0.10360828042030334, -0.00012837040412705392, 0.20602886378765106, -0.19280388951301575, 0.014834407716989517, 0.0570586621761322, -0.03366382420063019, 0.0061224703676998615, -0.08074496686458588, -0.02010968141257763, -0.00833974126726389, -0.020361537113785744, -0.1164812445640564, 0.010541609488427639, 0.04348275437951088, 0.06405669450759888, -0.013434837572276592, -0.08890311419963837, 0.06011183559894562, 0.013375537469983101, -0.11197894811630249, 0.23087823390960693, -0.11256899684667587, -0.23288831114768982, -0.12642496824264526, -0.09125322848558426, -0.011648864485323429, -0.008547701872885227, 0.024908987805247307, -0.11611195653676987, -0.10033982247114182, -0.002634481294080615, 0.1200370043516159, -0.018198465928435326, 0.08031809329986572, 0.07677499949932098, 0.02462157793343067, -0.03190728649497032, -0.10337651520967484, -0.00028517426108010113, 0.039366476237773895, -0.04961986467242241, 0.038509026169776917, 0.034430790692567825, 0.0457627959549427, 0.18214358389377594, 0.011299504898488522, 0.01297291461378336, -0.051890742033720016, 0.145633265376091, -0.11326411366462708, -0.03274973854422569, 0.21298009157180786, 0.04337966442108154, -0.0037754117511212826, 0.166614830493927, 0.0022303613368421793, -0.021418163552880287, 0.07499632239341736, -0.016575481742620468, -0.06089472398161888, -0.27132144570350647, -0.13153740763664246, -0.06844216585159302, 0.0884777382016182, -0.03830616921186447, 0.03620748594403267, 0.035011447966098785, 0.05822054669260979, -0.008084306493401527, 0.06620316207408905, -0.027309583500027657, 0.12759719789028168, 0.21875469386577606, 0.06798093765974045, 0.08814381808042526, -0.028623756021261215, -0.04031915217638016, 0.06545018404722214, -0.047964151948690414, 0.1350986510515213, 0.00210153148509562, 0.07907691597938538, 0.012442464008927345, 0.013061718083918095, 0.06430848687887192, 0.050245996564626694, 0.005625930614769459, -0.014654489234089851, 0.021389948204159737, -0.044932521879673004, 0.05490623787045479, -0.01361791044473648, -0.10128103941679001, -0.0072847153060138226, -0.036688413470983505, 0.08798585832118988, 0.08131901174783707, -0.028138626366853714, 0.030286148190498352, -0.24235017597675323, -0.05318327620625496, 0.06473622471094131, -0.05438307672739029, -0.09103740006685257, 0.022513369098305702, -0.016631782054901123, -0.1311972439289093, 0.0868544727563858, -0.011741002090275288, 0.08284129202365875, -0.11328296363353729, 0.054015208035707474, 0.03115612454712391, 0.06378009915351868, -0.010778636671602726, 0.11605659872293472, -0.3574569821357727, 0.14302398264408112, 0.04525010287761688, 0.0585520900785923, -0.032656941562891006, 0.035427436232566833, -0.019736861810088158, 0.06306111067533493, 0.03613286837935448, 0.0015588070964440703, 0.03857932612299919, 0.023532144725322723, -0.12380576133728027, 0.05607876926660538, 0.01972884126007557, -0.04864325374364853, 0.03151437267661095, -0.015309573151171207, 0.00980503112077713, 0.03193068876862526, -0.07489696145057678, -0.11954237520694733, -0.12249703705310822, 0.0724840834736824, -0.02833181992173195, 0.10553411394357681, -0.0949404239654541, -0.1017899289727211, -0.01154080405831337, 0.11701636016368866, -0.13018371164798737, -0.10446424037218094, -0.14543458819389343, -0.0024879719130694866, 0.0468338243663311, -0.07617487013339996, 0.07248333096504211, 0.0007165645365603268, 0.08985502272844315, -0.03851480409502983, -0.105103500187397, 0.10596023499965668, -0.0735815241932869, -0.1845039278268814, -0.05159608647227287, 0.11019415408372879, 0.10597829520702362, 0.061892181634902954, -0.01762099750339985, 0.002627615351229906, -0.08710857480764389, -0.13354884088039398, 0.0019490086706355214, -0.003619739320129156, 0.06997157633304596, 0.030686886981129646, 0.07163023203611374, -0.08973808586597443, -0.029617641121149063, -0.07092350721359253, 0.07143158465623856, 0.17868725955486298, -0.02529238350689411, 0.10557912290096283, 0.06712916493415833, -0.05346202477812767, -0.22905419766902924, -0.04060044884681702, 0.006546988617628813, 0.03434533253312111, -0.0417700856924057, -0.2603490650653839, 0.14992015063762665, -0.03336353227496147, -0.06238055229187012, -0.021979760378599167, -0.14911429584026337, -0.13862578570842743, 0.19529153406620026, 0.023637399077415466, 0.09843695163726807, -0.09016675502061844, -0.030206389725208282, -0.09687232971191406, -0.20985184609889984, 0.21391688287258148, -0.16312511265277863, 0.043762870132923126, -0.01814400777220726, 0.1775541752576828, -0.018950358033180237, -0.033741384744644165, 0.13254137337207794, 0.04565657302737236, 0.003381960326805711, -0.015064959414303303, -0.0869690403342247, 0.21574655175209045, 0.0404205322265625, 0.00023648139904253185, -0.06185615435242653, 0.04675917699933052, -0.10915745049715042, -0.04626650735735893, -0.023354902863502502, 0.11609463393688202, -0.035210464149713516, -0.14422093331813812, -0.10573183000087738, 0.018097788095474243, -0.023338012397289276, -0.04875597730278969, 0.20852459967136383, 0.06660659611225128, 0.030271848663687706, 0.023279430344700813, 0.21846453845500946, -0.008807821199297905, 0.05117590352892876, 0.0016757172998040915, -0.01847502961754799, 0.05109686031937599, -0.17459626495838165, 0.0324547216296196, 0.09452251344919205, 0.041141875088214874, 0.04110446572303772, 0.09757103025913239, -0.053605642169713974, -0.02091403305530548, 0.09805586189031601, -0.22390656173229218, -0.0490570105612278, -0.09020380675792694, 0.1416398286819458, -0.12644392251968384, 0.08921132236719131, 0.10943692922592163, -0.03419070690870285, -0.06390289217233658, 0.04680965840816498, 0.04068799316883087, -0.039715852588415146, 0.12414779514074326, 0.04664304852485657, 0.11153461039066315, -0.13969479501247406, 0.058586012572050095, 0.06577230989933014, 0.0437915064394474, -0.03380024433135986, 0.08656750619411469, -0.13884258270263672, -0.009366301819682121, 0.02863599918782711, 0.1870429515838623, -0.11871105432510376, -0.049012888222932816, -0.18135632574558258, -0.15868915617465973, 0.030636364594101906, 0.14459189772605896, 0.10902988910675049, 0.09003607928752899, -0.05458352342247963, 0.013843245804309845, -0.09713202714920044, 0.08459781110286713, 0.005261360667645931, 0.030839229002594948, -0.14776232838630676, 0.016619887202978134, -0.010012367740273476, 0.04193929582834244, -0.06118357181549072, 0.005693513434380293, -0.1459742784500122, 0.021773429587483406, -0.069001205265522, -0.055055633187294006, -0.04352845251560211, -0.023780034855008125, 0.00399506650865078, -0.04729226604104042, -0.024519000202417374, 0.03512803092598915, -0.04045788198709488, 0.058159273117780685, 0.023794177919626236, 0.06390014290809631, -0.11291755735874176, 0.06541306525468826, 0.07318757474422455, -0.06398507952690125, 0.12958228588104248, 0.08792155236005783, 0.07137124985456467, 0.09874490648508072, -0.18329915404319763, 0.0257770586758852, 0.06084328144788742, 0.04790924862027168, 0.04445677250623703, -0.12767741084098816, -0.0058440822176635265, -0.00030383499688468874, 0.030794939026236534, 0.0520068034529686, 0.04481857642531395, -0.09428161382675171, 0.06428897380828857, -0.05449584871530533, -0.021168315783143044, -0.048650842159986496, 0.034690313041210175, 0.13566121459007263, 0.02815745770931244, 0.11796989291906357, -0.05890361964702606, 0.002965993247926235, -0.08490131795406342, 0.029970156028866768, -0.02534250169992447, -0.07294206321239471, -0.10030902177095413, -0.04231467843055725, 0.05400000885128975, 0.006539626512676477, 0.20790547132492065, 0.0109012546017766, -0.022319508716464043, 0.055985063314437866, 0.17854204773902893, 0.06770138442516327, 0.03574318066239357, 0.11652711778879166, 0.10376311838626862, 0.04210745915770531, 0.08410879224538803, -0.010335885919630527, 0.1258620321750641, 0.09653474390506744, 0.1648865044116974, 0.04192133620381355, -0.01482220459729433, 0.03847740590572357, 0.016313884407281876, -0.09836836904287338, -0.15942569077014923, 0.012309764511883259, -0.07097160071134567, 0.0305766761302948, -0.041823145002126694, 0.19656778872013092, 0.11698845028877258, -0.1017056331038475, 0.04963140934705734, -0.02057306468486786, -0.10170990228652954, -0.10739225149154663, -0.004458441864699125, -0.039132315665483475, -0.17803752422332764, 0.027460431680083275, -0.13438551127910614, -0.07522346079349518, 0.14086861908435822, -0.0012483370956033468, -0.03662097826600075, 0.09066411107778549, 0.18468502163887024, -0.01655934751033783, 0.03177834302186966, -0.007191323209553957, -0.016091225668787956, -0.018781213089823723, 0.003994643222540617, -0.08034814149141312, 0.057942166924476624, 0.009549890644848347, -0.01219344511628151, -0.03275343030691147, 0.011104615405201912, -0.016632061451673508, -0.062074147164821625, 0.013712472282350063, -0.003520411206409335, -0.08774737268686295, 0.17236915230751038, -0.010692339390516281, 0.06331933289766312, 0.045485276728868484, 0.2808881103992462, -0.05996038392186165, -0.14459629356861115, -0.09204000979661942, 0.10702067613601685, -0.02017020806670189, 0.03868277743458748, 0.0544118694961071, -0.08945472538471222, -0.09335634857416153, 0.17251498997211456, 0.11221310496330261, 0.0034052585251629353, -0.009203623048961163, -0.010991684161126614, -0.010550525970757008, 0.006455277092754841, 0.01619860902428627, 0.07208032160997391, 0.17501237988471985, -0.07020387798547745, -0.04755570366978645, -0.049729231745004654, -0.048828043043613434, -0.054288867861032486, 0.054318949580192566, 0.038498952984809875, -0.03912721201777458, -0.017285333946347237, 0.06021960452198982, -0.18159672617912292, -0.1082448810338974, -0.09057100862264633, -0.10772879421710968, -0.09155850857496262, -0.05898476019501686, 0.14807642996311188, 0.002862411318346858, 0.01309814490377903, -0.010941008105874062, 0.02517819218337536, 0.020399173721671104, 0.05576423183083534, -0.1449955254793167, -0.012456029653549194, 0.1333567500114441, -0.1695428490638733, 0.13903392851352692, -0.0656224712729454, 0.062057752162218094, 0.11952157318592072, 0.02517482452094555, -0.07624302059412003, 0.007833842188119888, 0.036696262657642365, -0.13243094086647034, 0.009407199919223785, 0.11838074028491974, -0.004467021208256483, 0.08052518218755722, 0.005257238168269396, 0.022059818729758263, 0.0021422426216304302, -0.019423989579081535, -0.06981226056814194, -0.05133325234055519, -0.04032248258590698, -0.08187370747327805, 0.051393624395132065, 0.07512816786766052, -0.06018799543380737, 0.023166242986917496, -0.0746450126171112, 0.04584803804755211, -0.015619237907230854, 0.006321766879409552, -0.011691068299114704, -0.15208351612091064, -0.05709490552544594, 0.1940516084432602, 0.017939291894435883, -0.15055739879608154, 0.01629195176064968, -0.05504368618130684, -0.003845454193651676, -0.1486867219209671, 0.08689909428358078, 0.08795714378356934, -0.03178509697318077, -0.0246585700660944, -0.00906673539429903, -0.025741709396243095, 0.08520861715078354, -0.16734477877616882, -0.0861779972910881 ]
null
null
transformers
# BatterySciBERT-cased for QA **Language model:** batteryscibert-cased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 3 base_LM_model = "batteryscibert-cased" max_seq_len = 386 learning_rate = 2e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 79.66, "f1": 87.43, ``` Evaluated on the battery device dataset. ``` "precision": 65.09, "recall": 84.56, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batteryscibert-cased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batteryscibert-cased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatterySciBERT-cased for QA Language model: batteryscibert-cased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-cased for QA \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatterySciBERT-cased for QA \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 67, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatterySciBERT-cased for QA \r\nLanguage model: batteryscibert-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.07918901741504669, 0.030719337984919548, -0.004270212724804878, 0.058000996708869934, 0.2310180813074112, 0.03763307258486748, 0.12892302870750427, 0.0430271290242672, -0.1080101728439331, -0.027432426810264587, 0.03357856348156929, -0.02925903908908367, 0.07349278777837753, 0.07380899786949158, 0.015072333626449108, -0.11344629526138306, 0.04173228144645691, -0.03555799275636673, 0.013595445081591606, 0.14523719251155853, 0.11850740760564804, -0.08880425244569778, 0.11594665050506592, 0.014074096456170082, -0.10229967534542084, 0.06594649702310562, -0.010987377725541592, -0.07650550454854965, 0.1011410653591156, -0.08119576424360275, 0.13211417198181152, 0.049154408276081085, 0.02882625162601471, -0.1808760017156601, 0.039000339806079865, 0.031200453639030457, 0.02812241017818451, 0.027764076367020607, 0.02154645323753357, -0.04487905278801918, 0.06316760182380676, 0.01767442375421524, 0.04256672039628029, 0.022271860390901566, -0.06056593731045723, -0.22201603651046753, -0.06126191467046738, -0.05815236270427704, 0.08191192895174026, 0.14418698847293854, -0.030577832832932472, 0.09084802120923996, -0.04630926623940468, 0.07285185158252716, 0.16221944987773895, -0.10448329150676727, -0.0061498722061514854, 0.10470163822174072, 0.04017265513539314, -0.0021690665744245052, -0.013729850761592388, -0.07683597505092621, 0.00024067088088486344, 0.056412313133478165, 0.04621526598930359, -0.059590473771095276, -0.14777587354183197, 0.10534479469060898, -0.10391762107610703, -0.008091632276773453, 0.19431309401988983, -0.04765773564577103, -0.07433180510997772, 0.003731544828042388, -0.119802325963974, -0.027507685124874115, -0.02100237086415291, -0.015105262398719788, -0.0244315005838871, 0.001641205046325922, -0.07048788666725159, 0.028198543936014175, -0.022561529651284218, -0.10529490560293198, -0.12957589328289032, 0.09736089408397675, 0.05474654957652092, 0.038976818323135376, -0.12380062788724899, 0.07634852081537247, 0.0017372181173413992, -0.1554226130247116, -0.03197810798883438, -0.06249720975756645, -0.07945221662521362, 0.003912968561053276, 0.0155382314696908, 0.006018758285790682, 0.12411905080080032, 0.08261813968420029, -0.031912557780742645, 0.012751682661473751, 0.024531442672014236, -0.06710639595985413, 0.008385416120290756, 0.14545302093029022, -0.025036772713065147, -0.13121667504310608, -0.011999557726085186, 0.030322996899485588, -0.002592070261016488, -0.01463227067142725, -0.06805732101202011, -0.006300115957856178, 0.05503122881054878, 0.06311449408531189, 0.015125957317650318, 0.03286505490541458, -0.05490740016102791, -0.03602861985564232, 0.04489736258983612, -0.10669991374015808, -0.04915912449359894, -0.030335260555148125, 0.012618133798241615, 0.01609661430120468, -0.0528615340590477, 0.02391786128282547, -0.03446408733725548, -0.020361503586173058, -0.07772136479616165, -0.08368993550539017, -0.13121487200260162, -0.08398383855819702, -0.035294316709041595, 0.12357425689697266, 0.06472190469503403, -0.11982089281082153, -0.3385011553764343, 0.07985257357358932, 0.0016641005640849471, 0.007493345066905022, -0.03863171488046646, -0.03599381446838379, -0.04279733821749687, -0.03055167943239212, -0.03702493757009506, 0.10681343823671341, -0.06402125954627991, 0.06388572603464127, 0.16174238920211792, 0.026598364114761353, -0.0463380366563797, 0.04232293739914894, -0.1479349583387375, 0.005027570761740208, -0.13773421943187714, 0.0431632436811924, -0.1109132468700409, -0.001343111740425229, -0.11790289729833603, -0.07889620959758759, 0.11789880692958832, -0.010176575742661953, 0.07825139164924622, 0.1435428112745285, -0.09327269345521927, -0.12015487998723984, 0.036877427250146866, -0.14022673666477203, -0.14593620598316193, 0.0476439893245697, -0.0156186418607831, 0.022132690995931625, 0.020825199782848358, 0.047915417701005936, 0.1349753588438034, -0.16547612845897675, -0.1666756272315979, -0.02747044526040554, 0.038877688348293304, -0.013396231457591057, 0.0883881226181984, 0.040573153644800186, -0.06659140437841415, 0.07181917130947113, -0.06319978833198547, 0.05081838369369507, -0.0467061847448349, -0.1060137003660202, -0.055097099393606186, -0.03447284549474716, 0.11504257470369339, 0.009355559013783932, -0.05512308329343796, 0.018410412594676018, -0.09188275784254074, 0.1217012107372284, 0.11661465466022491, -0.04249647259712219, -0.007094309199601412, -0.1919575333595276, 0.08615388721227646, -0.034282535314559937, 0.01308585237711668, -0.11588596552610397, -0.16578833758831024, 0.08252320438623428, -0.038745611906051636, -0.06012074276804924, -0.04356734827160835, 0.02123793214559555, 0.028378969058394432, -0.10756862163543701, -0.030241841450333595, -0.14896805584430695, -0.023057712242007256, 0.011029507033526897, -0.1493798941373825, 0.04625009745359421, -0.03023126721382141, 0.09909145534038544, 0.027262737974524498, 0.02251802384853363, 0.1184900626540184, 0.03531450405716896, -0.029440950602293015, -0.01586269959807396, -0.020627684891223907, 0.06510625034570694, -0.031158791854977608, -0.038493089377880096, -0.031792622059583664, -0.018846336752176285, 0.0017922415863722563, -0.03559410199522972, -0.11793050169944763, 0.06866029649972916, 0.14525845646858215, 0.11050647497177124, 0.013604072853922844, -0.09051083028316498, -0.000047681045543868095, -0.039167921990156174, -0.10000161826610565, -0.01650088094174862, 0.22238023579120636, -0.0032922951504588127, 0.01979803666472435, -0.03394503891468048, 0.014194768853485584, -0.004644650034606457, 0.03882793337106705, 0.06348244100809097, 0.08286995440721512, 0.1119069755077362, 0.04381870850920677, 0.0663168653845787, 0.06615118682384491, -0.06872140616178513, 0.12965257465839386, -0.08604712784290314, -0.03758968412876129, -0.0371658131480217, -0.10421066731214523, -0.024836134165525436, 0.26748666167259216, -0.13711011409759521, 0.06436074525117874, 0.062965527176857, -0.007028899621218443, 0.028785455971956253, -0.03671411797404289, -0.010053899139165878, -0.01884712278842926, -0.05479903519153595, -0.15988300740718842, 0.027027027681469917, 0.05216893553733826, 0.010796409100294113, -0.0207627285271883, -0.0565776564180851, 0.02146022394299507, 0.009102504700422287, -0.07256869971752167, 0.22610045969486237, -0.11285777390003204, -0.2039809674024582, -0.11124663054943085, -0.0801113173365593, -0.030409028753638268, -0.07195577025413513, 0.017685294151306152, -0.12964974343776703, -0.09756901860237122, 0.019889196380972862, 0.18085859715938568, 0.013649547472596169, 0.07033315300941467, 0.03523793816566467, 0.007982950657606125, 0.004863590467721224, -0.11045974493026733, 0.01228160411119461, 0.05512142926454544, -0.10439882427453995, 0.03370295837521553, 0.04992910102009773, 0.0819878876209259, 0.10559211671352386, 0.010489173233509064, 0.016284003853797913, -0.052150093019008636, 0.23142880201339722, -0.10491761565208435, -0.03507585451006889, 0.29795992374420166, 0.12918326258659363, -0.0032845630776137114, 0.2010762244462967, -0.03380470350384712, -0.015370127744972706, 0.10239937901496887, -0.011498569510877132, -0.052265699952840805, -0.24743026494979858, -0.08832918107509613, -0.03334017097949982, 0.05340205505490303, -0.06673971563577652, 0.011812053620815277, -0.04243452474474907, 0.09345050156116486, 0.014613766223192215, -0.0193504486232996, -0.09622699022293091, 0.09559351950883865, 0.23718084394931793, 0.059122879058122635, 0.10042460262775421, -0.00847676768898964, 0.015988018363714218, 0.06597654521465302, 0.055468205362558365, 0.16593709588050842, -0.051996875554323196, 0.041068993508815765, 0.03544751554727554, 0.14778324961662292, 0.035143304616212845, 0.03392516076564789, -0.0057678744196891785, 0.006926997564733028, 0.016716161742806435, -0.04087870568037033, 0.008414572104811668, -0.020339470356702805, -0.0340961329638958, 0.016537336632609367, -0.007247979287058115, 0.15520364046096802, 0.049357518553733826, -0.011163187213242054, 0.019537322223186493, -0.20350691676139832, -0.0272489283233881, 0.045182351022958755, -0.041495345532894135, -0.09486037492752075, 0.03047124296426773, 0.02234969288110733, -0.12689776718616486, 0.03981203958392143, -0.03348970413208008, 0.06946159899234772, -0.06492554396390915, 0.05322427675127983, 0.06365013122558594, 0.12276631593704224, 0.016384029760956764, 0.12133124470710754, -0.40804046392440796, 0.10066993534564972, 0.04419691860675812, 0.08501166105270386, 0.03379733860492706, 0.06341466307640076, -0.0512794554233551, -0.03902854770421982, 0.05492868274450302, -0.000725527061149478, 0.06343080848455429, 0.04198349639773369, -0.1517987847328186, 0.14301127195358276, 0.012772805988788605, -0.0013169727753847837, 0.07345238327980042, -0.025866767391562462, 0.03383912146091461, -0.010731975547969341, -0.015879478305578232, -0.13899201154708862, -0.09357248246669769, 0.05296599119901657, -0.06564567238092422, 0.10753882676362991, -0.14878778159618378, -0.08261582255363464, 0.06241453066468239, 0.057141102850437164, -0.0758785754442215, -0.11911176145076752, -0.14686639606952667, 0.004343586508184671, 0.12273474037647247, -0.10950120538473129, 0.07511842250823975, 0.010605988092720509, 0.060102008283138275, -0.02114074118435383, -0.10435852408409119, 0.05204596742987633, -0.08505436778068542, -0.14324523508548737, -0.029762493446469307, 0.03141281381249428, 0.09834005683660507, 0.0563807338476181, 0.002013443037867546, -0.009639212861657143, -0.1351267397403717, -0.12266770750284195, -0.03637023642659187, -0.05042112246155739, 0.03548213094472885, 0.020464247092604637, 0.13495218753814697, -0.029042966663837433, -0.02875438891351223, -0.02972138300538063, 0.0052115716971457005, 0.14283250272274017, -0.01566583290696144, 0.05114035680890083, 0.1238023042678833, -0.020446235314011574, -0.20246487855911255, -0.06139074265956879, 0.07612859457731247, 0.0567619688808918, -0.03838324546813965, -0.26033109426498413, 0.13878342509269714, -0.0189274363219738, -0.06514456123113632, -0.11295251548290253, -0.14156940579414368, -0.11814835667610168, 0.18524877727031708, 0.07385049015283585, 0.03977899253368378, -0.07191134244203568, -0.01846291311085224, -0.05554449185729027, -0.2553277015686035, 0.12433037161827087, -0.1548236608505249, 0.022100510075688362, -0.016068460419774055, 0.23365803062915802, -0.03162093088030815, -0.02943185716867447, 0.14141999185085297, 0.0642557367682457, -0.022467192262411118, 0.030266737565398216, -0.0008906861767172813, 0.12178010493516922, 0.0341331921517849, 0.04086481034755707, 0.000961078389082104, 0.11274918913841248, -0.12003538012504578, -0.03869641199707985, -0.056547507643699646, 0.0898444727063179, -0.04252389445900917, -0.16596975922584534, -0.09311138093471527, 0.04530331864953041, -0.03970084339380264, -0.062494393438100815, 0.08118950575590134, 0.040621597319841385, -0.0045250882394611835, 0.09591560810804367, 0.16456088423728943, 0.03221552073955536, -0.07307298481464386, 0.0368352010846138, -0.010251876898109913, 0.05870014801621437, -0.14277854561805725, 0.03172040730714798, 0.10917249321937561, 0.028515763580799103, 0.02182246372103691, 0.044610798358917236, -0.0828094482421875, 0.003342048032209277, 0.07531838864088058, -0.15036383271217346, -0.1531575620174408, -0.06853388994932175, 0.17938226461410522, -0.20011983811855316, 0.058094579726457596, 0.1132497563958168, -0.012673042714595795, -0.07295197993516922, 0.02466745674610138, 0.04001593962311745, -0.013427281752228737, 0.17462071776390076, 0.09199444949626923, 0.1334276795387268, -0.130473330616951, 0.05756952986121178, 0.031957823783159256, 0.024780813604593277, -0.04418172314763069, 0.01908142864704132, -0.09899406880140305, 0.011657491326332092, 0.02783786691725254, 0.11765274405479431, -0.1324334442615509, -0.08002492785453796, -0.16080805659294128, -0.13258294761180878, -0.0010429725516587496, 0.07591775059700012, 0.09052591770887375, 0.07229504734277725, -0.031973179429769516, 0.022416817024350166, -0.08071408420801163, 0.0991964191198349, 0.0032442191150039434, 0.0261086355894804, -0.109978586435318, -0.07929422706365585, -0.028636472299695015, 0.109983429312706, -0.05644981935620308, -0.0008134038071148098, -0.09111957997083664, 0.027401141822338104, -0.1500413566827774, -0.06979810446500778, -0.03494149446487427, -0.01986525021493435, 0.02764706499874592, -0.07616341859102249, -0.03724534437060356, 0.0654372051358223, -0.044850002974271774, 0.05803632363677025, 0.042728081345558167, 0.021557964384555817, -0.12662512063980103, 0.09587404131889343, 0.07720813155174255, -0.053971096873283386, 0.11360514909029007, 0.06678375601768494, 0.07146535068750381, 0.11709997802972794, -0.07689747959375381, 0.021634023636579514, 0.002353402553126216, 0.08216524869203568, 0.07692795991897583, -0.13850921392440796, 0.0030698133632540703, -0.006091054063290358, 0.015383006073534489, 0.02446296252310276, 0.05288040265440941, -0.09968336671590805, 0.07199763506650925, -0.05681579187512398, 0.026955023407936096, -0.052709925919771194, 0.0433318167924881, 0.10201916843652725, 0.08407741785049438, 0.10926682502031326, -0.04121943935751915, -0.005997381638735533, -0.11917618662118912, 0.014185220934450626, -0.030645519495010376, -0.01378981489688158, -0.04743511229753494, -0.04140540584921837, 0.0688539519906044, 0.0345039963722229, 0.1423952430486679, -0.008988498710095882, 0.06357467919588089, 0.0165481548756361, 0.11964421719312668, 0.09898975491523743, -0.004640680737793446, 0.07392986118793488, 0.0971410721540451, 0.09207750856876373, 0.12012507021427155, -0.024865062907338142, 0.05965103209018707, 0.11402386426925659, 0.11363711208105087, 0.06800349801778793, 0.0518229715526104, 0.04684216529130936, 0.04956836998462677, -0.102318175137043, -0.08604764938354492, -0.006731743458658457, -0.0687078908085823, 0.012870224192738533, -0.00702187092974782, 0.18641021847724915, 0.0819012001156807, -0.11961335688829422, 0.0827370211482048, -0.003611244959756732, -0.10833767801523209, -0.10021883249282837, 0.10006482899188995, -0.0447140708565712, -0.21344943344593048, 0.031148763373494148, -0.15478435158729553, -0.11510111391544342, 0.1958940625190735, 0.0034625090193003416, -0.008325877599418163, 0.13278263807296753, 0.2058236449956894, -0.035234175622463226, -0.04039959982037544, -0.011898603290319443, 0.017868150025606155, 0.05609087273478508, 0.03146227449178696, -0.03964419290423393, 0.10339998453855515, 0.04758034273982048, 0.007737176027148962, -0.07605388015508652, 0.010818975046277046, -0.00766877131536603, -0.030013522133231163, -0.01653178595006466, 0.0036795474588871002, -0.07893127202987671, 0.18830378353595734, -0.0019965420942753553, 0.08156147599220276, 0.01637903042137623, 0.31058961153030396, -0.052575573325157166, -0.16654983162879944, -0.13519981503486633, 0.07569373399019241, -0.05836506560444832, 0.05832753702998161, 0.05608406290411949, -0.07945869863033295, -0.12553805112838745, 0.1470264196395874, 0.024024207144975662, -0.024051301181316376, -0.0479910783469677, 0.03200417011976242, -0.018586840480566025, -0.05706864967942238, 0.03650769963860512, 0.08166589587926865, 0.20571540296077728, -0.07196301966905594, -0.013625569641590118, -0.05495684966444969, -0.07900816947221756, -0.00509007228538394, 0.01731993444263935, 0.010365349240601063, -0.02373749390244484, -0.01987512782216072, 0.050113990902900696, -0.11382365971803665, -0.09788757562637329, -0.10333968698978424, -0.029878493398427963, -0.11773249506950378, -0.06429926306009293, 0.2118491232395172, 0.014240401796996593, 0.007947705686092377, -0.04236886650323868, 0.07525146752595901, 0.045922935009002686, 0.04053337499499321, -0.09740439802408218, -0.0672442764043808, 0.1986541748046875, -0.14725586771965027, 0.141484797000885, -0.04999391734600067, 0.0849037915468216, 0.11235231906175613, -0.025812828913331032, -0.11689793318510056, 0.05677786469459534, 0.05699389800429344, -0.1757124811410904, -0.0011778209591284394, 0.13426683843135834, 0.02083873562514782, 0.11125072091817856, -0.005066053941845894, 0.0808832123875618, -0.015209767036139965, -0.007982412353157997, -0.03142517805099487, -0.052284128963947296, -0.022490890696644783, -0.07922790199518204, 0.04573345184326172, 0.07667941600084305, -0.057868003845214844, 0.07156592607498169, -0.07769055664539337, 0.09843406081199646, -0.03591735661029816, 0.08921390771865845, 0.018600091338157654, -0.11911483108997345, -0.06917061656713486, 0.14050938189029694, -0.06151963025331497, -0.10913971811532974, 0.013238618150353432, -0.007812007796019316, 0.0019001459004357457, -0.0906921997666359, 0.14146463572978973, 0.0691775530576706, -0.004087354987859726, -0.02209480106830597, 0.032936081290245056, -0.05547196418046951, 0.06509033590555191, -0.17109094560146332, -0.094106525182724 ]
null
null
transformers
# BatterySciBERT-cased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the [SciBERT-cased](https://huggingface.co/allenai/scibert_scivocab_cased) weights. It was introduced in [this paper](paper_link) and first released in [this repository](https://github.com/ShuHuang/batterybert). This model is case-sensitive: it makes a difference between english and English. ## Model description BatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the [SciBERT-cased](https://huggingface.co/allenai/scibert_scivocab_cased) weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the [SciBERT-cased](https://huggingface.co/allenai/scibert_scivocab_cased) weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at [Github](https://github.com/ShuHuang/batterybert/blob/main/corpus.txt). ## Training procedure ### Preprocessing The texts are tokenized using WordPiece and a vocabulary size of 31,116. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=batterybert) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='batterydata/batteryscibert-cased') >>> unmasker("Hello I'm a <mask> model.") ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batteryscibert-cased') model = BertModel.from_pretrained('batterydata/batteryscibert-cased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batteryscibert-cased') model = TFBertModel.from_pretrained('batterydata/batteryscibert-cased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Evaluation results Final loss: 1.0505. ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": ["exbert"], "datasets": ["batterypapers"]}
fill-mask
batterydata/batteryscibert-cased
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "exbert", "en", "dataset:batterypapers", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatterySciBERT-cased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-cased weights. It was introduced in this paper and first released in this repository. This model is case-sensitive: it makes a difference between english and English. ## Model description BatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-cased weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github. ## Training procedure ### Preprocessing The texts are tokenized using WordPiece and a vocabulary size of 31,116. The inputs of the model are then of the form: The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by '[MASK]'. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Evaluation results Final loss: 1.0505. ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-cased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-cased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is case-sensitive: it makes a difference between english and English.", "## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-cased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are tokenized using WordPiece and a vocabulary size of 31,116. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.0505.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatterySciBERT-cased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-cased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is case-sensitive: it makes a difference between english and English.", "## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-cased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-cased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are tokenized using WordPiece and a vocabulary size of 31,116. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.0505.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 61, 82, 283, 97, 3, 131, 113, 128, 49, 11, 50 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatterySciBERT-cased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-cased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is case-sensitive: it makes a difference between english and English.## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-cased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.045523080974817276, 0.06799842417240143, -0.0057398974895477295, 0.047050315886735916, 0.07623998820781708, 0.0152371134608984, 0.1584051549434662, -0.005189898423850536, -0.008393499068915844, 0.05165661498904228, 0.03522811457514763, 0.0027335737831890583, 0.0664839968085289, 0.07607081532478333, 0.0996999740600586, -0.2664535641670227, 0.06323351711034775, -0.10992493480443954, 0.08061287552118301, 0.0682177022099495, 0.07505840808153152, -0.08828286081552505, 0.023560959845781326, 0.04012952744960785, 0.051219720393419266, -0.013542224653065205, -0.011629599146544933, -0.08223625272512436, 0.09160348773002625, 0.03432594612240791, 0.08690354973077774, 0.00948361773043871, 0.03653550148010254, -0.04950942471623421, 0.02709926851093769, 0.04495616257190704, 0.008302506990730762, 0.0731615200638771, 0.046246759593486786, -0.009816569276154041, 0.10048381984233856, 0.03744646534323692, 0.0858331248164177, 0.037716858088970184, -0.0900932252407074, -0.12172872573137283, -0.026169857010245323, 0.008217976428568363, -0.01188684906810522, 0.09598683565855026, -0.02535688504576683, 0.09253409504890442, -0.02935188263654709, 0.06952493637800217, 0.18408504128456116, -0.09430693089962006, -0.016865810379385948, 0.004665395710617304, 0.017438868060708046, 0.00861096940934658, -0.002647372893989086, -0.037777017802000046, -0.022025106474757195, 0.0358707457780838, 0.07710260897874832, 0.004946846980601549, 0.04663030803203583, -0.04223234951496124, -0.1552293598651886, -0.032117825001478195, 0.10388831794261932, -0.04876099154353142, -0.09687258303165436, -0.13536371290683746, -0.09326780587434769, 0.10518715530633926, 0.003317426424473524, 0.04094995930790901, -0.024558430537581444, 0.01905740052461624, 0.04166954755783081, -0.050706423819065094, -0.06352962553501129, -0.08004016429185867, -0.1020098552107811, 0.19095008075237274, 0.03197925537824631, 0.048568930476903915, -0.053629979491233826, 0.06834081560373306, -0.0835394486784935, -0.07333158701658249, -0.06359337270259857, -0.04619671776890755, -0.0770111083984375, 0.006533863488584757, -0.021155284717679024, -0.2119177132844925, -0.031215740367770195, 0.1260863095521927, 0.002899925922974944, -0.004332526586949825, -0.011331763118505478, 0.02259732410311699, 0.11606886237859726, 0.06867458671331406, -0.10465274006128311, 0.11239029467105865, 0.0073363035917282104, -0.039693597704172134, 0.04280206188559532, -0.06397907435894012, -0.01275235041975975, 0.013579950667917728, -0.040737830102443695, 0.019999707117676735, 0.04464736580848694, 0.05847024545073509, -0.02750730887055397, -0.0511857308447361, 0.07139565795660019, -0.11716804653406143, -0.04973818361759186, -0.0036051361821591854, -0.029385121539235115, 0.011340150609612465, 0.0527222715318203, -0.08098439127206802, -0.057773809880018234, 0.04997431859374046, -0.07285754382610321, -0.07620339095592499, -0.11657217144966125, -0.1444210559129715, -0.03602484241127968, 0.01409967616200447, -0.06674645096063614, -0.07866360992193222, -0.2604616582393646, -0.02611405961215496, 0.03641916811466217, 0.0050374469719827175, 0.012244864366948605, -0.02076539769768715, 0.008728470653295517, -0.0003442996530793607, 0.007930557243525982, -0.02388073317706585, -0.011302035301923752, 0.0026893315371125937, -0.033169932663440704, 0.08309219032526016, 0.020761564373970032, 0.004064319655299187, -0.07772566378116608, 0.0011593465460464358, -0.21441452205181122, 0.10001876205205917, -0.08227947354316711, -0.038465894758701324, -0.029398007318377495, -0.007749256677925587, 0.014268306083977222, 0.012000973336398602, 0.027870705351233482, 0.12693744897842407, -0.12737075984477997, -0.08892319351434708, 0.2068580538034439, -0.18093335628509521, 0.05623714253306389, 0.0969911441206932, -0.06419555097818375, 0.07229813188314438, 0.10723094642162323, -0.02409418299794197, 0.12630067765712738, -0.12418445199728012, -0.06297724694013596, 0.061569176614284515, -0.061305779963731766, 0.102823905646801, 0.10297238081693649, -0.0036659373436123133, -0.12330576777458191, 0.03713050112128258, 0.016763800755143166, 0.030628899112343788, -0.01940503716468811, -0.024462321773171425, 0.03226844593882561, -0.017291588708758354, 0.07199262082576752, -0.007963070645928383, -0.003500071819871664, 0.02209417335689068, -0.08923845738172531, 0.07223247736692429, 0.12895013391971588, -0.05799184367060661, -0.015460503287613392, -0.12232011556625366, 0.06047819182276726, -0.03606737405061722, -0.033528830856084824, -0.16788305342197418, -0.011910139583051205, 0.043855298310518265, -0.12179955095052719, 0.09406834840774536, -0.016821319237351418, 0.03678421676158905, 0.1251002997159958, -0.04028293490409851, 0.020751656964421272, -0.03562271222472191, -0.03305777534842491, -0.01867912895977497, -0.13972975313663483, -0.004492520820349455, -0.05006936192512512, 0.03856737166643143, 0.053388286381959915, 0.0396893173456192, -0.04742942377924919, -0.04094203934073448, 0.01027734111994505, -0.059990137815475464, 0.03996644541621208, -0.026700256392359734, -0.01894070953130722, -0.013254260644316673, -0.03762875869870186, 0.03737422078847885, -0.0615866556763649, 0.02749784290790558, -0.16469097137451172, -0.1282626837491989, 0.05279207229614258, 0.0741744339466095, -0.10128732770681381, 0.060462288558483124, -0.017903074622154236, 0.00603283429518342, -0.12068518996238708, -0.038280095905065536, 0.21245385706424713, 0.0073975990526378155, 0.04457175359129906, -0.08399524539709091, 0.035606708377599716, 0.028588637709617615, -0.020027030259370804, -0.004944269545376301, 0.07377072423696518, 0.06213347613811493, -0.0866389349102974, 0.01738726533949375, -0.11137422919273376, -0.018446320667862892, 0.16221438348293304, 0.03582486882805824, -0.058974020183086395, -0.07293404638767242, -0.033308107405900955, 0.03245671093463898, 0.05459629371762276, -0.04994507506489754, 0.04656657949090004, 0.04933773726224899, -0.012841652147471905, 0.024108266457915306, -0.025853794068098068, 0.04007098823785782, 0.06278814375400543, 0.008063167333602905, -0.0919264480471611, -0.07663486897945404, -0.01906603015959263, 0.07221417129039764, 0.05318692699074745, -0.011969432234764099, -0.00495134387165308, -0.03251517936587334, -0.0977766215801239, 0.16852925717830658, -0.09499187767505646, -0.239618182182312, -0.1724669486284256, -0.019162971526384354, 0.008237122558057308, 0.03435929864645004, -0.016678903251886368, -0.004098503850400448, -0.08002661168575287, -0.08322326838970184, 0.13405954837799072, -0.025813028216362, 0.013925650157034397, -0.01552486326545477, -0.02656841278076172, -0.05390847474336624, -0.11048480123281479, 0.02749086730182171, -0.024326328188180923, -0.07941801100969315, 0.010385453701019287, -0.04074057564139366, 0.051211223006248474, 0.10100983828306198, 0.019201988354325294, -0.04306791350245476, -0.026631822809576988, 0.18738384544849396, -0.03432077541947365, 0.0711057186126709, 0.1582627296447754, -0.0016238230746239424, 0.04610577970743179, 0.05487748235464096, 0.014028224162757397, -0.011925681494176388, 0.05695027858018875, -0.013075468130409718, -0.09492981433868408, -0.1902649700641632, -0.08450313657522202, -0.04574337229132652, 0.03415296599268913, 0.06351315975189209, 0.006933222059160471, 0.0335301011800766, 0.07236485928297043, -0.05257144570350647, -0.023518342524766922, 0.06590679287910461, 0.06650164723396301, -0.03540125489234924, -0.030375320464372635, 0.08682181686162949, -0.049072008579969406, -0.029148999601602554, 0.11759399622678757, -0.13609857857227325, 0.15909171104431152, -0.07509707659482956, 0.14469502866268158, 0.04950137063860893, 0.015797829255461693, 0.07580632716417313, 0.15817013382911682, -0.06252477318048477, 0.048889823257923126, -0.042665671557188034, -0.07788078486919403, -0.013040809892117977, 0.04389047622680664, 0.015531227923929691, 0.02279498428106308, -0.03567495942115784, 0.012900205329060555, 0.0012040893780067563, 0.1516185849905014, 0.01189420185983181, -0.13455384969711304, -0.12343553453683853, 0.008402688428759575, -0.006825379561632872, -0.08446846157312393, 0.02361210808157921, 0.10369237512350082, -0.08743897825479507, 0.07180729508399963, 0.002012314274907112, 0.053767211735248566, -0.08281264454126358, 0.013122469186782837, -0.09062830358743668, 0.10162898898124695, -0.05248408019542694, 0.11344657093286514, -0.18048720061779022, 0.041983213275671005, 0.029628530144691467, 0.09524056315422058, -0.12001506984233856, 0.038285352289676666, 0.01039082370698452, -0.06261110305786133, 0.14630256593227386, 0.008543204516172409, -0.06222188472747803, 0.007231743540614843, -0.11768703907728195, 0.00041389494435861707, 0.043956588953733444, -0.06582190096378326, 0.05353255942463875, 0.020345252007246017, 0.010790280997753143, 0.011756135150790215, 0.012001407332718372, -0.047512225806713104, -0.14054176211357117, 0.06912567466497421, -0.04218257591128349, -0.0774608850479126, -0.07892755419015884, -0.06603042781352997, -0.02419308014214039, 0.1271224021911621, -0.1513572782278061, -0.09529322385787964, -0.12470173090696335, -0.025440648198127747, 0.12135879695415497, -0.08678486198186874, 0.03816897049546242, -0.023469045758247375, 0.14687930047512054, -0.08697386085987091, -0.11812625080347061, 0.05776461213827133, -0.062356431037187576, -0.14228415489196777, -0.04533962160348892, 0.0326305627822876, 0.2017105221748352, 0.0644175261259079, -0.03831973299384117, 0.060946155339479446, -0.04312329366803169, -0.1163264736533165, -0.020280594006180763, 0.1384252905845642, -0.016162272542715073, 0.1180337443947792, -0.008099996484816074, -0.09358549118041992, -0.017428118735551834, 0.006779191084206104, 0.08057837933301926, 0.15733085572719574, -0.04787033051252365, 0.15156593918800354, 0.2690946161746979, -0.0774693489074707, -0.24192403256893158, -0.056284017860889435, 0.048137079924345016, 0.06833607703447342, 0.002636537654325366, -0.29201772809028625, 0.061733707785606384, 0.02924373187124729, -0.04022175073623657, -0.07973584532737732, -0.19113297760486603, -0.12989109754562378, 0.15178896486759186, 0.02970125339925289, 0.1119178906083107, 0.004704439081251621, -0.016184985637664795, -0.04973173886537552, 0.07940074801445007, 0.16553561389446259, -0.06232856959104538, 0.0920926183462143, 0.025014350190758705, -0.0026180422864854336, 0.037204816937446594, -0.06891396641731262, 0.08454913645982742, -0.03565886244177818, 0.04469888657331467, -0.0528486929833889, -0.017842935398221016, 0.126461923122406, -0.01221452560275793, 0.1072152778506279, 0.09131132811307907, 0.084876649081707, 0.00016254671209026128, -0.11461954563856125, -0.06455574184656143, 0.0784192606806755, -0.01705116219818592, -0.09523844718933105, -0.10749618709087372, 0.08408604562282562, 0.049089331179857254, -0.021358158439397812, 0.0182886254042387, -0.028537170961499214, 0.0692397803068161, 0.1212046667933464, 0.1289779096841812, 0.05873013287782669, -0.049336861819028854, 0.009158930741250515, -0.01593247428536415, 0.08114161342382431, -0.008500630035996437, 0.03191981837153435, 0.056020352989435196, 0.06067582219839096, 0.08148263394832611, 0.05821891501545906, -0.18455465137958527, -0.0012616580352187157, 0.020938606932759285, -0.15501420199871063, -0.16972646117210388, -0.028003161773085594, 0.09900148957967758, -0.12909680604934692, -0.02825448475778103, 0.11480867862701416, -0.0249765757471323, -0.025148961693048477, -0.010428992100059986, 0.04407903924584389, 0.025937464088201523, 0.13055230677127838, 0.024659382179379463, 0.0469961017370224, -0.07624495029449463, 0.06275084614753723, 0.05816393718123436, 0.017624299973249435, 0.027245206758379936, 0.07658297568559647, -0.07712332904338837, -0.02878776378929615, -0.005501303356140852, 0.1132337749004364, 0.004826247692108154, -0.031061269342899323, -0.04755234345793724, -0.09503591060638428, 0.06644299626350403, 0.15960480272769928, 0.019621573388576508, 0.0510326586663723, -0.08520673960447311, 0.04173446074128151, -0.05048951506614685, 0.037206582725048065, -0.055377859622240067, -0.02721599116921425, 0.0320008285343647, 0.11426930129528046, 0.015250833705067635, 0.021342020481824875, -0.06146807596087456, -0.08972954005002975, -0.09442324191331863, 0.006846224423497915, 0.012514257803559303, 0.034048277884721756, -0.07435774058103561, -0.015674524009227753, -0.005331138614565134, 0.0389830656349659, 0.04230698570609093, 0.006765960715711117, -0.03018982894718647, 0.023410888388752937, -0.049417782574892044, -0.008648299612104893, -0.1115691065788269, 0.039261236786842346, 0.05066400766372681, -0.04572688415646553, 0.10157015919685364, 0.043592799454927444, -0.017750011757016182, 0.009154455736279488, -0.10712669789791107, 0.020696932449936867, -0.03816569223999977, 0.049488428980112076, -0.0361022874712944, -0.08880206942558289, 0.017417695373296738, -0.025866104289889336, -0.032562367618083954, -0.008480300195515156, -0.0076094139367341995, -0.06671842187643051, 0.1527014970779419, 0.11011344194412231, -0.005823523737490177, -0.09872075170278549, 0.046770207583904266, 0.008798923343420029, 0.014948609285056591, 0.05461882799863815, -0.030247226357460022, 0.05615796521306038, -0.10080419480800629, 0.0023406611289829016, 0.013676485046744347, 0.006401600316166878, 0.023335186764597893, -0.03988000378012657, 0.0644647628068924, 0.005479688290506601, 0.1498723030090332, 0.02021016925573349, -0.05903248488903046, -0.0064532021060585976, 0.07443210482597351, 0.046186663210392, -0.0070735616609454155, 0.009463849477469921, 0.03252175450325012, -0.0283617302775383, 0.056777819991111755, 0.003858113894239068, -0.016156906262040138, 0.08746511489152908, 0.13202619552612305, -0.014670930802822113, 0.1879710704088211, -0.022695034742355347, 0.042334068566560745, 0.010173927992582321, -0.058166686445474625, -0.05665415897965431, -0.00275045377202332, 0.023545272648334503, 0.0056818644516170025, 0.013591325841844082, 0.0546804741024971, -0.056127551943063736, 0.1630162000656128, 0.06663282215595245, -0.052224788814783096, -0.12875932455062866, -0.22342318296432495, -0.007860004901885986, 0.06475773453712463, 0.0006694082985632122, -0.12081241607666016, -0.035648930817842484, 0.07381407171487808, 0.024681231006979942, 0.004932595882564783, 0.07047445327043533, -0.08009263873100281, -0.04881390929222107, 0.0363464429974556, -0.025438986718654633, -0.005941110197454691, 0.005292884074151516, -0.017153894528746605, 0.046620823442935944, 0.11728695780038834, 0.041701558977365494, 0.05529163405299187, 0.04438740015029907, 0.029968978837132454, -0.009067105129361153, -0.07587838172912598, -0.00786441471427679, -0.04026798903942108, -0.027805188670754433, 0.14792589843273163, -0.0026388666592538357, -0.032956551760435104, -0.000808695622254163, 0.1375272274017334, -0.047882549464702606, -0.0939425602555275, -0.12456146627664566, 0.2351711541414261, -0.07182266563177109, 0.018624672666192055, 0.03858436271548271, -0.07166220247745514, -0.04573313146829605, 0.19470296800136566, 0.15958625078201294, 0.02543407306075096, 0.007235580589622259, 0.02620537206530571, -0.007937698625028133, 0.01064593717455864, 0.09273688495159149, -0.02344539202749729, 0.36969515681266785, -0.038015298545360565, 0.18440663814544678, -0.024211741983890533, -0.010250333696603775, -0.02517731674015522, 0.03804948180913925, -0.026624318212270737, 0.034791249781847, -0.054125890135765076, 0.059327226132154465, -0.13211555778980255, -0.2537861168384552, -0.023841172456741333, 0.03838615119457245, -0.07592599838972092, -0.004121727775782347, 0.020054331049323082, 0.042551860213279724, 0.07620715349912643, -0.0005683943745680153, 0.042439643293619156, 0.015464740805327892, 0.03261837735772133, -0.11222727596759796, -0.07818136364221573, 0.08565661311149597, -0.014179439283907413, 0.1835518479347229, 0.023794418200850487, 0.031967949122190475, 0.09380848705768585, -0.041959598660469055, -0.10403823852539062, 0.06727910041809082, -0.005507636349648237, -0.08849386870861053, 0.009749392978847027, 0.14503887295722961, -0.035057295113801956, 0.1171862930059433, 0.015940338373184204, -0.01881704106926918, 0.046888768672943115, -0.058715421706438065, -0.05783495306968689, -0.08167858421802521, 0.06753842532634735, -0.07844490557909012, 0.11414279788732529, 0.12390980869531631, -0.014930400997400284, -0.032770149409770966, -0.035400934517383575, 0.03719867020845413, -0.07859575748443604, 0.013350642286241055, -0.0365205816924572, -0.11445332318544388, -0.0491795539855957, -0.007982802577316761, 0.06501396000385284, -0.25994905829429626, -0.01667979545891285, -0.0003987766394857317, -0.05981513485312462, 0.0067153749987483025, 0.014161065220832825, 0.018688669428229332, 0.030589159578084946, -0.014546835795044899, -0.006617068778723478, -0.0037934191059321165, 0.03267943114042282, -0.1287386119365692, -0.05698909983038902 ]
null
null
transformers
# BatterySciBERT-uncased for Battery Abstract Classification **Language model:** batteryscibert-uncased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 14 base_LM_model = "batteryscibert-uncased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 97.12, "Test accuracy": 97.47, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/batteryscibert-uncased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/batteryscibert-uncased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatterySciBERT-uncased for Battery Abstract Classification Language model: batteryscibert-uncased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatterySciBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 74, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatterySciBERT-uncased for Battery Abstract Classification \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.0690784826874733, 0.10396218299865723, -0.004250640515238047, 0.05132226273417473, 0.24002067744731903, 0.040813639760017395, 0.20327043533325195, 0.043149519711732864, -0.08336584270000458, -0.033805962651968, 0.039766497910022736, 0.03456681966781616, 0.056017156690359116, 0.06823539733886719, -0.047821398824453354, -0.1662082076072693, 0.03689992427825928, -0.019848769530653954, 0.027264438569545746, 0.11987078934907913, 0.13373039662837982, -0.07764999568462372, 0.10805616527795792, -0.005728617776185274, -0.11437921226024628, 0.02189979888498783, 0.013335684314370155, -0.1277475655078888, 0.12013041973114014, -0.06493115425109863, 0.15910544991493225, 0.047205034643411636, 0.034381307661533356, -0.1621675044298172, 0.0335073322057724, 0.013455169275403023, -0.002838660730049014, 0.07325334846973419, 0.028042802587151527, -0.08491908758878708, 0.09755062311887741, -0.022664247080683708, 0.07562797516584396, 0.025348152965307236, -0.09982112795114517, -0.1385398507118225, -0.04495517536997795, 0.011972356587648392, 0.07584460824728012, 0.11349990963935852, -0.01759120263159275, 0.07026377320289612, -0.05483994632959366, 0.06873174756765366, 0.17949508130550385, -0.10542353242635727, 0.006624917965382338, 0.11980380117893219, -0.0002634049451444298, -0.013596629723906517, -0.06412555277347565, -0.021097172051668167, 0.008664575405418873, 0.03899136930704117, 0.049666523933410645, -0.030037760734558105, -0.164014995098114, 0.08917342126369476, -0.09447204321622849, -0.002517832675948739, 0.2424980252981186, -0.07347315549850464, -0.03749807924032211, -0.043741900473833084, -0.09228292107582092, -0.07101096212863922, -0.031956497579813004, 0.025167368352413177, -0.0288456492125988, 0.021592851728200912, -0.01222413033246994, 0.04704861342906952, -0.04475664347410202, -0.06802251189947128, -0.11255361884832382, 0.10104712098836899, 0.005729182157665491, 0.0016477735480293632, -0.09333478659391403, 0.12405264377593994, 0.0672082006931305, -0.1654464453458786, 0.017958367243409157, -0.0632435753941536, -0.03643982112407684, -0.027609996497631073, 0.012680428102612495, -0.014032714068889618, 0.09492789208889008, 0.09378429502248764, -0.027171699330210686, 0.00018649644334800541, 0.00421549379825592, -0.00973480474203825, 0.007564963772892952, 0.11601094901561737, -0.02493002824485302, -0.04725835844874382, 0.02429310232400894, 0.009029757231473923, 0.03655748814344406, -0.04590153321623802, -0.12163614481687546, -0.0024723534006625414, 0.08441995829343796, 0.07249186187982559, -0.023229029029607773, 0.05671032890677452, -0.08407670259475708, -0.06254737824201584, 0.07970631867647171, -0.09957437217235565, -0.022119266912341118, 0.008738422766327858, 0.021618733182549477, 0.04919375851750374, -0.017167892307043076, 0.028831781819462776, -0.024463443085551262, 0.07486578077077866, -0.0898752436041832, -0.04964280128479004, -0.09428725391626358, -0.08900691568851471, -0.0016587144928053021, 0.06384722888469696, 0.05786523595452309, -0.12026845663785934, -0.2843734920024872, 0.05472700297832489, 0.0018787371227517724, -0.003358440939337015, -0.067132368683815, -0.07263782620429993, -0.06317262351512909, 0.02115979790687561, -0.0033589203376322985, 0.022432295605540276, -0.06557977944612503, 0.052873555570840836, 0.09083971381187439, 0.03558553755283356, -0.0727042406797409, 0.040295470505952835, -0.13567952811717987, -0.008538400754332542, -0.18796733021736145, 0.025991525501012802, -0.09819226711988449, 0.08878067135810852, -0.0707787349820137, -0.041670553386211395, 0.12038711458444595, -0.02422800473868847, 0.019783874973654747, 0.15605361759662628, -0.09450598061084747, -0.14240066707134247, 0.05298104137182236, -0.13646627962589264, -0.09233267605304718, 0.05277858301997185, -0.03771282359957695, 0.030017079785466194, 0.06704918295145035, 0.07837902754545212, 0.07479558140039444, -0.10056508332490921, -0.11339245736598969, -0.00039165603811852634, 0.027772383764386177, 0.003933920990675688, 0.09723062813282013, 0.034638337790966034, -0.04272632300853729, 0.06148598715662956, -0.01995578408241272, 0.056467585265636444, -0.038080260157585144, -0.09464368969202042, -0.023553533479571342, -0.03256973251700401, 0.1395905464887619, 0.027365777641534805, 0.010244394652545452, -0.0030151621904224157, -0.08353912830352783, 0.14037065207958221, 0.10074849426746368, -0.04486285522580147, 0.007512132171541452, -0.12977665662765503, 0.07013987749814987, 0.04042975604534149, 0.016318872570991516, -0.1155087873339653, -0.10587280988693237, 0.04740789160132408, 0.031059231609106064, -0.02958235889673233, -0.0796409547328949, -0.002959297737106681, 0.03524082526564598, -0.1035284623503685, -0.01651616394519806, -0.07691500335931778, 0.000029085265850881115, 0.0004795980639755726, -0.2233097404241562, 0.08415255695581436, -0.036004044115543365, 0.17500187456607819, -0.1123163029551506, 0.06085330620408058, 0.14339783787727356, 0.08058246970176697, -0.017575891688466072, -0.007904096506536007, 0.010905034840106964, 0.036799877882003784, -0.06547210365533829, -0.048192426562309265, 0.005619456991553307, -0.002524344250559807, -0.06707064807415009, 0.004524784162640572, -0.1479223668575287, 0.09887196123600006, 0.16863150894641876, -0.003652357030659914, -0.03978893160820007, -0.040159162133932114, 0.00776778906583786, -0.023018022999167442, -0.0933830514550209, 0.00025011325487867, 0.2097088247537613, -0.020693086087703705, 0.07370792329311371, -0.04011712968349457, -0.010767667554318905, -0.007980677299201488, 0.02880094200372696, 0.044828418642282486, 0.11711963266134262, 0.09734371304512024, -0.06302829086780548, 0.10303883999586105, 0.05607191473245621, -0.06676926463842392, 0.16079451143741608, -0.06530056893825531, -0.01613817922770977, -0.0038513867184519768, -0.0999477431178093, -0.0014081053668633103, 0.20246557891368866, -0.16377554833889008, 0.020277736708521843, 0.055266089737415314, -0.02629007026553154, 0.01559604611247778, -0.07471256703138351, -0.025108730420470238, -0.014940986409783363, -0.020779917016625404, -0.10935796797275543, 0.014046251773834229, 0.03739634156227112, 0.06512149423360825, -0.010915586724877357, -0.08218694478273392, 0.0709652528166771, 0.010253630578517914, -0.10348007827997208, 0.22563369572162628, -0.12624581158161163, -0.24587886035442352, -0.12308990955352783, -0.06306897848844528, -0.025290805846452713, -0.01132821012288332, 0.02663891762495041, -0.11306234449148178, -0.09501384943723679, -0.005041621159762144, 0.09479901194572449, -0.021008430048823357, 0.0674143135547638, 0.07352966070175171, 0.017896868288517, -0.027417302131652832, -0.10327031463384628, 0.007314383052289486, 0.03981306776404381, -0.052884843200445175, 0.04664209857583046, 0.031052084639668465, 0.0508919321000576, 0.18982112407684326, 0.010433572344481945, 0.013312582857906818, -0.04876840114593506, 0.14034511148929596, -0.11552008241415024, -0.03693290799856186, 0.20194631814956665, 0.03995664417743683, 0.0030806444119662046, 0.17201825976371765, 0.004956925753504038, -0.029500365257263184, 0.08020088821649551, -0.024280305951833725, -0.04969537630677223, -0.27279943227767944, -0.13030366599559784, -0.05902913957834244, 0.0907125249505043, -0.035434335470199585, 0.03605647385120392, 0.04174090549349785, 0.053343504667282104, -0.005620548967272043, 0.06390123069286346, -0.032901469618082047, 0.12160155922174454, 0.22286026179790497, 0.06439965963363647, 0.08298283070325851, -0.02791459858417511, -0.027777167037129402, 0.07122009247541428, -0.04613663628697395, 0.12877750396728516, 0.002049392322078347, 0.0871005654335022, 0.011206978000700474, 0.009666983038187027, 0.06977275013923645, 0.04206506162881851, 0.01082642748951912, -0.00247158482670784, 0.021017542108893394, -0.04353751987218857, 0.04477864131331444, -0.017733730375766754, -0.08981145173311234, -0.008519740775227547, -0.04242659732699394, 0.07590202242136002, 0.08786142617464066, -0.049521952867507935, 0.036017272621393204, -0.22347354888916016, -0.05186392739415169, 0.058288272470235825, -0.04762721061706543, -0.09275836497545242, 0.024640044197440147, -0.015288718044757843, -0.12564995884895325, 0.07926922291517258, -0.011255993507802486, 0.0757790207862854, -0.12532880902290344, 0.05571121722459793, 0.030412733554840088, 0.06290382891893387, -0.01480147521942854, 0.1156325712800026, -0.3574489951133728, 0.1287582516670227, 0.0487130843102932, 0.05417498201131821, -0.018287021666765213, 0.031093720346689224, -0.007422878872603178, 0.05984123796224594, 0.02997511439025402, 0.0032087210565805435, 0.03363579139113426, 0.032002341002225876, -0.13216130435466766, 0.05322229117155075, 0.02796158567070961, -0.05245821550488472, 0.028503727167844772, -0.0121328579261899, 0.021472161635756493, 0.02591126598417759, -0.08353855460882187, -0.1372334510087967, -0.12574097514152527, 0.09401869028806686, -0.027195759117603302, 0.11876065284013748, -0.0980815514922142, -0.09781035035848618, -0.007498768158257008, 0.12056344747543335, -0.13431806862354279, -0.10759475827217102, -0.14556024968624115, 0.0025429148226976395, 0.05261348560452461, -0.06966057419776917, 0.0625293105840683, 0.00008033032645471394, 0.0778135359287262, -0.04719929397106171, -0.10218928754329681, 0.1021733507514, -0.05897201597690582, -0.1871669441461563, -0.05633721128106117, 0.09292801469564438, 0.09738803654909134, 0.060592327266931534, -0.020356103777885437, -0.0007726267795078456, -0.08272213488817215, -0.13025733828544617, 0.0020677417051047087, 0.0017301221378147602, 0.05492641404271126, 0.04923269897699356, 0.06721960008144379, -0.09450776875019073, -0.030919505283236504, -0.0759832039475441, 0.08162417262792587, 0.18658168613910675, -0.021599527448415756, 0.10677996277809143, 0.09774438291788101, -0.06073254346847534, -0.23924620449543, -0.044714782387018204, 0.004015081096440554, 0.02481055073440075, -0.049147624522447586, -0.27132296562194824, 0.12135423719882965, -0.021820586174726486, -0.06459108740091324, -0.018983975052833557, -0.15310607850551605, -0.13465909659862518, 0.19882464408874512, 0.015483930706977844, 0.09768830984830856, -0.10092505067586899, -0.03968733921647072, -0.09831102937459946, -0.18494769930839539, 0.22321166098117828, -0.15919849276542664, 0.054060108959674835, -0.01661236584186554, 0.17821934819221497, -0.010970661416649818, -0.03179197385907173, 0.13904543220996857, 0.04271603375673294, 0.008237370289862156, -0.01956736482679844, -0.09062936902046204, 0.20924468338489532, 0.030066117644309998, 0.010411207564175129, -0.0466231107711792, 0.05326497554779053, -0.10444027930498123, -0.040729690343141556, -0.019361380487680435, 0.11018431186676025, -0.03747175261378288, -0.14782564342021942, -0.101396344602108, 0.030947251245379448, -0.02601715922355652, -0.04544258117675781, 0.19468259811401367, 0.06134725734591484, 0.03571276366710663, 0.03308340162038803, 0.2111268788576126, 0.016604308038949966, 0.0625796765089035, 0.0037093295250087976, -0.022884700447320938, 0.05737243592739105, -0.1857381910085678, 0.02548159845173359, 0.0948096364736557, 0.04156365618109703, 0.047551535069942474, 0.0901731327176094, -0.06322010606527328, -0.020909860730171204, 0.09985776245594025, -0.227189302444458, -0.02938011661171913, -0.09108958393335342, 0.15105724334716797, -0.13121719658374786, 0.0797818973660469, 0.11823712289333344, -0.03030487895011902, -0.05788479000329971, 0.05062170326709747, 0.0418006107211113, -0.0272049680352211, 0.12634453177452087, 0.059813208878040314, 0.10548454523086548, -0.13566447794437408, 0.08261416852474213, 0.08306244760751724, 0.028715040534734726, -0.03108282946050167, 0.09095511585474014, -0.13728556036949158, -0.02020341157913208, 0.001999892061576247, 0.19324500858783722, -0.11498314142227173, -0.047862399369478226, -0.1831957846879959, -0.1376638114452362, 0.018588759005069733, 0.13747423887252808, 0.10476479679346085, 0.0958777368068695, -0.047742605209350586, 0.008592521771788597, -0.0954258069396019, 0.08676877617835999, 0.002750702667981386, 0.029768850654363632, -0.14451920986175537, 0.014837505295872688, -0.012844673357903957, 0.037631306797266006, -0.05885789543390274, -0.0011693082051351666, -0.15601316094398499, 0.015110567212104797, -0.053247276693582535, -0.0554569736123085, -0.03229252249002457, -0.02023467794060707, 0.00723861251026392, -0.039914343506097794, -0.015340665355324745, 0.024797452613711357, -0.044802650809288025, 0.05162384733557701, 0.028258394449949265, 0.062495701014995575, -0.10899213701486588, 0.061106547713279724, 0.06625485420227051, -0.05834253877401352, 0.1359826624393463, 0.0844738557934761, 0.07081587612628937, 0.0960438922047615, -0.18079470098018646, 0.02456723339855671, 0.04423508420586586, 0.04165241867303848, 0.03371778503060341, -0.11609989404678345, -0.015879519283771515, -0.008098861202597618, 0.028477372601628304, 0.057173360139131546, 0.05667288601398468, -0.09862284362316132, 0.07106101512908936, -0.04956725984811783, -0.009417022578418255, -0.05813824012875557, 0.03631249815225601, 0.13500401377677917, 0.039088841527700424, 0.12014266848564148, -0.06304151564836502, 0.005308473948389292, -0.0780959203839302, 0.027593472972512245, -0.01855144463479519, -0.06920640915632248, -0.0867001935839653, -0.033863410353660583, 0.05139018967747688, 0.0068186866119503975, 0.20231011509895325, 0.006134646944701672, -0.026401879265904427, 0.05210581049323082, 0.16712993383407593, 0.04763951152563095, 0.03470837324857712, 0.11322874575853348, 0.10913825035095215, 0.04412892460823059, 0.07278802990913391, -0.010668866336345673, 0.11195529252290726, 0.0790649950504303, 0.1492612063884735, 0.03443662077188492, -0.007009717635810375, 0.035823892802000046, 0.013203728944063187, -0.09452244639396667, -0.13580051064491272, 0.0458141565322876, -0.0553445890545845, 0.032735154032707214, -0.04325961694121361, 0.17459511756896973, 0.11116188019514084, -0.09060534834861755, 0.05163436010479927, -0.018803779035806656, -0.10906627774238586, -0.10311990976333618, -0.0010672130156308413, -0.037845443934202194, -0.17078223824501038, 0.024619126692414284, -0.13509763777256012, -0.0670018345117569, 0.1440298855304718, -0.006861486006528139, -0.04099597409367561, 0.08827938884496689, 0.1674502193927765, -0.01757759228348732, 0.0445159450173378, -0.013928902335464954, -0.02561839483678341, -0.02694578468799591, -0.0017854943871498108, -0.07230853289365768, 0.08122031390666962, 0.015777505934238434, -0.009717532433569431, -0.029743555933237076, 0.006551725789904594, -0.01778968796133995, -0.06383505463600159, 0.015330805443227291, 0.0031996446195989847, -0.08340934664011002, 0.16765190660953522, -0.005758845247328281, 0.05108754336833954, 0.04746827483177185, 0.2843785881996155, -0.06204164773225784, -0.15331599116325378, -0.09523755311965942, 0.0991716980934143, -0.026272213086485863, 0.03733867406845093, 0.052494294941425323, -0.1043982282280922, -0.08713531494140625, 0.152216836810112, 0.11896383762359619, -0.01538004819303751, -0.008129923604428768, -0.0066183386370539665, -0.010172725655138493, -0.004043458495289087, 0.020728284493088722, 0.061911433935165405, 0.15376093983650208, -0.06401777267456055, -0.03585587814450264, -0.046706900000572205, -0.05482301115989685, -0.05790737271308899, 0.059358369559049606, 0.037894751876592636, -0.03335624560713768, -0.03064034879207611, 0.070323146879673, -0.1806734949350357, -0.10601462423801422, -0.0982714593410492, -0.10920106619596481, -0.09695009142160416, -0.06309991329908371, 0.15245971083641052, -0.006552234757691622, 0.007449693977832794, -0.014203856699168682, 0.030618220567703247, 0.009560991078615189, 0.044525861740112305, -0.13746444880962372, -0.010742759332060814, 0.12256136536598206, -0.16478633880615234, 0.13891679048538208, -0.05570047348737717, 0.06523812562227249, 0.11458505690097809, 0.015250107273459435, -0.07386354357004166, 0.020702114328742027, 0.0394371896982193, -0.12361464649438858, 0.010818014852702618, 0.12692369520664215, -0.004986007232218981, 0.1004326343536377, 0.011999729089438915, 0.018308300524950027, 0.002477552043274045, -0.0006437833653762937, -0.07998152822256088, -0.05464095622301102, -0.03386872634291649, -0.07169034332036972, 0.0620175339281559, 0.07570038735866547, -0.06201664358377457, 0.02501056343317032, -0.06487882882356644, 0.04291238263249397, -0.017303938046097755, 0.0030305064283311367, -0.01693563535809517, -0.13504286110401154, -0.05581027641892433, 0.18480798602104187, 0.026949957013130188, -0.16364219784736633, 0.024032944813370705, -0.05838531255722046, 0.0021909435745328665, -0.14683635532855988, 0.09790390729904175, 0.078441321849823, -0.0335853174328804, -0.025152992457151413, -0.0054936278611421585, -0.02461504377424717, 0.07674535363912582, -0.16592977941036224, -0.09128380566835403 ]
null
null
transformers
# BatterySciBERT-uncased for QA **Language model:** batteryscibert-uncased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 3 base_LM_model = "batteryscibert-uncased" max_seq_len = 386 learning_rate = 2e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 79.81, "f1": 87.66, ``` Evaluated on the battery device dataset. ``` "precision": 66.65, "recall": 85.29, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/batteryscibert-uncased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/batteryscibert-uncased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BatterySciBERT-uncased for QA Language model: batteryscibert-uncased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-uncased for QA \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BatterySciBERT-uncased for QA \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 69, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BatterySciBERT-uncased for QA \r\nLanguage model: batteryscibert-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.07842893153429031, 0.03512030467391014, -0.004168678540736437, 0.06044456735253334, 0.2351476401090622, 0.04356449469923973, 0.14292463660240173, 0.04534607008099556, -0.0868968665599823, -0.02720780298113823, 0.02862444333732128, -0.031762778759002686, 0.0724911093711853, 0.06883685290813446, 0.009129505604505539, -0.10524318367242813, 0.03885357454419136, -0.042949944734573364, 0.0426681824028492, 0.138103649020195, 0.11126072704792023, -0.08451574295759201, 0.11271123588085175, 0.013143558986485004, -0.10323674231767654, 0.06851489096879959, -0.023593880236148834, -0.07189091295003891, 0.08874589949846268, -0.09487683326005936, 0.13964076340198517, 0.04102899879217148, 0.03183910623192787, -0.18616048991680145, 0.03338339924812317, 0.03003288432955742, 0.03186758980154991, 0.018852703273296356, 0.0341448150575161, -0.045069370418787, 0.040725793689489365, 0.013936683535575867, 0.05256171151995659, 0.014544439502060413, -0.05551951006054878, -0.22250811755657196, -0.05384383350610733, -0.06517817825078964, 0.09111291170120239, 0.1500571072101593, -0.03303421661257744, 0.09178055822849274, -0.04953631013631821, 0.08043408393859863, 0.1527714729309082, -0.1098645031452179, -0.0021196678280830383, 0.11202773451805115, 0.031376518309116364, -0.028251660987734795, -0.012367292307317257, -0.09003017097711563, -0.00022476814046967775, 0.055555783212184906, 0.021028677001595497, -0.05737406015396118, -0.1600378155708313, 0.09279883652925491, -0.10589950531721115, 0.002365369349718094, 0.20381753146648407, -0.04823804646730423, -0.0684153363108635, -0.0018416860839352012, -0.11829794198274612, -0.019573215395212173, -0.02069975808262825, -0.003757522441446781, -0.02796170860528946, 0.0019483587238937616, -0.060680728405714035, 0.03763330727815628, -0.022950192913413048, -0.10537306219339371, -0.12430629134178162, 0.10489846020936966, 0.07011616230010986, 0.043009646236896515, -0.13121230900287628, 0.0815456286072731, 0.004121307749301195, -0.16739124059677124, -0.03080601617693901, -0.06270278245210648, -0.08192586153745651, -0.004711683373898268, 0.017254836857318878, 0.008525305427610874, 0.12086905539035797, 0.08387459814548492, -0.06085502728819847, 0.01636306755244732, 0.03140250965952873, -0.06569823622703552, 0.022238001227378845, 0.13621503114700317, -0.027841433882713318, -0.13082963228225708, -0.019181368872523308, 0.02831888198852539, 0.0008059716201387346, -0.00912799034267664, -0.06470555067062378, 0.007695897947996855, 0.05969870090484619, 0.06481324881315231, 0.010480745695531368, 0.02669842354953289, -0.06380344927310944, -0.03976459428668022, 0.03036171942949295, -0.10518332570791245, -0.04042791575193405, -0.02863214910030365, 0.003187985625118017, 0.019915543496608734, -0.04448634758591652, 0.020777232944965363, -0.0486479252576828, -0.01012862753123045, -0.07263032346963882, -0.0809774100780487, -0.13084709644317627, -0.08118638396263123, -0.0405658595263958, 0.12256163358688354, 0.07593132555484772, -0.11106394976377487, -0.34870949387550354, 0.07381858676671982, 0.004137859214097261, -0.006650824099779129, -0.0353815071284771, -0.04544442147016525, -0.03822761029005051, -0.02845548279583454, -0.03065653145313263, 0.07897114753723145, -0.06658805161714554, 0.050568751990795135, 0.15464116632938385, 0.02689800038933754, -0.027236614376306534, 0.04120813310146332, -0.15039631724357605, -0.00009117786976275966, -0.12110944092273712, 0.04404866322875023, -0.1132020652294159, 0.010038300417363644, -0.11734696477651596, -0.08586766570806503, 0.1159876212477684, -0.014824348501861095, 0.08360790461301804, 0.14310361444950104, -0.08681480586528778, -0.11216314882040024, 0.05504179373383522, -0.132351353764534, -0.13482853770256042, 0.05038117617368698, -0.013963251374661922, 0.030943680554628372, 0.020319290459156036, 0.06018180400133133, 0.11304619163274765, -0.16769954562187195, -0.1579016000032425, -0.019056376069784164, 0.047530367970466614, -0.0028805152978748083, 0.08002857863903046, 0.03483913093805313, -0.0697852373123169, 0.07115942239761353, -0.044595539569854736, 0.04112158343195915, -0.048830315470695496, -0.1068277508020401, -0.054688066244125366, -0.03584150969982147, 0.09711983054876328, 0.017164792865514755, -0.04940610006451607, 0.011270575225353241, -0.09407777339220047, 0.14079201221466064, 0.1257544606924057, -0.05303078517317772, -0.010150874964892864, -0.18394702672958374, 0.08316618204116821, -0.027140868827700615, 0.01101149432361126, -0.11260978132486343, -0.17058762907981873, 0.0911717340350151, -0.04537590965628624, -0.07248087227344513, -0.04174092039465904, 0.01937541738152504, 0.03493528068065643, -0.10139631479978561, -0.03630286082625389, -0.12943115830421448, -0.020704513415694237, 0.012905625626444817, -0.17095117270946503, 0.05537313222885132, -0.03189026936888695, 0.10392532497644424, 0.02862715534865856, 0.021436650305986404, 0.09678810089826584, 0.038764838129282, -0.042628463357686996, -0.018505627289414406, -0.009904450736939907, 0.058236125856637955, -0.029667165130376816, -0.042832545936107635, -0.026551896706223488, -0.01269008219242096, 0.014054007828235626, -0.04855038970708847, -0.13008558750152588, 0.04245644062757492, 0.1446182131767273, 0.09922619163990021, 0.0017838265048339963, -0.06515492498874664, 0.0047021228820085526, -0.048773664981126785, -0.09987422078847885, -0.00036359752994030714, 0.22243693470954895, -0.008042663335800171, 0.02288397029042244, -0.03588029369711876, 0.0020989307668060064, -0.01086251437664032, 0.030014406889677048, 0.05844688415527344, 0.08331616967916489, 0.09481799602508545, 0.012985343113541603, 0.05992838367819786, 0.054839201271533966, -0.07927033305168152, 0.13814854621887207, -0.0717056542634964, -0.03175560012459755, -0.04246405139565468, -0.0934228003025055, -0.031068595126271248, 0.2616545557975769, -0.152490496635437, 0.07036950439214706, 0.05842012166976929, -0.015750519931316376, 0.03413708880543709, -0.03319624438881874, -0.012414440512657166, -0.018847770988941193, -0.05065936967730522, -0.1544438898563385, 0.020478466525673866, 0.052853845059871674, 0.008849434554576874, -0.019619734957814217, -0.06340686231851578, 0.01772833615541458, 0.007143383380025625, -0.07414988428354263, 0.2356971949338913, -0.11966650933027267, -0.19177046418190002, -0.09963314235210419, -0.08060120046138763, -0.029330046847462654, -0.075795017182827, 0.023518793284893036, -0.1296626776456833, -0.10726714134216309, 0.01641438901424408, 0.15931981801986694, 0.0038098234217613935, 0.06282071769237518, 0.05753553286194801, 0.0019127876730635762, 0.006049876566976309, -0.10860409587621689, 0.012589126825332642, 0.05430051311850548, -0.10898496210575104, 0.03259453549981117, 0.05074567347764969, 0.07158711552619934, 0.11752685159444809, 0.018901214003562927, 0.021023785695433617, -0.04401008412241936, 0.23924149572849274, -0.09511449187994003, -0.03289947658777237, 0.2818983793258667, 0.12624302506446838, 0.002460454124957323, 0.2095806896686554, -0.02994905784726143, -0.0200978834182024, 0.10004810988903046, -0.00446215970441699, -0.0485098622739315, -0.2520424723625183, -0.08757483959197998, -0.033022474497556686, 0.05132020637392998, -0.05423441529273987, 0.011982562020421028, -0.03102892078459263, 0.08431229740381241, 0.004817829467356205, -0.01962951011955738, -0.08817273378372192, 0.1060018241405487, 0.22120477259159088, 0.06465093791484833, 0.09859218448400497, -0.003418239299207926, 0.02478785067796707, 0.0665021687746048, 0.04799402877688408, 0.1619098037481308, -0.05487314611673355, 0.03311658650636673, 0.039832040667533875, 0.1382857710123062, 0.02568499557673931, 0.026103001087903976, -0.018570637330412865, 0.01470203697681427, 0.025326410308480263, -0.04080745577812195, 0.027900880202651024, -0.026266762986779213, -0.04725702479481697, 0.02018873579800129, -0.005654458422213793, 0.1621691882610321, 0.04900260269641876, -0.012863799929618835, 0.03667409345507622, -0.20249043405056, -0.027183258906006813, 0.04821740463376045, -0.03334987536072731, -0.10292793810367584, 0.03149664029479027, 0.014436564408242702, -0.12419481575489044, 0.04966997727751732, -0.018152713775634766, 0.06742192804813385, -0.0670776218175888, 0.050058379769325256, 0.08494150638580322, 0.12295577675104141, 0.016897311434149742, 0.12187246233224869, -0.39078980684280396, 0.08904817700386047, 0.04519534856081009, 0.09523621946573257, 0.03700421750545502, 0.0629759281873703, -0.05453747510910034, -0.04378226026892662, 0.06413894891738892, 0.001906593213789165, 0.07559438049793243, 0.04756716266274452, -0.152164027094841, 0.12187019735574722, 0.00145339360460639, 0.011291390284895897, 0.06245829910039902, -0.024571644142270088, 0.029304053634405136, -0.005717080552130938, -0.04491466283798218, -0.148317351937294, -0.0895586684346199, 0.05388139933347702, -0.0613008551299572, 0.08609427511692047, -0.13684052228927612, -0.07999018579721451, 0.05632577836513519, 0.07327187061309814, -0.07837602496147156, -0.1326957494020462, -0.1371372789144516, 0.003121413988992572, 0.11646182090044022, -0.10958287864923477, 0.07188466191291809, -0.00016843352932482958, 0.047751691192388535, -0.021818306297063828, -0.11309168487787247, 0.04667048528790474, -0.09178400784730911, -0.1360364705324173, -0.029085490852594376, 0.028278307989239693, 0.10486773401498795, 0.049205392599105835, 0.006299398373812437, -0.013675585389137268, -0.1470683068037033, -0.12344522029161453, -0.036821961402893066, -0.04892977327108383, 0.05284174904227257, 0.02474674955010414, 0.12625201046466827, -0.03486719727516174, -0.026430228725075722, -0.025967812165617943, 0.0005579471471719444, 0.14693765342235565, -0.015220520086586475, 0.04615454003214836, 0.1398378610610962, -0.024419289082288742, -0.1969301402568817, -0.0670500099658966, 0.08827541768550873, 0.0438230074942112, -0.04719555750489235, -0.25990578532218933, 0.14346635341644287, -0.019758515059947968, -0.06310465931892395, -0.11864454299211502, -0.14268921315670013, -0.12054767459630966, 0.20090703666210175, 0.0734810158610344, 0.05966951698064804, -0.07440953701734543, -0.018557308241724968, -0.05057753250002861, -0.2421753853559494, 0.13151225447654724, -0.15018266439437866, 0.022205466404557228, -0.014509797096252441, 0.23394346237182617, -0.02856481820344925, -0.033365506678819656, 0.14199113845825195, 0.06525921076536179, -0.02160174399614334, 0.02043120563030243, 0.0088139483705163, 0.12278073281049728, 0.02728119306266308, 0.05683024600148201, 0.016836533322930336, 0.11499016731977463, -0.1143018826842308, -0.04089956730604172, -0.05480034276843071, 0.08941332995891571, -0.0532018318772316, -0.1656356304883957, -0.08134431391954422, 0.04874001070857048, -0.03006146103143692, -0.059713445603847504, 0.08418580144643784, 0.05387403070926666, -0.012058250606060028, 0.09226764738559723, 0.16239602863788605, 0.04698503017425537, -0.06310365349054337, 0.040413059294223785, -0.011242171749472618, 0.05570186302065849, -0.13234233856201172, 0.026380425319075584, 0.10043071210384369, 0.02990962192416191, 0.013849934563040733, 0.044818729162216187, -0.08565833419561386, -0.012545603327453136, 0.07439415901899338, -0.15747755765914917, -0.1394323855638504, -0.06813428550958633, 0.1769743114709854, -0.18592099845409393, 0.05573606491088867, 0.10398596525192261, 0.0005544576561078429, -0.07051179558038712, 0.026707839220762253, 0.04715375602245331, -0.012147998437285423, 0.1780480444431305, 0.08150642365217209, 0.13138394057750702, -0.13326257467269897, 0.06742773205041885, 0.04578980430960655, 0.04900852590799332, -0.0320291593670845, 0.033520378172397614, -0.10167663544416428, 0.006631297059357166, 0.03274121880531311, 0.13152611255645752, -0.13276642560958862, -0.08173917979001999, -0.1603720635175705, -0.12287307530641556, -0.0017764875665307045, 0.06723064929246902, 0.09248089045286179, 0.0658544972538948, -0.026556236669421196, 0.015097065828740597, -0.08826834708452225, 0.10765589028596878, 0.013722777366638184, 0.020198872312903404, -0.11050592362880707, -0.09375909715890884, -0.02177618071436882, 0.11100083589553833, -0.05284173786640167, -0.004631747491657734, -0.10614775121212006, 0.03011203370988369, -0.14718183875083923, -0.060266848653554916, -0.03923286497592926, -0.01899798773229122, 0.025618255138397217, -0.06824205815792084, -0.03262420743703842, 0.07118755578994751, -0.03671777993440628, 0.05690419673919678, 0.04750979319214821, 0.022776257246732712, -0.1280083805322647, 0.08746954053640366, 0.0676652267575264, -0.056231338530778885, 0.10781411081552505, 0.08056815713644028, 0.07387341558933258, 0.10198702663183212, -0.07298411428928375, 0.01555189024657011, -0.003941377624869347, 0.08109710365533829, 0.07462282478809357, -0.14256907999515533, -0.001615607994608581, -0.007378754206001759, 0.010705157183110714, 0.02094985917210579, 0.06568454951047897, -0.09667064994573593, 0.08256873488426208, -0.05476094037294388, 0.024431535974144936, -0.060829706490039825, 0.041288938373327255, 0.10793774574995041, 0.08142134547233582, 0.10951521247625351, -0.04372100904583931, -0.0002610220108181238, -0.10953781008720398, 0.00592719204723835, -0.02138381637632847, -0.011318663135170937, -0.042838387191295624, -0.037792835384607315, 0.06997786462306976, 0.03393497318029404, 0.13953877985477448, -0.01210732664912939, 0.06810116022825241, 0.010750568471848965, 0.12469012290239334, 0.08966021984815598, -0.005665712058544159, 0.06554877012968063, 0.08053289353847504, 0.09385432302951813, 0.1200733333826065, -0.029513752087950706, 0.06688498705625534, 0.1078912541270256, 0.11446603387594223, 0.06144792214035988, 0.05335377901792526, 0.04181985557079315, 0.04193946719169617, -0.11785682290792465, -0.09430718421936035, 0.002639648038893938, -0.06695012748241425, 0.016830643638968468, -0.00243353471159935, 0.17152373492717743, 0.07839296758174896, -0.12164070457220078, 0.07964499294757843, -0.0037735705263912678, -0.12256422638893127, -0.11233016103506088, 0.084245964884758, -0.040005624294281006, -0.200325146317482, 0.03631840646266937, -0.15665103495121002, -0.10648013651371002, 0.1768389642238617, 0.011247700080275536, -0.009926783852279186, 0.13732941448688507, 0.19741179049015045, -0.025082042440772057, -0.046656083315610886, -0.007099710404872894, 0.016897903755307198, 0.05712209641933441, 0.03330502659082413, -0.050417210906744, 0.11480960249900818, 0.0483088418841362, 0.002658778103068471, -0.0787147805094719, 0.008089384995400906, 0.006075519137084484, -0.029002279043197632, -0.012654030695557594, 0.009009324014186859, -0.07737164944410324, 0.19222071766853333, -0.0014187548076733947, 0.08286633342504501, 0.014665513299405575, 0.30447128415107727, -0.055784329771995544, -0.18110191822052002, -0.12751170992851257, 0.06861815601587296, -0.0501774400472641, 0.054206158965826035, 0.06268470734357834, -0.08985113352537155, -0.12973538041114807, 0.14951670169830322, 0.028492921963334084, -0.014987373724579811, -0.04889483377337456, 0.03292786702513695, -0.01919807679951191, -0.05546928569674492, 0.025103572756052017, 0.0791480541229248, 0.20121897757053375, -0.07586833089590073, -0.002446456579491496, -0.051270145922899246, -0.08191947638988495, -0.028556516394019127, 0.013230843469500542, 0.016662128269672394, -0.030160365626215935, -0.01476901862770319, 0.05063929036259651, -0.12061324715614319, -0.09283068776130676, -0.08959435671567917, -0.031094364821910858, -0.12028836458921432, -0.062035731971263885, 0.22387386858463287, 0.0032046183478087187, 0.0004886768292635679, -0.04664310812950134, 0.07481683790683746, 0.05238084867596626, 0.03544951230287552, -0.09494360536336899, -0.07421691715717316, 0.17988455295562744, -0.14708858728408813, 0.14870408177375793, -0.046551115810871124, 0.09282371401786804, 0.11321684718132019, -0.0292922705411911, -0.11280292272567749, 0.05175929144024849, 0.06189151108264923, -0.17304563522338867, 0.004535106010735035, 0.14312133193016052, 0.00601576641201973, 0.11438258737325668, -0.010653113946318626, 0.07164814323186874, -0.022589700296521187, 0.005404217168688774, -0.03841139003634453, -0.05505617335438728, -0.017309624701738358, -0.07524815201759338, 0.0495261512696743, 0.07692661881446838, -0.06202492117881775, 0.05993405357003212, -0.06908412277698517, 0.10106106847524643, -0.036105766892433167, 0.08812519907951355, 0.020252514630556107, -0.11726286262273788, -0.05479319021105766, 0.126521497964859, -0.03155903145670891, -0.10569379478693008, 0.018539775162935257, 0.0008034592028707266, -0.003503053914755583, -0.10472987592220306, 0.1398942917585373, 0.06396153569221497, -0.0013574460754171014, -0.024261625483632088, 0.029927851632237434, -0.05623675510287285, 0.05874989926815033, -0.16884745657444, -0.10140406340360641 ]
null
null
transformers
# BatterySciBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the [SciBERT-uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) weights. It was introduced in [this paper](paper_link) and first released in [this repository](https://github.com/ShuHuang/batterybert). This model is uncased: it does not make a difference between english and English. ## Model description BatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the [SciBERT-uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the [SciBERT-uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at [Github](https://github.com/ShuHuang/batterybert/blob/main/corpus.txt). ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 31,090. The inputs of the model are then of the form: ``` [CLS] Sentence A [SEP] Sentence B [SEP] ``` The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by `[MASK]`. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=batterybert) to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='batterydata/batteryscibert-uncased') >>> unmasker("Hello I'm a <mask> model.") ``` Here is how to use this model to get the features of a given text in PyTorch: ```python from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batteryscibert-uncased') model = BertModel.from_pretrained('batterydata/batteryscibert-uncased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` and in TensorFlow: ```python from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('batterydata/batteryscibert-uncased') model = TFBertModel.from_pretrained('batterydata/batteryscibert-uncased') text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` ## Evaluation results Final loss: 1.095. ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": ["exbert"], "datasets": ["batterypapers"]}
fill-mask
batterydata/batteryscibert-uncased
[ "transformers", "pytorch", "tensorboard", "bert", "fill-mask", "exbert", "en", "dataset:batterypapers", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BatterySciBERT-uncased model Pretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-uncased weights. It was introduced in this paper and first released in this repository. This model is uncased: it does not make a difference between english and English. ## Model description BatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-uncased weights. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. ## Training data The BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github. ## Training procedure ### Preprocessing The texts are lowercased and tokenized using WordPiece and a vocabulary size of 31,090. The inputs of the model are then of the form: The details of the masking procedure for each sentence are the following: - 15% of the tokens are masked. - In 80% of the cases, the masked tokens are replaced by '[MASK]'. - In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace. - In the 10% remaining cases, the masked tokens are left as is. ### Pretraining The model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\(\beta_{1} = 0.9\\) and \\(\beta_{2} = 0.999\\), a weight decay of 0.01, learning rate warmup for 10,000 steps and linear decay of the learning rate after. ## Intended uses & limitations You can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked) to make decisions, such as sequence classification, token classification or question answering. For tasks such as text generation you should look at model like GPT2. ### How to use You can use this model directly with a pipeline for masked language modeling: Here is how to use this model to get the features of a given text in PyTorch: and in TensorFlow: ## Evaluation results Final loss: 1.095. ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BatterySciBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.", "## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 31,090. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.095.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BatterySciBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.", "## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.", "## Training data\n\nThe BatterySciBERT model was pretrained on the full text of battery papers only, after initialized from the SciBERT-uncased weights. The paper corpus contains a total of 400,366 battery research papers that are published from 2000 to June 2021, from the publishers Royal Society of Chemistry (RSC), Elsevier, and Springer. The list of DOIs can be found at Github.", "## Training procedure", "### Preprocessing\n\nThe texts are lowercased and tokenized using WordPiece and a vocabulary size of 31,090. The inputs of the model are\nthen of the form:\n\n\n\nThe details of the masking procedure for each sentence are the following:\n- 15% of the tokens are masked.\n- In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n- In the 10% remaining cases, the masked tokens are left as is.", "### Pretraining\n\n\nThe model was trained on 8 NVIDIA DGX A100 GPUs for 1,000,000 steps with a batch size of 256. The sequence length was limited to 512 tokens. The optimizer used is Adam with a learning rate of 2e-5, \\\\(\\beta_{1} = 0.9\\\\) and \\\\(\\beta_{2} = 0.999\\\\), a weight decay of 0.01,\nlearning rate warmup for 10,000 steps and linear decay of the learning rate after.", "## Intended uses & limitations\n\nYou can use the raw model for masked language modeling, but it's mostly intended to be fine-tuned on a downstream task.\nSee the model hub to look for fine-tuned versions on a task that\ninterests you.\n\nNote that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)\nto make decisions, such as sequence classification, token classification or question answering. For tasks such as text\ngeneration you should look at model like GPT2.", "### How to use\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\n\nand in TensorFlow:", "## Evaluation results\n\nFinal loss: 1.095.", "## Authors\nShu Huang: 'sh2009 [at] URL'\n\nJacqueline Cole: 'jmc61 [at] URL'\n\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 61, 85, 284, 98, 3, 135, 113, 128, 49, 10, 50 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #fill-mask #exbert #en #dataset-batterypapers #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BatterySciBERT-uncased model\n\nPretrained model on a large corpus of battery research papers using a masked language modeling (MLM) objective, starting with the SciBERT-uncased weights. It was introduced in\nthis paper and first released in\nthis repository. This model is uncased: it does not make a difference\nbetween english and English.## Model description\n\nBatterySciBERT is a transformers model pretrained on a large corpus of battery research papers in a self-supervised fashion, starting with the SciBERT-uncased weights. This means\nit was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. \n\nMore precisely, it was pretrained with the Masked language modeling (MLM) objective. Taking a sentence, the model\nrandomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict\nthe masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one\nafter the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to\nlearn a bidirectional representation of the sentence.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs." ]
[ -0.050156813114881516, 0.06265944242477417, -0.006124672945588827, 0.04659106209874153, 0.09742653369903564, 0.022577708587050438, 0.13966883718967438, 0.006064188666641712, -0.0002157637500204146, 0.04666838422417641, 0.038157299160957336, 0.015551126562058926, 0.05739665403962135, 0.05689741298556328, 0.10000777244567871, -0.24663527309894562, 0.04260016977787018, -0.10623487830162048, 0.09856683015823364, 0.07739134877920151, 0.059567470103502274, -0.07976590842008591, 0.029639791697263718, 0.026280643418431282, 0.0224410742521286, -0.012751711532473564, -0.008154947310686111, -0.0638190507888794, 0.0931379646062851, 0.04535112902522087, 0.08267829567193985, -0.015095279552042484, 0.022149013355374336, -0.055083852261304855, 0.026549257338047028, 0.05625363439321518, -0.004448513966053724, 0.06955474615097046, 0.040365442633628845, -0.014966635964810848, 0.1395910382270813, 0.03614573925733566, 0.0665888786315918, 0.050361573696136475, -0.10448608547449112, -0.12073911726474762, -0.03151904419064522, 0.025090903043746948, 0.017348414286971092, 0.10145450383424759, -0.03508845344185829, 0.07404948025941849, -0.034136924892663956, 0.07845965027809143, 0.18742898106575012, -0.09688696265220642, -0.00854232907295227, 0.014993652701377869, 0.019793447107076645, 0.01683623157441616, -0.009853459894657135, -0.040790192782878876, -0.031906627118587494, 0.03827986493706703, 0.06539887189865112, -0.013805761933326721, 0.009002789855003357, -0.047377072274684906, -0.14684070646762848, -0.0381920225918293, 0.09493591636419296, -0.039984576404094696, -0.09382278472185135, -0.12235784530639648, -0.08584855496883392, 0.05745339393615723, 0.014397619292140007, 0.05187438428401947, -0.027168650180101395, 0.021552108228206635, 0.049585796892642975, -0.05326586216688156, -0.06847860664129257, -0.0791921615600586, -0.11217333376407623, 0.19711503386497498, 0.026651687920093536, 0.05224636197090149, -0.05791691690683365, 0.07825538516044617, -0.08649440109729767, -0.06710711866617203, -0.06875339150428772, -0.04026932269334793, -0.10285437852144241, 0.007483739405870438, -0.025264674797654152, -0.18262767791748047, -0.03147704154253006, 0.12350686639547348, -0.031509287655353546, 0.00021926370391156524, -0.002728939289227128, 0.03347023203969002, 0.10658679157495499, 0.06785419583320618, -0.08918781578540802, 0.09734590351581573, 0.011494080536067486, -0.05162236839532852, 0.05488661304116249, -0.061116598546504974, -0.009729858487844467, 0.01624324358999729, -0.060181207954883575, 0.03900716081261635, 0.03353426605463028, 0.05764218047261238, -0.026826420798897743, -0.043294697999954224, 0.04415971040725708, -0.11175129562616348, -0.04113874211907387, 0.003740268759429455, -0.02048921398818493, 0.015867261216044426, 0.0691240057349205, -0.08517244458198547, -0.057520508766174316, 0.05802081897854805, -0.06541923433542252, -0.05888015031814575, -0.12284137308597565, -0.1406097114086151, -0.02550739422440529, -0.004063861910253763, -0.06502212584018707, -0.08516788482666016, -0.24846608936786652, -0.04649348556995392, 0.0300587248057127, 0.007292384747415781, 0.010126526467502117, -0.015409029088914394, 0.003297645365819335, -0.0033400163520127535, 0.012544399127364159, -0.030319515615701675, -0.0029901964589953423, 0.003585173049941659, -0.03275556117296219, 0.08914888650178909, 0.01218586228787899, 0.011309893801808357, -0.09472006559371948, 0.004655627533793449, -0.19840754568576813, 0.09478112310171127, -0.06781250238418579, -0.009139004163444042, -0.02731877751648426, 0.008072524331510067, -0.014603211544454098, 0.0265817791223526, 0.025032944977283478, 0.13078050315380096, -0.12666769325733185, -0.08356283605098724, 0.20433945953845978, -0.1665598303079605, 0.052175868302583694, 0.09549649804830551, -0.05095831677317619, 0.08438055962324142, 0.11136848479509354, -0.011968818493187428, 0.10251767933368683, -0.09315313398838043, -0.053555428981781006, 0.04626390337944031, -0.08833396434783936, 0.10073690861463547, 0.09592948108911514, -0.02402554079890251, -0.09584738314151764, 0.02389167994260788, 0.028227901086211205, 0.031138991937041283, -0.013341374695301056, -0.030093906447291374, 0.03508012741804123, -0.01909368298947811, 0.0491887666285038, 0.010858747176826, -0.014448960311710835, 0.030716082081198692, -0.09120627492666245, 0.06366220861673355, 0.11708573997020721, -0.061809979379177094, 0.001781361410394311, -0.10603685677051544, 0.03963842988014221, -0.018658354878425598, -0.017695125192403793, -0.17413824796676636, -0.02190624736249447, 0.023185765370726585, -0.1194617971777916, 0.10205493867397308, -0.012566624209284782, 0.02995259128510952, 0.11199105530977249, -0.04367505759000778, 0.02416679635643959, -0.028067994862794876, -0.040764156728982925, -0.020710989832878113, -0.1368417739868164, 0.0027479117270559072, -0.053983237594366074, 0.05994020774960518, 0.028785573318600655, 0.040331173688173294, -0.04043436422944069, -0.052142709493637085, 0.005708094220608473, -0.05684863403439522, 0.04086242616176605, -0.014130951836705208, -0.018987052142620087, -0.01729138381779194, -0.013121484778821468, 0.037051405757665634, -0.07663321495056152, 0.014946481212973595, -0.18184784054756165, -0.15901054441928864, 0.05469753220677376, 0.05556892231106758, -0.12069279700517654, 0.05748455971479416, -0.004208545666188002, -0.002786970464512706, -0.12487180531024933, -0.05201460048556328, 0.19831320643424988, 0.017697317525744438, 0.029164180159568787, -0.059004053473472595, 0.027412107214331627, 0.015347610227763653, -0.019425097852945328, 0.020196275785565376, 0.0906151756644249, 0.06551545858383179, -0.09710460156202316, 0.022962545976042747, -0.10913462191820145, -0.05505836755037308, 0.18558038771152496, 0.04434364289045334, -0.05434161052107811, -0.06586819142103195, -0.015505693852901459, 0.03358881175518036, 0.0385783351957798, -0.08171537518501282, 0.0295444093644619, 0.0440264530479908, -0.00979885458946228, 0.019939405843615532, -0.029114559292793274, 0.03593812882900238, 0.055519673973321915, 0.014263946563005447, -0.09935428947210312, -0.05997539684176445, -0.02516763098537922, 0.06742790341377258, 0.035866573452949524, 0.0016606514109298587, 0.007063425146043301, -0.03716079145669937, -0.10868076235055923, 0.17794716358184814, -0.07681870460510254, -0.22654280066490173, -0.1676316112279892, -0.022521862760186195, -0.005395475309342146, 0.031342457979917526, 0.009491308592259884, -0.01624205708503723, -0.06845547258853912, -0.07499067485332489, 0.13233132660388947, -0.0009600773337297142, -0.005729575641453266, -0.0262887142598629, -0.02829272858798504, -0.06824911385774612, -0.10613924264907837, 0.022179799154400826, -0.02522813156247139, -0.09009768068790436, 0.009537586942315102, -0.03724454715847969, 0.050013694912195206, 0.12906387448310852, 0.018526457250118256, -0.030949348583817482, -0.02237834222614765, 0.17558790743350983, -0.042586036026477814, 0.06788432598114014, 0.16295161843299866, -0.01176867913454771, 0.03955449163913727, 0.06998885422945023, 0.01994200609624386, -0.01209203153848648, 0.057746272534132004, 0.008823301643133163, -0.11435918509960175, -0.19532175362110138, -0.07241097837686539, -0.06146468222141266, 0.04142821580171585, 0.04740108177065849, 0.007586400490254164, 0.043013352900743484, 0.06893281638622284, -0.05415003001689911, -0.042342767119407654, 0.05323363468050957, 0.07547858357429504, -0.00959321204572916, -0.0350218303501606, 0.09013242274522781, -0.05693229287862778, -0.011424942873418331, 0.10149102658033371, -0.13150592148303986, 0.17541258037090302, -0.06193006783723831, 0.13922086358070374, 0.05552295222878456, -0.01800715923309326, 0.06775186955928802, 0.15142636001110077, -0.06305980682373047, 0.031165283173322678, -0.03796816244721413, -0.07980762422084808, -0.015648579224944115, 0.04811937361955643, -0.005239816848188639, 0.017060773447155952, -0.02817634865641594, 0.03262471780180931, 0.012427253648638725, 0.13794384896755219, 0.006363617721945047, -0.13241656124591827, -0.10642757266759872, 0.007262115832418203, -0.0008085676236078143, -0.08777089416980743, 0.016429170966148376, 0.09579580277204514, -0.09609674662351608, 0.05524710193276405, 0.0028498941101133823, 0.058028385043144226, -0.07090406119823456, 0.009332316927611828, -0.07542002201080322, 0.1031603068113327, -0.04881051555275917, 0.10535955429077148, -0.1625368446111679, 0.0480467714369297, 0.02286648005247116, 0.10066592693328857, -0.09990508109331131, 0.025983257219195366, 0.01598707213997841, -0.04071963578462601, 0.1511395275592804, 0.020396661013364792, -0.09476640820503235, 0.018773822113871574, -0.07863544672727585, 0.0019782150629907846, 0.03562146797776222, -0.06009532883763313, 0.053677357733249664, 0.019114237278699875, 0.010561059229075909, 0.001962889451533556, 0.023306962102651596, -0.04190300405025482, -0.13547492027282715, 0.059312984347343445, -0.022896988317370415, -0.05038994923233986, -0.0566696859896183, -0.05259120464324951, 0.0117691894993186, 0.13828399777412415, -0.16180522739887238, -0.10188735276460648, -0.1215452030301094, -0.03232715651392937, 0.10185147076845169, -0.09166478365659714, 0.04231174662709236, -0.025228194892406464, 0.12144951522350311, -0.07359310239553452, -0.10372646898031235, 0.06582950055599213, -0.06522706151008606, -0.1342034488916397, -0.03858599066734314, 0.05045876279473305, 0.1890847384929657, 0.06756916642189026, -0.03185536712408066, 0.05498524755239487, -0.041408028453588486, -0.12785033881664276, -0.025171730667352676, 0.14310014247894287, -0.02395591326057911, 0.12473928183317184, -0.01876942254602909, -0.08855699747800827, -0.018146812915802002, -0.02165122702717781, 0.11308713257312775, 0.16763584315776825, -0.054461635649204254, 0.13948982954025269, 0.26491379737854004, -0.08354176580905914, -0.24113065004348755, -0.05501781404018402, 0.03966950997710228, 0.0667102187871933, 0.02229434624314308, -0.273319810628891, 0.06697418540716171, 0.0452076718211174, -0.02813275158405304, -0.06864653527736664, -0.19640815258026123, -0.12890633940696716, 0.148711696267128, 0.03682227432727814, 0.10694507509469986, -0.023855283856391907, -0.01312390249222517, -0.060310009866952896, 0.06481354683637619, 0.16411235928535461, -0.09026926755905151, 0.10351274162530899, 0.020307347178459167, 0.02081998810172081, 0.03361993655562401, -0.06531541794538498, 0.08674939721822739, -0.019327353686094284, 0.040178313851356506, -0.049559276551008224, -0.03018820844590664, 0.12380511313676834, -0.013842780143022537, 0.10918143391609192, 0.06901662796735764, 0.0673668161034584, -0.007486466318368912, -0.09724489599466324, -0.07705911993980408, 0.07883285731077194, -0.024042129516601562, -0.08699259907007217, -0.09471209347248077, 0.07783543318510056, 0.05364149808883667, -0.017718560993671417, 0.016575220972299576, -0.06298227608203888, 0.07470332831144333, 0.13095369935035706, 0.13270792365074158, 0.04995688050985336, -0.058985717594623566, 0.0024638629984110594, -0.0179840587079525, 0.08290717750787735, -0.02962270937860012, 0.031248601153492928, 0.05724715441465378, 0.07478537410497665, 0.07692289352416992, 0.06410007178783417, -0.16652151942253113, 0.005882426630705595, 0.033335231244564056, -0.1466674655675888, -0.13893842697143555, -0.016194621101021767, 0.08466119319200516, -0.13226577639579773, -0.029597146436572075, 0.10635732114315033, -0.023095333948731422, -0.019211968407034874, 0.0023140194825828075, 0.04806853458285332, 0.012523642741143703, 0.13137617707252502, 0.019186770543456078, 0.04405095428228378, -0.07691054046154022, 0.059403806924819946, 0.04907573387026787, 0.025771381333470345, 0.01615866646170616, 0.1143711507320404, -0.07105045020580292, -0.04584784805774689, -0.03959324210882187, 0.09046077728271484, -0.040602877736091614, -0.041219793260097504, -0.028888899832963943, -0.0914081260561943, 0.053755030035972595, 0.14885197579860687, 0.014626684598624706, 0.053778838366270065, -0.07754608988761902, 0.03166718780994415, -0.05149148404598236, 0.039329927414655685, -0.044333577156066895, -0.02554047480225563, 0.020561911165714264, 0.12557755410671234, 0.019743235781788826, 0.022133074700832367, -0.051352594047784805, -0.08781570196151733, -0.10858911275863647, 0.01921842060983181, 0.01818055845797062, 0.016341380774974823, -0.06677044183015823, -0.019430529326200485, -0.0026248213835060596, 0.045234713703393936, 0.026905786246061325, 0.002608124865218997, -0.031109176576137543, 0.02196144312620163, -0.04788489639759064, -0.011301432736217976, -0.1127382218837738, 0.03626447543501854, 0.04610540345311165, -0.040269386023283005, 0.09625858813524246, 0.04422326758503914, -0.012831361033022404, 0.00903769489377737, -0.11126728355884552, 0.037222668528556824, -0.029209667816758156, 0.028442751616239548, -0.039050810039043427, -0.1068413034081459, 0.005664583761245012, -0.03037157654762268, -0.034464411437511444, 0.0003073138650506735, -0.008676843717694283, -0.07154178619384766, 0.15482759475708008, 0.09755992144346237, -0.011116674169898033, -0.10896765440702438, 0.04456785321235657, -0.0018296154448762536, 0.02697184309363365, 0.06358616054058075, -0.02453465573489666, 0.06812281161546707, -0.10159949213266373, 0.00975134875625372, 0.02108858898282051, 0.014261221513152122, 0.019340386614203453, -0.04152578487992287, 0.060931023210287094, 0.012156881392002106, 0.12688498198986053, 0.02023913525044918, -0.05750849470496178, -0.002391807036474347, 0.08079112321138382, 0.03984846919775009, -0.0243474543094635, -0.0031429689843207598, 0.020358119159936905, -0.034483652561903, 0.04356240853667259, 0.006235388107597828, -0.026487762108445168, 0.06696562469005585, 0.15741969645023346, -0.022309208288788795, 0.17694091796875, -0.012611319310963154, 0.033211320638656616, 0.0027815166395157576, -0.06996367126703262, -0.04741896688938141, 0.02671891637146473, 0.00983651913702488, -0.015219324268400669, 0.015957899391651154, 0.09438713639974594, -0.050632864236831665, 0.1659349501132965, 0.04990209639072418, -0.05242406949400902, -0.12799900770187378, -0.23210878670215607, 0.008890220895409584, 0.05761956050992012, -0.006643294356763363, -0.10673901438713074, -0.025764837861061096, 0.07904241979122162, 0.018185144290328026, -0.0029069138690829277, 0.10040202736854553, -0.09406111389398575, -0.05849102884531021, 0.014298638328909874, -0.03138275817036629, 0.008419632911682129, 0.012471168301999569, 0.01079691480845213, 0.049627527594566345, 0.09772539883852005, 0.0498826839029789, 0.06133843958377838, 0.041002076119184494, 0.03194061294198036, -0.02906576171517372, -0.06972511857748032, -0.020055346190929413, -0.0177592933177948, -0.015086513943970203, 0.1723344624042511, 0.0038265667390078306, -0.03692216798663139, 0.00591686088591814, 0.14738790690898895, -0.06089340150356293, -0.09567137807607651, -0.11448097974061966, 0.24479712545871735, -0.045037299394607544, 0.012669186107814312, 0.014205912128090858, -0.07644297927618027, -0.023295918479561806, 0.18385455012321472, 0.14309275150299072, 0.026845773681998253, 0.0070687853731215, 0.02128850482404232, -0.0002376698248554021, 0.0009576622978784144, 0.07528956979513168, -0.005954192019999027, 0.38325008749961853, -0.041392676532268524, 0.19619709253311157, -0.017011890187859535, -0.014285196550190449, -0.03786221146583557, 0.024448132142424583, -0.019702123478055, 0.02873094566166401, -0.041546717286109924, 0.0685582235455513, -0.15213556587696075, -0.2355353683233261, -0.022893261164426804, 0.039558932185173035, -0.06832877546548843, -0.008646626956760883, 0.024308206513524055, 0.048201192170381546, 0.08559362590312958, -0.007740630302578211, 0.02924943156540394, 0.027810221537947655, 0.030180098488926888, -0.10648684203624725, -0.08810717612504959, 0.07986811548471451, -0.010645990259945393, 0.17387080192565918, 0.006337565835565329, 0.04620450735092163, 0.09486771374940872, -0.043594930320978165, -0.10258354246616364, 0.06781864911317825, -0.012538495473563671, -0.10539240390062332, 0.0129663972184062, 0.1535032093524933, -0.0390649177134037, 0.10980738699436188, 0.009206090122461319, -0.016947675496339798, 0.04492586478590965, -0.03542429953813553, -0.032941948622465134, -0.07972213625907898, 0.07285048067569733, -0.07618092745542526, 0.1259818822145462, 0.13677890598773956, -0.00605544401332736, -0.03461449593305588, -0.0439426563680172, 0.025935590267181396, -0.05188906565308571, 0.005221770144999027, -0.02962014637887478, -0.10207395255565643, -0.0266928318887949, 0.023421822115778923, 0.05166767165064812, -0.24200744926929474, -0.029275719076395035, -0.005418610759079456, -0.060242220759391785, -0.005320003721863031, 0.006485911086201668, 0.02384578250348568, 0.025449542328715324, -0.01571575179696083, -0.056521542370319366, 0.004329377319663763, 0.04369591549038887, -0.1295972764492035, -0.07216496020555496 ]
null
null
transformers
# BERT-base-cased for Battery Abstract Classification **Language model:** bert-base-cased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 15 base_LM_model = "bert-base-cased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 96.84, "Test accuracy": 96.83, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/bert-base-cased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/bert-base-cased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BERT-base-cased for Battery Abstract Classification Language model: bert-base-cased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BERT-base-cased for Battery Abstract Classification \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BERT-base-cased for Battery Abstract Classification \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 71, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BERT-base-cased for Battery Abstract Classification \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.0644843727350235, 0.08330554515123367, -0.0034968836698681116, 0.051429383456707, 0.24801141023635864, 0.016842391341924667, 0.18042923510074615, 0.041375622153282166, -0.10087358206510544, -0.04119560495018959, 0.048845257610082626, 0.03551759198307991, 0.05058338865637779, 0.0712677612900734, -0.05502569302916527, -0.18833637237548828, 0.036978013813495636, 0.016822529956698418, 0.000913276569917798, 0.10959042608737946, 0.13697457313537598, -0.09884942322969437, 0.11652502417564392, 0.014170125126838684, -0.10319194197654724, 0.028500162065029144, 0.014743178151547909, -0.14727835357189178, 0.130541130900383, -0.03345576301217079, 0.16272948682308197, 0.02992299385368824, 0.05167853459715843, -0.14470553398132324, 0.030245870351791382, 0.017303485423326492, 0.0063231755048036575, 0.08654635399580002, 0.03142862394452095, -0.07274696230888367, 0.06331393867731094, -0.004852426704019308, 0.07435517758131027, 0.026204325258731842, -0.09014729410409927, -0.16322948038578033, -0.02303069457411766, 0.019966553896665573, 0.10299341380596161, 0.10741189122200012, -0.010174238122999668, 0.059801895171403885, -0.049674101173877716, 0.07340618968009949, 0.20319345593452454, -0.14872118830680847, -0.0007454867591150105, 0.08938844501972198, 0.01990348845720291, -0.016414152458310127, -0.06522298604249954, -0.02015843614935875, 0.025560738518834114, 0.04227519407868385, 0.08848655968904495, -0.02925245091319084, -0.1436881273984909, 0.10017183423042297, -0.10196198523044586, 0.006857726257294416, 0.2408287674188614, -0.058144424110651016, -0.03847009316086769, -0.02529456466436386, -0.0805341824889183, -0.07619505375623703, -0.050177618861198425, 0.010504728183150291, -0.016912007704377174, 0.015367980115115643, -0.015581578016281128, 0.03699139505624771, -0.05887705460190773, -0.07381898909807205, -0.10367965698242188, 0.11347409337759018, 0.011892315000295639, 0.0005171127850189805, -0.10463689267635345, 0.12099961191415787, 0.06567995995283127, -0.16364671289920807, 0.012877732515335083, -0.05498993396759033, -0.016968512907624245, -0.028438134118914604, 0.004539632238447666, -0.007073354907333851, 0.07545574009418488, 0.09272967278957367, -0.0023854905739426613, 0.008384162560105324, -0.01056303083896637, -0.011397100053727627, -0.00824444554746151, 0.16313068568706512, -0.04608164727687836, -0.050501883029937744, 0.04528450593352318, -0.001785307307727635, 0.019269945099949837, -0.03589024394750595, -0.14364586770534515, -0.003852994879707694, 0.08840984851121902, 0.0507592037320137, -0.045842245221138, 0.07391887903213501, -0.07215566188097, -0.060503747314214706, 0.09079226851463318, -0.10499317944049835, -0.011378439143300056, 0.019265808165073395, 0.026975098997354507, 0.03852321580052376, -0.009334791451692581, 0.025085240602493286, -0.02297709323465824, 0.07663251459598541, -0.09027084708213806, -0.0497976616024971, -0.10536283999681473, -0.1139426901936531, 0.0006581579800695181, 0.054126840084791183, 0.05865143984556198, -0.12936101853847504, -0.25887563824653625, 0.07206401973962784, 0.011898658238351345, -0.0077283382415771484, -0.06170656532049179, -0.06554646790027618, -0.08113740384578705, 0.022802112624049187, -0.012149978429079056, 0.06175104156136513, -0.06939026713371277, 0.05186503380537033, 0.09411827474832535, 0.029080329462885857, -0.07664693892002106, 0.0506778322160244, -0.1179809421300888, -0.01089886948466301, -0.20376303791999817, 0.0300101637840271, -0.09964205324649811, 0.08128634840250015, -0.07185027003288269, -0.05584009736776352, 0.12562590837478638, -0.009158944711089134, 0.024566899985074997, 0.14736005663871765, -0.08977869153022766, -0.14149385690689087, 0.03899715840816498, -0.11989517509937286, -0.09325437247753143, 0.05360295623540878, -0.04598341882228851, 0.010785683058202267, 0.08053160458803177, 0.05879412218928337, 0.051695309579372406, -0.1131751537322998, -0.08529651910066605, 0.024197790771722794, 0.0263188648968935, -0.020278293639421463, 0.1084371879696846, 0.05456945672631264, -0.0702570304274559, 0.06556013226509094, -0.058617331087589264, 0.06718185544013977, -0.036406226456165314, -0.08057145774364471, -0.005034451372921467, -0.027345968410372734, 0.12269970029592514, 0.01831245794892311, 0.030304480344057083, -0.008507604710757732, -0.09729630500078201, 0.16507768630981445, 0.07947839051485062, -0.051242511719465256, 0.008273201063275337, -0.13967597484588623, 0.08116212487220764, 0.015165342018008232, 0.0191500224173069, -0.11650826036930084, -0.08658993989229202, 0.0391056090593338, 0.07922385632991791, -0.029674142599105835, -0.055188052356243134, 0.00775719340890646, 0.03787671774625778, -0.10037785023450851, -0.014151469804346561, -0.06975335627794266, 0.007908660918474197, -0.019709091633558273, -0.22166799008846283, 0.05111445486545563, -0.043117210268974304, 0.13242508471012115, -0.11914592236280441, 0.05609365925192833, 0.1313173919916153, 0.06992824375629425, -0.011271577328443527, 0.006774624809622765, -0.010941853746771812, 0.03914574533700943, -0.0778074786067009, -0.03439078480005264, 0.009680851362645626, 0.010470523498952389, -0.060058970004320145, 0.025245577096939087, -0.12622596323490143, 0.12936092913150787, 0.1725660115480423, -0.018004771322011948, -0.04281788691878319, -0.04255321994423866, 0.0022666899021714926, -0.01583295315504074, -0.0829399898648262, -0.02714928612112999, 0.24941308796405792, -0.014916413463652134, 0.1038341373205185, -0.047292131930589676, -0.017252931371331215, -0.00471732672303915, 0.029418693855404854, 0.03631202504038811, 0.11600729078054428, 0.09076736867427826, -0.038698870688676834, 0.1177632287144661, 0.05401315167546272, -0.08213245123624802, 0.14387555420398712, -0.06081000342965126, -0.019954413175582886, 0.014130499213933945, -0.10365808010101318, -0.005859035532921553, 0.1964719146490097, -0.20022939145565033, 0.002971978159621358, 0.06040072813630104, -0.03314264863729477, 0.005108712241053581, -0.08902986347675323, -0.00867493823170662, -0.004925869405269623, -0.019366608932614326, -0.09311392903327942, -0.011381014250218868, 0.041074927896261215, 0.06859784573316574, 0.006505449302494526, -0.10515148937702179, 0.06774674355983734, 0.012170076370239258, -0.11106482148170471, 0.2341390997171402, -0.12488062679767609, -0.24140816926956177, -0.1364169716835022, -0.11627820134162903, -0.02352096699178219, 0.007738504093140364, 0.03380659595131874, -0.12468390166759491, -0.10114893317222595, -0.005312768276780844, 0.09899137914180756, -0.021439556032419205, 0.06560298800468445, 0.07922869175672531, 0.016251476481556892, -0.016711270436644554, -0.11899448931217194, 0.0012800154509022832, 0.02987886779010296, -0.03225863352417946, 0.03457020968198776, 0.03717375546693802, 0.03832841292023659, 0.18164239823818207, 0.0011001460952684283, 0.013860180042684078, -0.060141827911138535, 0.13536573946475983, -0.09117866307497025, -0.033692702651023865, 0.18924067914485931, 0.004530993290245533, -0.006806180812418461, 0.16801486909389496, 0.010683692060410976, -0.020589934661984444, 0.06956379860639572, -0.02292521670460701, -0.05310766398906708, -0.2622317969799042, -0.12151141464710236, -0.06087986379861832, 0.09460362046957016, -0.02003789320588112, 0.03493771329522133, 0.00724811339750886, 0.05523538962006569, 0.0005805635009892285, 0.08736110478639603, -0.014504854567348957, 0.12691815197467804, 0.21875479817390442, 0.051053281873464584, 0.10255467146635056, -0.014081349596381187, -0.04976934939622879, 0.058771368116140366, -0.037780340760946274, 0.1202879324555397, -0.00016915469313971698, 0.06999843567609787, 0.003367222845554352, 0.011633693240582943, 0.07228934019804001, 0.06731414049863815, -0.0053451042622327805, -0.0031083314679563046, 0.009852091781795025, -0.046518150717020035, 0.042030878365039825, -0.0008902492700144649, -0.11298903077840805, -0.012203674763441086, -0.05058182775974274, 0.045544885098934174, 0.061268724501132965, 0.013745833188295364, 0.029263805598020554, -0.2662377655506134, -0.06260348856449127, 0.057705409824848175, -0.0556221567094326, -0.09814812242984772, 0.028777867555618286, -0.018595637753605843, -0.12373137474060059, 0.07565607875585556, -0.01939912512898445, 0.0882057324051857, -0.09613793343305588, 0.060103099793195724, 0.03549594432115555, 0.06299600750207901, -0.009168640710413456, 0.11726731806993484, -0.3508847653865814, 0.15204760432243347, 0.041506823152303696, 0.054477326571941376, -0.02532699890434742, 0.03938627988100052, -0.00979435071349144, 0.05549491196870804, 0.03069419600069523, -0.01244976744055748, 0.09392252564430237, -0.02260330319404602, -0.13164184987545013, 0.0619686059653759, 0.00850240420550108, -0.04685873165726662, 0.04153880849480629, -0.011779462918639183, 0.006652978714555502, 0.0411265566945076, -0.05332190543413162, -0.11854492872953415, -0.11471585929393768, 0.060460224747657776, -0.018070081248879433, 0.13930760324001312, -0.09224177896976471, -0.11768464744091034, -0.03791256248950958, 0.14489321410655975, -0.12555716931819916, -0.10364583134651184, -0.13971592485904694, 0.02611556649208069, 0.03072546049952507, -0.06731561571359634, 0.06817848980426788, -0.01079985685646534, 0.10510410368442535, -0.031225282698869705, -0.11255575716495514, 0.11809481680393219, -0.07234155386686325, -0.17128513753414154, -0.05305328592658043, 0.09817489236593246, 0.09900037199258804, 0.05998390167951584, -0.012447042390704155, -0.007786608766764402, -0.07590685784816742, -0.11386482417583466, -0.017516981810331345, 0.02031324990093708, 0.08399046212434769, 0.02325223758816719, 0.014084478840231895, -0.11314737051725388, -0.03391752392053604, -0.046078868210315704, 0.09558660537004471, 0.16048714518547058, -0.03182348981499672, 0.10258972644805908, 0.09620834141969681, -0.05726669356226921, -0.24717609584331512, -0.030324436724185944, 0.0295749232172966, 0.03720661252737045, -0.05297970026731491, -0.2669736444950104, 0.15798257291316986, -0.02853807434439659, -0.06031402572989464, -0.037453893572092056, -0.09487777203321457, -0.1345556676387787, 0.2240075320005417, 0.031097518280148506, 0.11892060190439224, -0.09045898914337158, -0.023080013692378998, -0.09607894718647003, -0.1984201818704605, 0.2117990404367447, -0.1315169483423233, 0.052397470921278, -0.013804400339722633, 0.14574410021305084, -0.012816197238862514, -0.03487960994243622, 0.11853567510843277, 0.03246518597006798, 0.011910269036889076, -0.00784605648368597, -0.0932314321398735, 0.19864803552627563, 0.026288392022252083, 0.020393408834934235, -0.05307149142026901, 0.03889976441860199, -0.1057703047990799, -0.0490027591586113, -0.02212805300951004, 0.10566622763872147, -0.020202236250042915, -0.13997763395309448, -0.09286674112081528, 0.01427119318395853, -0.03163373842835426, -0.04442981258034706, 0.22235038876533508, 0.059187330305576324, 0.028338640928268433, 0.04155484959483147, 0.22462765872478485, -0.03823040425777435, 0.05619850009679794, 0.009814321994781494, -0.030578237026929855, 0.07752084732055664, -0.16307051479816437, 0.03855234012007713, 0.09883502125740051, 0.027668097987771034, 0.06176786869764328, 0.10448841750621796, -0.0541352853178978, -0.016106732189655304, 0.0909729078412056, -0.21212443709373474, -0.029640616849064827, -0.09848999232053757, 0.12207174301147461, -0.1349731981754303, 0.08500929921865463, 0.12295810133218765, -0.04856758937239647, -0.06118724122643471, 0.04890035092830658, 0.02527557499706745, -0.03536982461810112, 0.12679725885391235, 0.049250490963459015, 0.10318467020988464, -0.1389295756816864, 0.06077507883310318, 0.06829896569252014, 0.028287438675761223, -0.006023717578500509, 0.06657569110393524, -0.1394573450088501, -0.0140653345733881, 0.016747411340475082, 0.2207588255405426, -0.10048434138298035, -0.04844512417912483, -0.17779730260372162, -0.15690526366233826, 0.03847005218267441, 0.13138534128665924, 0.11090213805437088, 0.0801890641450882, -0.06780477613210678, 0.02144712023437023, -0.08566699922084808, 0.1054275706410408, 0.012750904075801373, 0.03808436542749405, -0.14902256429195404, 0.02604641579091549, -0.013293101452291012, 0.045832447707653046, -0.06564969569444656, -0.003814838593825698, -0.15275967121124268, 0.015445888973772526, -0.11211042106151581, -0.04727029800415039, -0.03584432229399681, -0.019962580874562263, -0.0009256203193217516, -0.0429253950715065, -0.016626087948679924, 0.03660362586379051, -0.05287287384271622, 0.05084196850657463, 0.021610818803310394, 0.062290534377098083, -0.1107647567987442, 0.05216392129659653, 0.05897929146885872, -0.06362687051296234, 0.12810875475406647, 0.08507736772298813, 0.05856352671980858, 0.09508923441171646, -0.16230902075767517, 0.009564101696014404, 0.0682159811258316, 0.04747899994254112, 0.05607827380299568, -0.11201977729797363, -0.0017094979993999004, 0.0017818064661696553, 0.023612720891833305, 0.03870072588324547, 0.08503368496894836, -0.09580910950899124, 0.04447377100586891, -0.04500007629394531, -0.015813980251550674, -0.0463973730802536, 0.032083507627248764, 0.13030989468097687, 0.051732566207647324, 0.12507563829421997, -0.05964336171746254, -0.003517686389386654, -0.08455781638622284, 0.028139661997556686, -0.025853119790554047, -0.0835794061422348, -0.11639972776174545, -0.0461922250688076, 0.05341260880231857, -0.0024157376028597355, 0.1942262351512909, 0.031890619546175, -0.027490926906466484, 0.055953387171030045, 0.12379128485918045, 0.0901838019490242, 0.04270406439900398, 0.14412593841552734, 0.09690841287374496, 0.02996865287423134, 0.06709416210651398, 0.013019229285418987, 0.13175322115421295, 0.10243181139230728, 0.1328229010105133, 0.04415362700819969, -0.028346829116344452, 0.046526387333869934, 0.017885081470012665, -0.08628885447978973, -0.16678540408611298, 0.01669476553797722, -0.06031446158885956, 0.055446721613407135, -0.047428566962480545, 0.20248499512672424, 0.11191053688526154, -0.09985119849443436, 0.04581838846206665, -0.026418685913085938, -0.095463827252388, -0.10370849817991257, -0.011096970178186893, -0.04692024365067482, -0.1783083826303482, 0.026008307933807373, -0.1472129076719284, -0.08213268965482712, 0.12006299197673798, 0.0016921099741011858, -0.03302696719765663, 0.08856895565986633, 0.15293359756469727, -0.028726935386657715, 0.04854114353656769, -0.008481660857796669, -0.015622107312083244, -0.025381406769156456, -0.002091354224830866, -0.059068962931632996, 0.06127307936549187, 0.0061377533711493015, -0.006166485138237476, -0.02392641454935074, 0.018360663205385208, -0.0026707975193858147, -0.06605048477649689, 0.011447272263467312, 0.0019490626873448491, -0.058779120445251465, 0.175041064620018, -0.010133858770132065, 0.047761596739292145, 0.04734912887215614, 0.26916196942329407, -0.06998245418071747, -0.11526279151439667, -0.10370219498872757, 0.14129991829395294, -0.00332409399561584, 0.042045239359140396, 0.04852437227964401, -0.08827220648527145, -0.09741572290658951, 0.19611547887325287, 0.12236897647380829, -0.006037252955138683, -0.011240548454225063, -0.014251595363020897, -0.012369678355753422, -0.003548760199919343, 0.03929559141397476, 0.08494186401367188, 0.16380970180034637, -0.06936924904584885, -0.06528738141059875, -0.043070897459983826, -0.044569022953510284, -0.06257715076208115, 0.054552026093006134, 0.04282711073756218, -0.052019067108631134, -0.02876475639641285, 0.04749377816915512, -0.1446523666381836, -0.13414880633354187, -0.11696334183216095, -0.10161983966827393, -0.09325607120990753, -0.057327013462781906, 0.12026646733283997, -0.0037156282924115658, 0.017357226461172104, -0.008466080762445927, 0.03925887867808342, 0.04559251293540001, 0.03971650078892708, -0.13094443082809448, -0.009788460098206997, 0.13826602697372437, -0.150150865316391, 0.13874508440494537, -0.054226383566856384, 0.05673127621412277, 0.11212045699357986, 0.04524108022451401, -0.07298169285058975, 0.014050968922674656, 0.023356754332780838, -0.1328241527080536, 0.01766057126224041, 0.09075227379798889, -0.008683155290782452, 0.0955127403140068, 0.0005409382283687592, -0.0014744148356840014, 0.00040784821612760425, -0.0260756928473711, -0.09847380220890045, -0.05831585451960564, -0.036396197974681854, -0.08901610970497131, 0.059172604233026505, 0.08723229914903641, -0.0604405403137207, 0.003340199589729309, -0.08054845035076141, 0.04561581835150719, -0.004048185423016548, -0.016746636480093002, -0.01396082155406475, -0.16696015000343323, -0.05640014261007309, 0.17705018818378448, 0.01084023155272007, -0.1837514489889145, 0.017866874113678932, -0.05859775096178055, 0.009888995438814163, -0.1380472034215927, 0.07401086390018463, 0.08400953561067581, -0.023333556950092316, -0.03126616030931473, -0.0006515762070193887, -0.028331412002444267, 0.08611869066953659, -0.17412568628787994, -0.09856399893760681 ]
null
null
transformers
# BERT-base-cased for QA **Language model:** bert-base-cased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 2 base_LM_model = "bert-base-cased" max_seq_len = 386 learning_rate = 5e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 81.30, "f1": 88.58, ``` Evaluated on the battery device dataset. ``` "precision": 67.02, "recall": 80.15, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/bert-base-cased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/bert-base-cased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BERT-base-cased for QA Language model: bert-base-cased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BERT-base-cased for QA \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BERT-base-cased for QA \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 66, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BERT-base-cased for QA \r\nLanguage model: bert-base-cased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.08264510333538055, 0.03704140707850456, -0.004181412979960442, 0.05439529940485954, 0.225198432803154, 0.030874891206622124, 0.11174658685922623, 0.04101616516709328, -0.10514313727617264, -0.028222301974892616, 0.04187488183379173, -0.032096099108457565, 0.07252109050750732, 0.08074328303337097, 0.004610520321875811, -0.10788297653198242, 0.04851280897855759, -0.012536558322608471, 0.0020435017067939043, 0.13797545433044434, 0.1188671663403511, -0.10084633529186249, 0.11711350828409195, 0.012419131584465504, -0.0950072854757309, 0.06747550517320633, -0.009153635241091251, -0.08097530901432037, 0.10537535697221756, -0.06524581462144852, 0.1286197006702423, 0.05192287266254425, 0.023126589134335518, -0.16670072078704834, 0.03791371360421181, 0.029599908739328384, 0.03278590366244316, 0.043871693313121796, 0.01858864352107048, -0.017991865053772926, 0.03174784407019615, 0.03978900983929634, 0.03597259148955345, 0.022859027609229088, -0.05026363581418991, -0.2323901504278183, -0.06615763157606125, -0.05747433006763458, 0.08007681369781494, 0.12892645597457886, -0.02710161730647087, 0.08741968870162964, -0.04670620709657669, 0.07213979214429855, 0.1782005876302719, -0.1278722584247589, -0.002810306381434202, 0.09693312644958496, 0.05702396109700203, 0.0020626962650567293, -0.012272888794541359, -0.07343656569719315, 0.00470301229506731, 0.051134735345840454, 0.05808793008327484, -0.05545942112803459, -0.14026570320129395, 0.11226722598075867, -0.09864280372858047, -0.007510367780923843, 0.207663431763649, -0.03880434110760689, -0.07910507172346115, 0.01660262979567051, -0.11700446158647537, -0.041877854615449905, -0.03070971369743347, -0.0255969800055027, -0.02188108302652836, -0.00645042397081852, -0.08095409721136093, 0.024563072249293327, -0.019210798665881157, -0.11709211766719818, -0.1248551532626152, 0.09410329908132553, 0.05942673608660698, 0.03456564620137215, -0.12108469009399414, 0.08517122268676758, -0.0009187859250232577, -0.15173041820526123, -0.03228890523314476, -0.058368440717458725, -0.07540114969015121, 0.011707553640007973, 0.008727168664336205, 0.011457383632659912, 0.11629577726125717, 0.08652487397193909, -0.01849316619336605, 0.012764779850840569, 0.004827816970646381, -0.06736411899328232, -0.004027221817523241, 0.1597355604171753, -0.039593931287527084, -0.1304716020822525, -0.003741430351510644, 0.027871785685420036, -0.01378125324845314, -0.015298290178179741, -0.07361383736133575, -0.006512124557048082, 0.07013451308012009, 0.0553625151515007, 0.0069966986775398254, 0.044691745191812515, -0.05322401225566864, -0.035431668162345886, 0.05657782033085823, -0.10430625826120377, -0.051820144057273865, -0.020617669448256493, 0.01602749153971672, -0.0022782988380640745, -0.04304155707359314, 0.024117181077599525, -0.031122762709856033, -0.022889677435159683, -0.08202962577342987, -0.10001585632562637, -0.12374454736709595, -0.09496018290519714, -0.03553048521280289, 0.11234934628009796, 0.058209843933582306, -0.1289951503276825, -0.32596367597579956, 0.09601391106843948, 0.001474618911743164, 0.008034530095756054, -0.04438721761107445, -0.025622155517339706, -0.04860571026802063, -0.03570782393217087, -0.04506609961390495, 0.13019242882728577, -0.06005194038152695, 0.07820682972669601, 0.16193529963493347, 0.021072667092084885, -0.05744168907403946, 0.05013955011963844, -0.13443082571029663, 0.005910898558795452, -0.1522812694311142, 0.04356621578335762, -0.11853786557912827, -0.013682028278708458, -0.11155779659748077, -0.07448755204677582, 0.11634550243616104, -0.00767933763563633, 0.08571280539035797, 0.14312990009784698, -0.097368024289608, -0.11737805604934692, 0.025623813271522522, -0.13394665718078613, -0.15504376590251923, 0.04876808822154999, -0.01808425970375538, 0.02043926902115345, 0.028248542919754982, 0.054227568209171295, 0.12920688092708588, -0.1650054156780243, -0.1665174961090088, -0.02311820350587368, 0.05257938802242279, -0.031164057552814484, 0.09500563889741898, 0.04737786203622818, -0.08148904889822006, 0.07698929309844971, -0.08240782469511032, 0.05429650470614433, -0.053730037063360214, -0.10009192675352097, -0.04983168840408325, -0.038892712444067, 0.12404855340719223, 0.0027309099677950144, -0.04355708882212639, 0.02238008938729763, -0.1000918596982956, 0.12050345540046692, 0.10768888890743256, -0.04981318116188049, -0.0005225501372478902, -0.1954769492149353, 0.0970403179526329, -0.04763329401612282, 0.021455122157931328, -0.12063133716583252, -0.15726128220558167, 0.08440874516963959, -0.02395663969218731, -0.07021045684814453, -0.03142626956105232, 0.018809499219059944, 0.026769908145070076, -0.1053374782204628, -0.024707544595003128, -0.14512255787849426, -0.01899678073823452, -0.0009867933113127947, -0.13153687119483948, 0.026653647422790527, -0.03206615522503853, 0.0809212252497673, 0.018881995230913162, 0.021702365949749947, 0.1262427568435669, 0.03161777928471565, -0.019195299595594406, -0.016997048631310463, -0.030292754992842674, 0.06583352386951447, -0.03139771521091461, -0.030050834640860558, -0.029553133994340897, -0.014639854431152344, -0.010582073591649532, -0.022141698747873306, -0.09766723960638046, 0.10225243866443634, 0.13953223824501038, 0.11828956753015518, 0.017308756709098816, -0.0869927927851677, -0.006957856938242912, -0.04002828150987625, -0.0837053582072258, -0.04516978934407234, 0.2248532921075821, 0.005932528525590897, 0.034028660506010056, -0.037873525172472, 0.013650812208652496, 0.00028297555400058627, 0.04800755903124809, 0.06605807691812515, 0.08225879073143005, 0.10294875502586365, 0.06823728233575821, 0.05948231369256973, 0.07352222502231598, -0.06061975657939911, 0.11687813699245453, -0.09365668147802353, -0.04090730473399162, -0.03242642804980278, -0.10646439343690872, -0.026679201051592827, 0.2661118507385254, -0.1445089727640152, 0.06491842865943909, 0.07100919634103775, -0.007063162978738546, 0.027086663991212845, -0.03360452130436897, -0.005843873601406813, -0.014240504242479801, -0.05693092942237854, -0.1704614758491516, 0.02126229926943779, 0.0568777397274971, 0.011838347651064396, -0.013561690226197243, -0.054921671748161316, 0.024051029235124588, 0.010956333950161934, -0.06817802786827087, 0.22689305245876312, -0.10229086130857468, -0.197395458817482, -0.11371965706348419, -0.11130207031965256, -0.03189079463481903, -0.0677868127822876, 0.026566486805677414, -0.12817184627056122, -0.09719689190387726, 0.017178013920783997, 0.18523889780044556, 0.024162881076335907, 0.07082296162843704, 0.03326484560966492, -0.005783569999039173, 0.013914594426751137, -0.11820117384195328, 0.01603982038795948, 0.04998500645160675, -0.10007922351360321, 0.027710426598787308, 0.049387525767087936, 0.08372186869382858, 0.0917244479060173, 0.005189064424484968, 0.015850521624088287, -0.05959349870681763, 0.21924936771392822, -0.0956045612692833, -0.037377215921878815, 0.2877155840396881, 0.10402745753526688, -0.0061632986180484295, 0.19504158198833466, -0.02953783981502056, -0.014077321626245975, 0.09458805620670319, -0.011998903006315231, -0.05070074647665024, -0.2431919276714325, -0.08050011843442917, -0.0291195809841156, 0.06444893777370453, -0.06758401542901993, 0.009725652635097504, -0.07366400212049484, 0.09517527371644974, 0.024621810764074326, -0.007426127325743437, -0.10076142102479935, 0.0878775343298912, 0.2372196763753891, 0.050240859389305115, 0.1127861961722374, -0.00874766893684864, 0.013797118328511715, 0.0647849589586258, 0.0637873038649559, 0.16863292455673218, -0.060834918171167374, 0.04585954546928406, 0.033358894288539886, 0.1585313230752945, 0.0337194986641407, 0.0508292019367218, -0.012091447599232197, 0.007119064684957266, 0.009284892119467258, -0.040888238698244095, 0.005513138137757778, -0.02786017395555973, -0.044975053519010544, 0.01041109673678875, -0.018814275041222572, 0.1261156052350998, 0.03416561707854271, 0.011934725567698479, 0.015326553955674171, -0.20962375402450562, -0.038605622947216034, 0.03806518018245697, -0.04675734415650368, -0.0989166870713234, 0.034315887838602066, 0.02432303875684738, -0.1208200603723526, 0.023937130346894264, -0.04036884382367134, 0.07437296956777573, -0.053038340061903, 0.059290193021297455, 0.06511423736810684, 0.11074800789356232, 0.01813720539212227, 0.11920711398124695, -0.39334872364997864, 0.11125018447637558, 0.04152580350637436, 0.07549130171537399, 0.03087960183620453, 0.060740675777196884, -0.05508652329444885, -0.04546588659286499, 0.052319712936878204, -0.013515375554561615, 0.10857013612985611, 0.014488381333649158, -0.15604496002197266, 0.15646620094776154, 0.0017422848613932729, -0.015124700963497162, 0.07917404174804688, -0.021416617557406425, 0.037233419716358185, -0.008624535985291004, 0.010562590323388577, -0.12917813658714294, -0.08375706523656845, 0.05281930789351463, -0.05813979730010033, 0.12338217347860336, -0.1514633148908615, -0.09302239865064621, 0.04349169135093689, 0.06475388258695602, -0.06814029812812805, -0.10990787297487259, -0.13902685046195984, 0.013392620719969273, 0.12250807136297226, -0.10501199215650558, 0.076146699488163, 0.008427119813859463, 0.06528404355049133, -0.01353498362004757, -0.10305662453174591, 0.05815249681472778, -0.07873865962028503, -0.1373354196548462, -0.02512708492577076, 0.0361248217523098, 0.08938776701688766, 0.052151158452034, 0.0016809969674795866, -0.017129166051745415, -0.1312790811061859, -0.11199172586202621, -0.05208436772227287, -0.05995055288076401, 0.048329174518585205, 0.016616888344287872, 0.11480037868022919, -0.03383709117770195, -0.029304342344403267, -0.018477847799658775, 0.009733067825436592, 0.1402968019247055, -0.02008664794266224, 0.056160323321819305, 0.12680524587631226, -0.012431671842932701, -0.21311867237091064, -0.06322373449802399, 0.08409848809242249, 0.0572953037917614, -0.028759069740772247, -0.25707659125328064, 0.15294066071510315, -0.019944148138165474, -0.0662572979927063, -0.10826724022626877, -0.10833268612623215, -0.11851160228252411, 0.2003939300775528, 0.07903246581554413, 0.05614453926682472, -0.06853899359703064, -0.023707721382379532, -0.051398035138845444, -0.27641838788986206, 0.12603560090065002, -0.14649611711502075, 0.02885783649981022, -0.018903369084000587, 0.22490772604942322, -0.026705943048000336, -0.030805127695202827, 0.13456599414348602, 0.05290687829256058, -0.023630572482943535, 0.03409254550933838, -0.013255451805889606, 0.11935871094465256, 0.028466401621699333, 0.04127839207649231, -0.0026503021363168955, 0.11818111687898636, -0.12029270827770233, -0.03855316713452339, -0.05593486875295639, 0.08031582832336426, -0.03716820478439331, -0.15496093034744263, -0.09570344537496567, 0.03497886285185814, -0.05072854831814766, -0.06166747212409973, 0.09299386292695999, 0.03900180011987686, -0.008335758931934834, 0.09980493783950806, 0.16206836700439453, 0.014273724518716335, -0.06929907947778702, 0.0361921489238739, -0.020316071808338165, 0.06929051876068115, -0.13831733167171478, 0.04184170812368393, 0.10914041846990585, 0.01714314892888069, 0.015423783101141453, 0.048145558685064316, -0.08114447444677353, 0.007460224907845259, 0.06975824385881424, -0.15072423219680786, -0.15014174580574036, -0.07191908359527588, 0.1694345474243164, -0.21570181846618652, 0.07405035942792892, 0.12254441529512405, -0.012910152785480022, -0.06929901987314224, 0.029264697805047035, 0.026358824223279953, -0.013433213345706463, 0.17636388540267944, 0.0840645283460617, 0.13589482009410858, -0.13035619258880615, 0.0550864115357399, 0.02846807800233364, 0.02466089278459549, -0.036165256053209305, -0.007852008566260338, -0.10285639762878418, 0.017118122428655624, 0.004650963004678488, 0.13401807844638824, -0.1156250536441803, -0.076687291264534, -0.16149191558361053, -0.13162446022033691, 0.007788771763443947, 0.07636693865060806, 0.08892370760440826, 0.07903323322534561, -0.045226965099573135, 0.025276711210608482, -0.07026023417711258, 0.11076758056879044, 0.004519676323980093, 0.033850137144327164, -0.11504705250263214, -0.07750300318002701, -0.031180402263998985, 0.12049885094165802, -0.05798372998833656, 0.0016156277852132916, -0.08589562773704529, 0.023874221369624138, -0.1799352765083313, -0.07206067442893982, -0.01787942461669445, -0.02152206562459469, 0.015561184845864773, -0.07890982925891876, -0.040547966957092285, 0.062131959944963455, -0.04688093811273575, 0.056656740605831146, 0.045474667102098465, 0.022108817473053932, -0.12908123433589935, 0.09613552689552307, 0.07868979871273041, -0.05395370349287987, 0.12233548611402512, 0.06563162058591843, 0.0654035285115242, 0.11674269288778305, -0.05963025614619255, 0.01673266664147377, 0.0017987785395234823, 0.07924831658601761, 0.07990594208240509, -0.13097216188907623, 0.0028416900895535946, -0.008042478002607822, 0.013911728747189045, 0.018248876556754112, 0.08203721046447754, -0.09912516176700592, 0.05738631263375282, -0.05848013237118721, 0.033675067126750946, -0.05011749267578125, 0.042718589305877686, 0.09605646133422852, 0.10041162371635437, 0.10048825293779373, -0.033916693180799484, -0.015672653913497925, -0.12042904645204544, 0.014551914297044277, -0.0307158213108778, -0.025700127705931664, -0.05691840127110481, -0.04629223421216011, 0.06438212841749191, 0.027030587196350098, 0.14526459574699402, -0.02092713490128517, 0.048152998089790344, 0.010343872010707855, 0.10381343215703964, 0.11289958655834198, -0.009817883372306824, 0.11049077659845352, 0.09628784656524658, 0.08286537975072861, 0.11593057960271835, -0.014210921712219715, 0.054201092571020126, 0.12084861844778061, 0.10733503848314285, 0.0738319605588913, 0.03686815872788429, 0.04776214808225632, 0.04681498557329178, -0.09954873472452164, -0.07877770066261292, -0.011385519988834858, -0.07222345471382141, 0.019375724717974663, -0.007217030040919781, 0.20621280372142792, 0.07270181179046631, -0.11559164524078369, 0.07764100283384323, -0.0006228848360478878, -0.10020511597394943, -0.09277047216892242, 0.10321077704429626, -0.045608051121234894, -0.21274161338806152, 0.028441738337278366, -0.15989038348197937, -0.12231793254613876, 0.19450049102306366, -0.0003877142444252968, -0.004617210011929274, 0.14881418645381927, 0.18591295182704926, -0.03601889684796333, -0.030477559193968773, -0.014439697377383709, 0.007220607716590166, 0.05716617405414581, 0.02513355389237404, -0.02252647839486599, 0.11320056021213531, 0.04778872802853584, 0.010157971642911434, -0.06984946131706238, 0.021167486906051636, -0.004289973992854357, -0.03471194580197334, -0.021302631124854088, 0.0064718411304056644, -0.0654880702495575, 0.19193197786808014, -0.011203180998563766, 0.08662601560354233, 0.011808705516159534, 0.3084222674369812, -0.05184539034962654, -0.1343214213848114, -0.13399474322795868, 0.07504415512084961, -0.04805907607078552, 0.05634341016411781, 0.05760745704174042, -0.0781264454126358, -0.12682180106639862, 0.1681845635175705, 0.028424225747585297, -0.02700587548315525, -0.0516635924577713, 0.04437751695513725, -0.01878647692501545, -0.05949920043349266, 0.05186179652810097, 0.0903342142701149, 0.19463075697422028, -0.062081217765808105, -0.0284646637737751, -0.04745526611804962, -0.07286201417446136, 0.0019941481295973063, 0.02952072210609913, 0.013780813664197922, -0.03207683935761452, -0.026353176683187485, 0.03883400559425354, -0.08937108516693115, -0.12409764528274536, -0.12325781583786011, -0.0352654829621315, -0.11697028577327728, -0.0667644813656807, 0.19246481359004974, 0.003422507317736745, 0.0039129736833274364, -0.042448997497558594, 0.08795791119337082, 0.0606444850564003, 0.03862352296710014, -0.08464600890874863, -0.06597180664539337, 0.20097419619560242, -0.13562895357608795, 0.1475033015012741, -0.045620404183864594, 0.07677597552537918, 0.10883981734514236, -0.02155524492263794, -0.11677210032939911, 0.06074211746454239, 0.05197424441576004, -0.17881695926189423, -0.006089129485189915, 0.11664463579654694, 0.01963537186384201, 0.11132092773914337, -0.01162458211183548, 0.07373013347387314, -0.021857012063264847, -0.0087065314874053, -0.04471992328763008, -0.05482380837202072, -0.030706724151968956, -0.08315915614366531, 0.044345781207084656, 0.07799014449119568, -0.059506434947252274, 0.06625138223171234, -0.08832447230815887, 0.09995781630277634, -0.029881194233894348, 0.07099077850580215, 0.027114544063806534, -0.12751056253910065, -0.06613272428512573, 0.14762921631336212, -0.07336082309484482, -0.13612410426139832, 0.01949276030063629, -0.009479361586272717, 0.010775535367429256, -0.08243679255247116, 0.13128776848316193, 0.06292546540498734, 0.0008033000631257892, -0.020975761115550995, 0.03064114786684513, -0.0615868866443634, 0.06834892928600311, -0.17663931846618652, -0.09677818417549133 ]
null
null
transformers
# BERT-base-uncased for Battery Abstract Classification **Language model:** bert-base-uncased **Language:** English **Downstream-task:** Text Classification **Training data:** training\_data.csv **Eval data:** val\_data.csv **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 13 base_LM_model = "bert-base-uncased" learning_rate = 2e-5 ``` ## Performance ``` "Validation accuracy": 96.79, "Test accuracy": 96.29, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline model_name = "batterydata/bert-base-uncased-abstract" # a) Get predictions nlp = pipeline('text-classification', model=model_name, tokenizer=model_name) input = {'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.'} res = nlp(input) # b) Load model & tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "Text Classification", "datasets": ["batterydata/paper-abstracts"], "metrics": "glue"}
text-classification
batterydata/bert-base-uncased-abstract
[ "transformers", "pytorch", "bert", "text-classification", "Text Classification", "en", "dataset:batterydata/paper-abstracts", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# BERT-base-uncased for Battery Abstract Classification Language model: bert-base-uncased Language: English Downstream-task: Text Classification Training data: training\_data.csv Eval data: val\_data.csv Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BERT-base-uncased for Battery Abstract Classification \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# BERT-base-uncased for Battery Abstract Classification \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 64, 73, 5, 2, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #text-classification #Text Classification #en #dataset-batterydata/paper-abstracts #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BERT-base-uncased for Battery Abstract Classification \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Text Classification\r\nTraining data: training\\_data.csv\r\nEval data: val\\_data.csv\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.06228888779878616, 0.11250824481248856, -0.0041213142685592175, 0.055747926235198975, 0.23391050100326538, 0.023527516052126884, 0.1881936639547348, 0.052966222167015076, -0.08502889424562454, -0.03743405267596245, 0.0547482855618, 0.04717966541647911, 0.048833444714546204, 0.05789508298039436, -0.06720832735300064, -0.17673532664775848, 0.03977479785680771, 0.005881349090486765, 0.013472475111484528, 0.10886987298727036, 0.12927909195423126, -0.08997895568609238, 0.11126676201820374, -0.004677884746342897, -0.09576999396085739, 0.02236267551779747, 0.009153943508863449, -0.13596896827220917, 0.11670473217964172, -0.03327525779604912, 0.14926248788833618, 0.038034070283174515, 0.03554176911711693, -0.15698052942752838, 0.02956457808613777, 0.01662442646920681, -0.004571729805320501, 0.08249885588884354, 0.030352018773555756, -0.0744418054819107, 0.07244908064603806, -0.025779983028769493, 0.06636882573366165, 0.03307770937681198, -0.09769592434167862, -0.1396782398223877, -0.03695763647556305, 0.005818197503685951, 0.08002575486898422, 0.10374927520751953, -0.015237929299473763, 0.06490843743085861, -0.05599529668688774, 0.07525669038295746, 0.18604770302772522, -0.147354856133461, -0.0024797506630420685, 0.09039535373449326, 0.009100791066884995, -0.007367027923464775, -0.06322057545185089, -0.009582627564668655, 0.019129926338791847, 0.04189402610063553, 0.06991736590862274, -0.032773882150650024, -0.17588426172733307, 0.08690977841615677, -0.09434165060520172, 0.0014400534564629197, 0.2399500459432602, -0.05715690180659294, -0.04002469405531883, -0.03641575574874878, -0.08460943400859833, -0.05869467556476593, -0.03839203342795372, 0.024105899035930634, -0.016455044969916344, 0.018934935331344604, -0.0037622088566422462, 0.028402777388691902, -0.04753527417778969, -0.06754370033740997, -0.10303446650505066, 0.10163094103336334, 0.012007116340100765, 0.0005543057341128588, -0.09265324473381042, 0.12462905049324036, 0.06640659272670746, -0.16450674831867218, 0.010185003280639648, -0.06238999962806702, -0.01806103065609932, -0.019939366728067398, 0.00809696689248085, -0.03028877079486847, 0.07979574054479599, 0.07876035571098328, -0.007167816627770662, 0.008735189214348793, -0.02753928117454052, -0.006618595216423273, -0.008782234974205494, 0.14299771189689636, -0.037063661962747574, -0.034234482795000076, 0.04005654528737068, 0.0011127671459689736, 0.029565850272774696, -0.04212339222431183, -0.13260436058044434, -0.00572064658626914, 0.0857323408126831, 0.06245562806725502, -0.03997192531824112, 0.0726185142993927, -0.08198544383049011, -0.06115931272506714, 0.08530742675065994, -0.09867432713508606, -0.011558505706489086, 0.024360543116927147, 0.016689972952008247, 0.0430651493370533, -0.0022710252087563276, 0.021265733987092972, -0.0394321046769619, 0.06651619076728821, -0.08386287093162537, -0.045655038207769394, -0.08960557729005814, -0.11361365765333176, 0.007351818960160017, 0.026584841310977936, 0.05096801742911339, -0.12581250071525574, -0.2691899240016937, 0.06312926858663559, 0.011732909828424454, -0.016300147399306297, -0.05299889296293259, -0.07840854674577713, -0.07482203841209412, 0.02836814895272255, -0.010807742364704609, 0.04306318610906601, -0.069627545773983, 0.0521429181098938, 0.07528194785118103, 0.03375735878944397, -0.0853121429681778, 0.05204472318291664, -0.11592930555343628, -0.012930687516927719, -0.19291748106479645, 0.04418192431330681, -0.10102511942386627, 0.08357453346252441, -0.06841110438108444, -0.04582708328962326, 0.12154911458492279, -0.01009679026901722, 0.02075367048382759, 0.1602809727191925, -0.11002321541309357, -0.1304006725549698, 0.05373359099030495, -0.10967063903808594, -0.0830642506480217, 0.06713944673538208, -0.04566376656293869, 0.017960872501134872, 0.08625209331512451, 0.08013303577899933, 0.046303994953632355, -0.0889616385102272, -0.09038616716861725, 0.021011577919125557, 0.037260718643665314, -0.005608543753623962, 0.10151350498199463, 0.03749207779765129, -0.052251677960157394, 0.0616006925702095, -0.039949674159288406, 0.05374772846698761, -0.03412657976150513, -0.08189122378826141, -0.013136292807757854, -0.03299499303102493, 0.13165296614170074, 0.018369687721133232, 0.03365183249115944, -0.011031944304704666, -0.09272517263889313, 0.1407436579465866, 0.07582927495241165, -0.056475307792425156, 0.014519610442221165, -0.12732979655265808, 0.07388757169246674, 0.026346001774072647, 0.02801337093114853, -0.12155815958976746, -0.070127934217453, 0.0436030849814415, 0.054260559380054474, -0.02307087555527687, -0.05253688991069794, -0.0001333885738858953, 0.035661131143569946, -0.09666076302528381, -0.007937748916447163, -0.05093143880367279, 0.0023517608642578125, -0.021788215264678, -0.2124955803155899, 0.0683596283197403, -0.045037344098091125, 0.13901160657405853, -0.12964563071727753, 0.060368310660123825, 0.13233090937137604, 0.0693414956331253, -0.008418215438723564, 0.0008288201643154025, 0.012555640190839767, 0.03859550878405571, -0.06960843503475189, -0.04283005744218826, 0.01834079623222351, 0.001701622735708952, -0.07897995412349701, 0.01522780116647482, -0.1207239180803299, 0.1105019599199295, 0.16464346647262573, -0.030538922175765038, -0.029077308252453804, -0.022898221388459206, 0.010090803727507591, -0.0126528674736619, -0.08034709095954895, -0.01187194511294365, 0.24448628723621368, -0.015188626945018768, 0.10868040472269058, -0.055525317788124084, -0.01141341682523489, -0.008199379779398441, 0.031025294214487076, 0.03428865969181061, 0.1318824738264084, 0.06641414016485214, -0.05921947956085205, 0.10345041006803513, 0.07104869931936264, -0.07144234329462051, 0.15426263213157654, -0.05510329455137253, -0.015618303790688515, 0.01719161868095398, -0.09022101759910583, -0.002254216931760311, 0.17269645631313324, -0.20520657300949097, 0.006758788134902716, 0.057648491114377975, -0.0234715286642313, 0.005774593912065029, -0.0784234032034874, -0.015824299305677414, -0.010016575455665588, -0.030012283474206924, -0.10688035935163498, -0.0017264644848182797, 0.04081502556800842, 0.07457762956619263, -0.00030436323140747845, -0.0853741243481636, 0.07028856873512268, 0.0018902960000559688, -0.10389673709869385, 0.23301544785499573, -0.13040190935134888, -0.24266749620437622, -0.1266438513994217, -0.10843027383089066, -0.04020921513438225, 0.006887568160891533, 0.039995767176151276, -0.12374518066644669, -0.09478618949651718, -0.023335497826337814, 0.08406578004360199, -0.01762387901544571, 0.0523306205868721, 0.07520759105682373, 0.011866090819239616, -0.01172390766441822, -0.11967357993125916, 0.003937906585633755, 0.026686077937483788, -0.04462660476565361, 0.04457720369100571, 0.02519657462835312, 0.04253540188074112, 0.1751375049352646, 0.004562335088849068, 0.013070221990346909, -0.052163366228342056, 0.13659681379795074, -0.08895877748727798, -0.03781469166278839, 0.179616779088974, 0.0034485759679228067, -0.0017073673661798239, 0.1624249666929245, 0.015338130295276642, -0.03389986604452133, 0.06649515777826309, -0.02049100026488304, -0.04916410148143768, -0.2813330888748169, -0.12108982354402542, -0.06233124062418938, 0.09197944402694702, -0.01991790533065796, 0.037169199436903, 0.009163430891931057, 0.047301359474658966, -0.005186950787901878, 0.08668655902147293, -0.012400493957102299, 0.11961372941732407, 0.2197643518447876, 0.05278347432613373, 0.09332844614982605, -0.02258853241801262, -0.03871815651655197, 0.06663646548986435, -0.040794458240270615, 0.12171178311109543, -0.002471982967108488, 0.0992620587348938, 0.010437468066811562, 0.004387148655951023, 0.06643186509609222, 0.06010770797729492, -0.003384131472557783, 0.0026897990610450506, 0.003109181299805641, -0.048515159636735916, 0.03801197558641434, -0.01021746639162302, -0.1052551418542862, -0.0032263733446598053, -0.05430279299616814, 0.037340108305215836, 0.07027610391378403, 0.010783469304442406, 0.04391885921359062, -0.23970302939414978, -0.05463477596640587, 0.06889119744300842, -0.051734473556280136, -0.10130787640810013, 0.02973628044128418, -0.008811239153146744, -0.11641179025173187, 0.08422783762216568, -0.01234609354287386, 0.0929177775979042, -0.1116175577044487, 0.060176391154527664, 0.01636788807809353, 0.05280478298664093, -0.014267712831497192, 0.1161787137389183, -0.35793936252593994, 0.149399071931839, 0.04763982072472572, 0.05648081749677658, -0.029696129262447357, 0.029487866908311844, 0.0016153118340298533, 0.060790788382291794, 0.04290509968996048, -0.00976119376718998, 0.09452445805072784, -0.0013249964686110616, -0.1331881880760193, 0.059303637593984604, 0.010138582438230515, -0.04012376815080643, 0.03115135058760643, -0.002486905548721552, 0.01627364195883274, 0.031769413501024246, -0.05510767549276352, -0.12983033061027527, -0.12057367712259293, 0.06791998445987701, -0.010283064097166061, 0.12029772251844406, -0.08255265653133392, -0.10603200644254684, -0.03724765032529831, 0.13458949327468872, -0.14089259505271912, -0.10212529450654984, -0.1326088011264801, 0.013342554681003094, 0.03331772983074188, -0.061750445514917374, 0.05091221258044243, -0.012987111695110798, 0.08812364190816879, -0.03219856321811676, -0.11108610779047012, 0.11034473776817322, -0.06079607084393501, -0.17588432133197784, -0.055902838706970215, 0.08894289284944534, 0.0978093221783638, 0.0593397282063961, -0.012198335491120815, -0.007176933344453573, -0.07938048988580704, -0.11553233116865158, -0.009355149231851101, 0.0313604474067688, 0.08027449995279312, 0.0557892806828022, 0.011438240297138691, -0.11713199317455292, -0.037496164441108704, -0.05339571461081505, 0.10449650883674622, 0.1763160228729248, -0.027459384873509407, 0.11444468796253204, 0.10882942378520966, -0.06977000832557678, -0.24971044063568115, -0.02823658287525177, 0.010695458389818668, 0.027239210903644562, -0.05331207439303398, -0.26733893156051636, 0.13956277072429657, -0.02047087997198105, -0.04930054396390915, -0.021953269839286804, -0.11889541149139404, -0.13173888623714447, 0.2212357521057129, 0.024212270975112915, 0.13552558422088623, -0.10141651332378387, -0.03552073612809181, -0.10697407275438309, -0.19261685013771057, 0.2180328667163849, -0.15384605526924133, 0.06694202125072479, -0.022598857060074806, 0.13745516538619995, -0.010338970459997654, -0.029232408851385117, 0.12530283629894257, 0.03932745009660721, 0.014946965500712395, -0.018840784206986427, -0.08670569211244583, 0.19614562392234802, 0.017628531903028488, 0.018216675147414207, -0.06766051054000854, 0.04270758852362633, -0.11780056357383728, -0.04229438304901123, -0.02205110713839531, 0.1051582619547844, -0.027293989434838295, -0.1250103861093521, -0.08885304629802704, 0.012824089266359806, -0.01419831346720457, -0.03439546003937721, 0.22151227295398712, 0.04574630409479141, 0.03434452414512634, 0.04784340038895607, 0.22979162633419037, -0.03128241002559662, 0.08058498054742813, 0.007299646735191345, -0.026318589225411415, 0.07120350748300552, -0.1761905699968338, 0.034665316343307495, 0.09920486062765121, 0.029843205586075783, 0.0626610741019249, 0.09645218402147293, -0.05030339956283569, -0.011621414683759212, 0.08064531534910202, -0.212094247341156, -0.02615799754858017, -0.08655808866024017, 0.1185525432229042, -0.12495137751102448, 0.08410985767841339, 0.12858489155769348, -0.039302099496126175, -0.05127720534801483, 0.05222406983375549, 0.029574519023299217, -0.0292669590562582, 0.12476669251918793, 0.044244375079870224, 0.09409109503030777, -0.14231261610984802, 0.08276525884866714, 0.07678377628326416, 0.02592042274773121, -0.007042738143354654, 0.07621146738529205, -0.1411515772342682, -0.027206961065530777, -0.0012386080343276262, 0.22684817016124725, -0.08830789476633072, -0.03999834507703781, -0.16621732711791992, -0.14831899106502533, 0.03712709993124008, 0.14482936263084412, 0.1067073792219162, 0.07942549139261246, -0.0742853507399559, 0.0051328581757843494, -0.07737234979867935, 0.09609048813581467, 0.010544170625507832, 0.029065735638141632, -0.13759946823120117, 0.015994062647223473, -0.015675708651542664, 0.04371277615427971, -0.06048965826630592, -0.005258271470665932, -0.15100052952766418, 0.014544866979122162, -0.11811091005802155, -0.0356510654091835, -0.02864055335521698, -0.014527828432619572, -0.012695887126028538, -0.03848934546113014, -0.016430525109171867, 0.019803613424301147, -0.05113907903432846, 0.04328387603163719, 0.02420561946928501, 0.06875330209732056, -0.11303975433111191, 0.04234533756971359, 0.052939895540475845, -0.06048055738210678, 0.1368902176618576, 0.0813215896487236, 0.04081137478351593, 0.09119941294193268, -0.18154922127723694, 0.004097743891179562, 0.05969056487083435, 0.041842736303806305, 0.04358239471912384, -0.09459254890680313, -0.010150075890123844, -0.00025700643891468644, 0.027939578518271446, 0.051267869770526886, 0.08549731969833374, -0.08815053850412369, 0.050645627081394196, -0.04855017736554146, -0.01391957700252533, -0.05111800134181976, 0.048990726470947266, 0.13715073466300964, 0.050480443984270096, 0.1273748129606247, -0.06120991334319115, -0.0005339878844097257, -0.07811082154512405, 0.024677470326423645, -0.01453984435647726, -0.09648581594228745, -0.11933444440364838, -0.04091356694698334, 0.05592930316925049, -0.007429986260831356, 0.2007342129945755, 0.015266919508576393, -0.0396571159362793, 0.050233740359544754, 0.12667769193649292, 0.055058520287275314, 0.034056954085826874, 0.15015465021133423, 0.09967391937971115, 0.02233796752989292, 0.06231576204299927, 0.0007000266341492534, 0.1228850707411766, 0.11068567633628845, 0.14208729565143585, 0.0615694634616375, -0.022159460932016373, 0.05310230702161789, 0.012079006060957909, -0.078270822763443, -0.1433866173028946, 0.042635053396224976, -0.043436191976070404, 0.06886336207389832, -0.05062376707792282, 0.18480359017848969, 0.09382547438144684, -0.09609347581863403, 0.043758027255535126, -0.02480035275220871, -0.10526397079229355, -0.10723698884248734, -0.01650266721844673, -0.04673634096980095, -0.15854859352111816, 0.02164088748395443, -0.14573310315608978, -0.05926952511072159, 0.12528851628303528, 0.0021575086284428835, -0.041160948574543, 0.08568399399518967, 0.1429181545972824, -0.024174345657229424, 0.07343293726444244, -0.015119243413209915, -0.024033205583691597, -0.026967398822307587, -0.0006402666913345456, -0.053341858088970184, 0.0724366083741188, 0.006936495658010244, 0.0003232832532376051, -0.01444681454449892, 0.012924161739647388, -0.019027572125196457, -0.07095050066709518, 0.010573884472250938, 0.01770639233291149, -0.06187279522418976, 0.18384554982185364, -0.0010223420104011893, 0.03237687423825264, 0.053502850234508514, 0.2646506130695343, -0.06471669673919678, -0.11146632581949234, -0.09377661347389221, 0.1279170960187912, -0.004802794195711613, 0.03768543154001236, 0.047409866005182266, -0.09217086434364319, -0.08618146926164627, 0.18549038469791412, 0.13589490950107574, -0.02560276724398136, -0.008441969752311707, -0.013166723772883415, -0.009787830524146557, 0.005225678905844688, 0.041465919464826584, 0.07764151692390442, 0.1380869597196579, -0.06399305164813995, -0.0519942045211792, -0.03400852903723717, -0.046500809490680695, -0.07479340583086014, 0.06387168169021606, 0.04251394793391228, -0.05046327784657478, -0.03274212032556534, 0.05654073879122734, -0.16317661106586456, -0.12017263472080231, -0.11249095946550369, -0.11457236111164093, -0.09636767208576202, -0.06200416758656502, 0.10689171403646469, -0.0003825396706815809, 0.007925822399556637, -0.01190898846834898, 0.03717846795916557, 0.041949767619371414, 0.02937210537493229, -0.12414459139108658, -0.008123173378407955, 0.12219151854515076, -0.16226962208747864, 0.13895805180072784, -0.044828854501247406, 0.04965692758560181, 0.10979652404785156, 0.03463933989405632, -0.06600768864154816, 0.022770829498767853, 0.02521081641316414, -0.11154990643262863, 0.02014343999326229, 0.09750209003686905, -0.013049328699707985, 0.09278073906898499, 0.016334282234311104, 0.006082457955926657, 0.0063730087131261826, -0.006831081118434668, -0.106648288667202, -0.06278988718986511, -0.033822715282440186, -0.0848069116473198, 0.06047157570719719, 0.09566563367843628, -0.05573442950844765, 0.0055627659894526005, -0.0727531760931015, 0.03632009029388428, -0.014883692376315594, -0.015006732195615768, -0.015535855665802956, -0.15088409185409546, -0.047347232699394226, 0.18015722930431366, 0.027753913775086403, -0.2088024616241455, 0.023445220664143562, -0.06906677782535553, 0.008249304257333279, -0.13349448144435883, 0.07916797697544098, 0.0768510028719902, -0.023315221071243286, -0.03050910495221615, -0.013833247125148773, -0.02604469656944275, 0.08922187983989716, -0.15779191255569458, -0.091957688331604 ]
null
null
transformers
# BERT-base-cased for QA **Language model:** bert-base-uncased **Language:** English **Downstream-task:** Extractive QA **Training data:** SQuAD v1 **Eval data:** SQuAD v1 **Code:** See [example](https://github.com/ShuHuang/batterybert) **Infrastructure**: 8x DGX A100 ## Hyperparameters ``` batch_size = 32 n_epochs = 3 base_LM_model = "bert-base-uncased" max_seq_len = 386 learning_rate = 3e-5 doc_stride=128 max_query_length=64 ``` ## Performance Evaluated on the SQuAD v1.0 dev set. ``` "exact": 80.93, "f1": 88.20, ``` Evaluated on the battery device dataset. ``` "precision": 62.19, "recall": 75.00, ``` ## Usage ### In Transformers ```python from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline model_name = "batterydata/bert-base-uncased-squad-v1" # a) Get predictions nlp = pipeline('question-answering', model=model_name, tokenizer=model_name) QA_input = { 'question': 'What is the electrolyte?', 'context': 'The typical non-aqueous electrolyte for commercial Li-ion cells is a solution of LiPF6 in linear and cyclic carbonates.' } res = nlp(QA_input) # b) Load model & tokenizer model = AutoModelForQuestionAnswering.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) ``` ## Authors Shu Huang: `sh2009 [at] cam.ac.uk` Jacqueline Cole: `jmc61 [at] cam.ac.uk` ## Citation BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
{"language": "en", "license": "apache-2.0", "tags": "question answering", "datasets": ["squad", "batterydata/battery-device-data-qa"], "metrics": "squad"}
question-answering
batterydata/bert-base-uncased-squad-v1
[ "transformers", "pytorch", "bert", "question-answering", "question answering", "en", "dataset:squad", "dataset:batterydata/battery-device-data-qa", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us
# BERT-base-cased for QA Language model: bert-base-uncased Language: English Downstream-task: Extractive QA Training data: SQuAD v1 Eval data: SQuAD v1 Code: See example Infrastructure: 8x DGX A100 ## Hyperparameters ## Performance Evaluated on the SQuAD v1.0 dev set. Evaluated on the battery device dataset. ## Usage ### In Transformers ## Authors Shu Huang: 'sh2009 [at] URL' Jacqueline Cole: 'jmc61 [at] URL' BatteryBERT: A Pre-trained Language Model for Battery Database Enhancement
[ "# BERT-base-cased for QA \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ "TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n", "# BERT-base-cased for QA \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100", "## Hyperparameters", "## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.", "## Usage", "### In Transformers", "## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ 67, 67, 5, 23, 3, 6, 50 ]
[ "passage: TAGS\n#transformers #pytorch #bert #question-answering #question answering #en #dataset-squad #dataset-batterydata/battery-device-data-qa #license-apache-2.0 #endpoints_compatible #region-us \n# BERT-base-cased for QA \r\nLanguage model: bert-base-uncased\r\nLanguage: English \r\nDownstream-task: Extractive QA \r\nTraining data: SQuAD v1\r\nEval data: SQuAD v1\r\nCode: See example \r\nInfrastructure: 8x DGX A100## Hyperparameters## Performance\r\nEvaluated on the SQuAD v1.0 dev set.\r\n\r\nEvaluated on the battery device dataset.## Usage### In Transformers## Authors\r\nShu Huang: 'sh2009 [at] URL'\r\n\r\nJacqueline Cole: 'jmc61 [at] URL'\r\n\r\nBatteryBERT: A Pre-trained Language Model for Battery Database Enhancement" ]
[ -0.08233732730150223, 0.04036601632833481, -0.004108112771064043, 0.05932421609759331, 0.22697557508945465, 0.033763810992240906, 0.1229172870516777, 0.04320504516363144, -0.10582517832517624, -0.031084313988685608, 0.04113712161779404, -0.02831280417740345, 0.07124070823192596, 0.07246532291173935, 0.0069641415029764175, -0.10598722845315933, 0.05146188661456108, -0.017674662172794342, -0.002047397196292877, 0.13380834460258484, 0.11374549567699432, -0.09296327084302902, 0.11312771588563919, 0.009924502111971378, -0.09377788007259369, 0.06567202508449554, -0.014292634092271328, -0.08079706132411957, 0.09910152107477188, -0.06517492979764938, 0.1299445629119873, 0.04785296693444252, 0.022727303206920624, -0.16283361613750458, 0.03652968630194664, 0.03633655980229378, 0.032211609184741974, 0.038860008120536804, 0.01740567944943905, -0.01690179482102394, 0.03681225702166557, 0.03922991082072258, 0.03748681768774986, 0.02381402626633644, -0.05335008725523949, -0.23106619715690613, -0.05749617516994476, -0.06653088331222534, 0.08038823306560516, 0.1316002607345581, -0.02623722143471241, 0.08661685883998871, -0.0507136732339859, 0.07282940298318863, 0.15820716321468353, -0.12160572409629822, -0.003719600150361657, 0.09678854048252106, 0.04896121099591255, -0.004386276472359896, -0.007879412733018398, -0.07874611020088196, 0.0037104252260178328, 0.05006525665521622, 0.052385225892066956, -0.054836858063936234, -0.15498240292072296, 0.1066850796341896, -0.10189666599035263, -0.003787756897509098, 0.20751984417438507, -0.036760471761226654, -0.07858390361070633, 0.01279725506901741, -0.1214166209101677, -0.03245248273015022, -0.02629871666431427, -0.021941669285297394, -0.019710468128323555, 0.0005018265801481903, -0.07847832143306732, 0.02544488199055195, -0.017214637249708176, -0.11813876032829285, -0.1212293729186058, 0.0971565693616867, 0.0620151124894619, 0.03566114604473114, -0.12146616727113724, 0.08569250255823135, -0.0007874012808315456, -0.1544034332036972, -0.03264669328927994, -0.06123218685388565, -0.07631833106279373, 0.008973993360996246, 0.009732798673212528, 0.003912169020622969, 0.12034264951944351, 0.08146876096725464, -0.01836935617029667, 0.01382566150277853, 0.006964772008359432, -0.06882869452238083, -0.002563093788921833, 0.14969182014465332, -0.045780133455991745, -0.11971910297870636, -0.0037884556222707033, 0.034688644111156464, -0.011095769703388214, -0.013380785472691059, -0.0722522959113121, -0.004838251043111086, 0.06885595619678497, 0.055783215910196304, 0.0023185862228274345, 0.04588336870074272, -0.05716433376073837, -0.0373460091650486, 0.04761822894215584, -0.10452453792095184, -0.05524956434965134, -0.02074248716235161, 0.008788935840129852, 0.0043857526034116745, -0.03964037820696831, 0.020831605419516563, -0.03466573357582092, -0.0325060673058033, -0.07936415821313858, -0.09772893041372299, -0.12944939732551575, -0.09835299849510193, -0.03476395830512047, 0.10648903995752335, 0.06125166267156601, -0.1287103146314621, -0.3359777629375458, 0.09320224076509476, 0.01093155425041914, 0.004858430474996567, -0.03733668848872185, -0.03410430997610092, -0.04548868536949158, -0.032239049673080444, -0.04307433217763901, 0.12313473224639893, -0.06161815673112869, 0.07468301057815552, 0.1603616178035736, 0.02375272661447525, -0.049676887691020966, 0.051247116178274155, -0.1361647993326187, 0.005042492412030697, -0.13931503891944885, 0.047269556671381, -0.11842622607946396, -0.008058467879891396, -0.11168720573186874, -0.07025453448295593, 0.11582060158252716, -0.004799260757863522, 0.0898713618516922, 0.14634156227111816, -0.09837780147790909, -0.11429271101951599, 0.025162724778056145, -0.1309940367937088, -0.1485922932624817, 0.05203777924180031, -0.018971510231494904, 0.025487804785370827, 0.02622775174677372, 0.050674233585596085, 0.12440378963947296, -0.16492460668087006, -0.1595999300479889, -0.014308548532426357, 0.05174211785197258, -0.025497831404209137, 0.09295172989368439, 0.043133269995450974, -0.08253125846385956, 0.07653046399354935, -0.07270560413599014, 0.05746777355670929, -0.04993514344096184, -0.10233239829540253, -0.04964292794466019, -0.042987145483493805, 0.1174352616071701, 0.006661520339548588, -0.04133366420865059, 0.02059081941843033, -0.10268904268741608, 0.12648431956768036, 0.11584709584712982, -0.05447368696331978, -0.005816091783344746, -0.19140036404132843, 0.08822073042392731, -0.052359286695718765, 0.022295281291007996, -0.11742890626192093, -0.1565975844860077, 0.0865698978304863, -0.039990007877349854, -0.0677952766418457, -0.025275535881519318, 0.02117026597261429, 0.03234255686402321, -0.1029978096485138, -0.027642641216516495, -0.14709049463272095, -0.018233370035886765, 0.001414251746609807, -0.13932354748249054, 0.03243473172187805, -0.034801505506038666, 0.08391126990318298, 0.026804041117429733, 0.023027455434203148, 0.11611592769622803, 0.03103822097182274, -0.02070462703704834, -0.018695570528507233, -0.019915202632546425, 0.06287269294261932, -0.032105691730976105, -0.03933852165937424, -0.025889378041028976, -0.010490783490240574, -0.005518466699868441, -0.03233467787504196, -0.10147310793399811, 0.0928315594792366, 0.14093369245529175, 0.11588888615369797, 0.012260069139301777, -0.07662000507116318, -0.0022243596613407135, -0.04186757281422615, -0.08440912514925003, -0.0350857637822628, 0.22667954862117767, 0.0033640204928815365, 0.038230206817388535, -0.03915559500455856, 0.015966080129146576, -0.0015175712760537863, 0.04597450792789459, 0.06232400983572006, 0.08060535788536072, 0.10382852703332901, 0.05286315828561783, 0.06307157874107361, 0.06848262995481491, -0.07078617811203003, 0.13082394003868103, -0.08954263478517532, -0.040720779448747635, -0.034508127719163895, -0.10059811919927597, -0.024162661284208298, 0.26431557536125183, -0.15924833714962006, 0.06554882228374481, 0.06638414412736893, -0.008059979416429996, 0.02860591560602188, -0.03444455936551094, -0.006200516130775213, -0.01510626170784235, -0.058726716786623, -0.16224442422389984, 0.019925838336348534, 0.04896024987101555, 0.013005352579057217, -0.013443369418382645, -0.0603608600795269, 0.0270993672311306, 0.009392465464770794, -0.06564730405807495, 0.22918204963207245, -0.11067597568035126, -0.19654448330402374, -0.10930535197257996, -0.09681224822998047, -0.037241268903017044, -0.06434846669435501, 0.022707337513566017, -0.13059739768505096, -0.09851446747779846, 0.015991326421499252, 0.17478930950164795, 0.019392235204577446, 0.06519918888807297, 0.03706338256597519, -0.002363404957577586, 0.018463097512722015, -0.12114239484071732, 0.017793677747249603, 0.04838882386684418, -0.10340847074985504, 0.028473732993006706, 0.04640521481633186, 0.07971616089344025, 0.09694257378578186, 0.010992995463311672, 0.016355130821466446, -0.054799195379018784, 0.2191661149263382, -0.09203644096851349, -0.040463924407958984, 0.2858467102050781, 0.1138538047671318, -0.005657169967889786, 0.1963973194360733, -0.026125362142920494, -0.019249556586146355, 0.09762493520975113, -0.012470249086618423, -0.05032650753855705, -0.24267275631427765, -0.075261689722538, -0.03155232220888138, 0.05990685150027275, -0.06187291815876961, 0.009860742837190628, -0.06790357083082199, 0.09395723789930344, 0.019405093044042587, -0.000024473294615745544, -0.09529406577348709, 0.09079355001449585, 0.23169125616550446, 0.05274002254009247, 0.10671032220125198, -0.006005631759762764, 0.014677586033940315, 0.063950315117836, 0.06405088305473328, 0.16335365176200867, -0.06280229985713959, 0.05319979041814804, 0.030752819031476974, 0.15223948657512665, 0.03163604438304901, 0.05038424953818321, -0.013130126520991325, 0.013368802145123482, 0.006226907018572092, -0.042227134108543396, 0.005304871127009392, -0.02447722852230072, -0.049649763852357864, 0.016761086881160736, -0.017835842445492744, 0.13181151449680328, 0.03564724698662758, 0.015663715079426765, 0.023076845332980156, -0.21419991552829742, -0.03872976079583168, 0.03941497579216957, -0.04077345132827759, -0.10418610274791718, 0.03240383043885231, 0.02306988276541233, -0.11735376715660095, 0.025519253686070442, -0.037192899733781815, 0.0708608403801918, -0.055204201489686966, 0.05761948227882385, 0.06747681647539139, 0.12294940650463104, 0.017357978969812393, 0.12157917767763138, -0.41073116660118103, 0.10371758043766022, 0.045130155980587006, 0.08578531444072723, 0.03463965654373169, 0.06314992904663086, -0.05028533190488815, -0.05512392148375511, 0.05398374795913696, -0.012818912044167519, 0.12147048115730286, 0.019029101356863976, -0.15723870694637299, 0.15016114711761475, -0.002062625717371702, -0.005509513895958662, 0.07411099225282669, -0.01887105405330658, 0.03602968901395798, -0.0032759413588792086, 0.004815050400793552, -0.14116781949996948, -0.08409556746482849, 0.053468819707632065, -0.05847488343715668, 0.11621348559856415, -0.14684122800827026, -0.09026582539081573, 0.04604145884513855, 0.06719302386045456, -0.07238823175430298, -0.11594750732183456, -0.1372605562210083, 0.005869285203516483, 0.11548518389463425, -0.1017613634467125, 0.07118165493011475, 0.0032597847748547792, 0.06035730987787247, -0.013381276279687881, -0.10855953395366669, 0.062237247824668884, -0.08271253108978271, -0.13660044968128204, -0.029798181727528572, 0.022404955700039864, 0.09291037917137146, 0.05105415731668472, 0.0012360934633761644, -0.021272683516144753, -0.1337161809206009, -0.10920193791389465, -0.05069749802350998, -0.04298917576670647, 0.04764114320278168, 0.02268972247838974, 0.11080406606197357, -0.02909649908542633, -0.024577006697654724, -0.019394664093852043, 0.0032764871139079332, 0.1401735097169876, -0.014775224961340427, 0.05363820865750313, 0.14001189172267914, -0.015559817664325237, -0.21115098893642426, -0.06189275532960892, 0.08450467139482498, 0.05646162107586861, -0.04307716339826584, -0.2656063735485077, 0.14306773245334625, -0.01728886365890503, -0.06391707062721252, -0.11285948008298874, -0.1144501119852066, -0.11782537400722504, 0.20664304494857788, 0.08288545906543732, 0.06688032299280167, -0.07058584690093994, -0.01776464283466339, -0.05737828090786934, -0.2627648115158081, 0.13098685443401337, -0.15432949364185333, 0.034624841064214706, -0.01953028328716755, 0.21286416053771973, -0.027195097878575325, -0.033565543591976166, 0.135403573513031, 0.05797842517495155, -0.015492522157728672, 0.03384882211685181, -0.0012714986223727465, 0.11994916200637817, 0.02568202279508114, 0.04866120219230652, -0.0012670470168814063, 0.11640598624944687, -0.11541599780321121, -0.04083337262272835, -0.05231054872274399, 0.08406542241573334, -0.03836920112371445, -0.1605449616909027, -0.0918300449848175, 0.03712160885334015, -0.044955722987651825, -0.059009019285440445, 0.08765457570552826, 0.04417470470070839, -0.009286434389650822, 0.10888928920030594, 0.1641775667667389, 0.018413428217172623, -0.06303144991397858, 0.039819780737161636, -0.016478976234793663, 0.07528345286846161, -0.1393822431564331, 0.03529844060540199, 0.10766949504613876, 0.017312631011009216, 0.02261035516858101, 0.047560740262269974, -0.08703887462615967, 0.00450174231082201, 0.06419365108013153, -0.14690272510051727, -0.14241763949394226, -0.07118149101734161, 0.1645306646823883, -0.2064841240644455, 0.06918592005968094, 0.1200861856341362, -0.016629325225949287, -0.06464719027280807, 0.030350401997566223, 0.02703387662768364, -0.012856281362473965, 0.17982663214206696, 0.0834670215845108, 0.13248635828495026, -0.12922072410583496, 0.06316835433244705, 0.02935277298092842, 0.029933130368590355, -0.029137758538126945, -0.008812269195914268, -0.10453183948993683, 0.012295570224523544, 0.00785012450069189, 0.14331349730491638, -0.11518675833940506, -0.07600440829992294, -0.16325949132442474, -0.12839658558368683, 0.006768000777810812, 0.07147970050573349, 0.09042215347290039, 0.0789489820599556, -0.04616035893559456, 0.027336355298757553, -0.07265792042016983, 0.10823366045951843, 0.003742536762729287, 0.030982768163084984, -0.11455989629030228, -0.07903716713190079, -0.029577437788248062, 0.1132727861404419, -0.05714207515120506, -0.0033168145455420017, -0.09286914020776749, 0.024220280349254608, -0.18211817741394043, -0.0684366524219513, -0.022671081125736237, -0.01753508485853672, 0.01867561787366867, -0.07395017147064209, -0.03394307196140289, 0.06414436548948288, -0.044684991240501404, 0.05882955715060234, 0.04843322932720184, 0.01853620447218418, -0.1270264983177185, 0.09015185385942459, 0.0692223608493805, -0.05849096551537514, 0.11957956850528717, 0.06913658231496811, 0.06730435788631439, 0.11312441527843475, -0.066995769739151, 0.017994629219174385, 0.00037781192804686725, 0.08068525791168213, 0.07583294063806534, -0.12815168499946594, 0.004838071297854185, -0.011052653193473816, 0.011964150704443455, 0.015544920228421688, 0.08112732321023941, -0.0961797833442688, 0.05907364934682846, -0.05725444480776787, 0.03996877372264862, -0.05713101848959923, 0.04794410243630409, 0.09976916015148163, 0.10311779379844666, 0.10571029782295227, -0.036612946540117264, -0.01889960654079914, -0.11847484111785889, 0.012589797377586365, -0.027316898107528687, -0.02404969185590744, -0.05688905343413353, -0.041722677648067474, 0.06641431897878647, 0.02829795330762863, 0.14165519177913666, -0.014617704786360264, 0.048242371529340744, 0.006105831358581781, 0.0944124162197113, 0.10987365990877151, -0.010221018455922604, 0.10490845888853073, 0.09634527564048767, 0.08728007972240448, 0.10965832322835922, -0.014904789626598358, 0.06275081634521484, 0.12436646968126297, 0.10125510394573212, 0.0697845071554184, 0.029525453224778175, 0.049354348331689835, 0.04693176597356796, -0.10340752452611923, -0.07950242608785629, -0.0011123502627015114, -0.06516122072935104, 0.024757947772741318, -0.009157800115644932, 0.19627772271633148, 0.07283085584640503, -0.11493434756994247, 0.077757827937603, -0.0009137274464592338, -0.10501705855131149, -0.09818019717931747, 0.09724146872758865, -0.04374343529343605, -0.21021464467048645, 0.029493756592273712, -0.1583610326051712, -0.12010534107685089, 0.191877081990242, 0.002686090534552932, -0.003763829357922077, 0.14186839759349823, 0.19153891503810883, -0.03275056555867195, -0.028665775433182716, -0.015508591197431087, 0.004902831744402647, 0.05444703251123428, 0.021795887500047684, -0.024101251736283302, 0.11140848696231842, 0.04726942628622055, 0.00909425038844347, -0.06790126115083694, 0.01796935684978962, 0.0011469261953607202, -0.03327450528740883, -0.02137726917862892, 0.00989561341702938, -0.07106970250606537, 0.19388245046138763, -0.005430401302874088, 0.07801692187786102, 0.01581036113202572, 0.3028891384601593, -0.05395740643143654, -0.14118807017803192, -0.1340785175561905, 0.08107198774814606, -0.05108758807182312, 0.05781121551990509, 0.06028646230697632, -0.07645168155431747, -0.12735363841056824, 0.1645081788301468, 0.026814831420779228, -0.0303504578769207, -0.04996722191572189, 0.03811672329902649, -0.021456733345985413, -0.06496179848909378, 0.052432890981435776, 0.08799803256988525, 0.18886560201644897, -0.06352667510509491, -0.02131616324186325, -0.04456266015768051, -0.0742589607834816, -0.007172923535108566, 0.022799525409936905, 0.012890635058283806, -0.031832486391067505, -0.023351293057203293, 0.0328470841050148, -0.09257828444242477, -0.11811418831348419, -0.11672601848840714, -0.02605140581727028, -0.11494872719049454, -0.0747869461774826, 0.1839233636856079, 0.006930411793291569, 0.005162521731108427, -0.04791936278343201, 0.0874776840209961, 0.062455419450998306, 0.03634749725461006, -0.08744633197784424, -0.06949494779109955, 0.19501858949661255, -0.13099661469459534, 0.1422777771949768, -0.043704912066459656, 0.08008944243192673, 0.10982989519834518, -0.021190373227000237, -0.11508993804454803, 0.06669976562261581, 0.05317945405840874, -0.1747073531150818, -0.0010079194325953722, 0.1137373223900795, 0.012152031995356083, 0.11794164776802063, -0.013473501428961754, 0.07305213809013367, -0.022567402571439743, -0.004221112933009863, -0.048698216676712036, -0.05753059685230255, -0.028225049376487732, -0.08175864070653915, 0.04721800610423088, 0.07911107689142227, -0.058590490370988846, 0.06067334860563278, -0.08634285628795624, 0.10348351299762726, -0.030540980398654938, 0.06781039386987686, 0.025166088715195656, -0.12093143910169601, -0.06570494920015335, 0.1402081549167633, -0.0641988143324852, -0.13414210081100464, 0.01651773601770401, -0.010261055082082748, 0.00537582952529192, -0.08269772678613663, 0.1298946738243103, 0.0643027126789093, 0.005507471971213818, -0.026424886658787727, 0.04083043336868286, -0.0635361298918724, 0.06939048320055008, -0.17124243080615997, -0.1012323647737503 ]
null
null
transformers
# ALBERT-Mongolian [pretraining repo link](https://github.com/bayartsogt-ya/albert-mongolian) ## Model description Here we provide pretrained ALBERT model and trained SentencePiece model for Mongolia text. Training data is the Mongolian wikipedia corpus from Wikipedia Downloads and Mongolian News corpus. ## Evaluation Result: ``` loss = 1.7478163 masked_lm_accuracy = 0.6838185 masked_lm_loss = 1.6687671 sentence_order_accuracy = 0.998125 sentence_order_loss = 0.007942731 ``` ## Fine-tuning Result on Eduge Dataset: ``` precision recall f1-score support ะฑะฐะนะณะฐะป ะพั€ั‡ะธะฝ 0.85 0.83 0.84 999 ะฑะพะปะพะฒัั€ะพะป 0.80 0.80 0.80 873 ัะฟะพั€ั‚ 0.98 0.98 0.98 2736 ั‚ะตั…ะฝะพะปะพะณะธ 0.88 0.93 0.91 1102 ัƒะปั ั‚ำฉั€ 0.92 0.85 0.89 2647 ัƒั€ะปะฐะณ ัะพั‘ะป 0.93 0.94 0.94 1457 ั…ัƒัƒะปัŒ 0.89 0.87 0.88 1651 ัะดะธะนะฝ ะทะฐัะฐะณ 0.83 0.88 0.86 2509 ัั€าฏาฏะป ะผัะฝะด 0.89 0.92 0.90 1159 accuracy 0.90 15133 macro avg 0.89 0.89 0.89 15133 weighted avg 0.90 0.90 0.90 15133 ``` ## Reference 1. [ALBERT - official repo](https://github.com/google-research/albert) 2. [WikiExtrator](https://github.com/attardi/wikiextractor) 3. [Mongolian BERT](https://github.com/tugstugi/mongolian-bert) 4. [ALBERT - Japanese](https://github.com/alinear-corp/albert-japanese) 5. [Mongolian Text Classification](https://github.com/sharavsambuu/mongolian-text-classification) 6. [You's paper](https://arxiv.org/abs/1904.00962) ## Citation ``` @misc{albert-mongolian, author = {Bayartsogt Yadamsuren}, title = {ALBERT Pretrained Model on Mongolian Datasets}, year = {2020}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/bayartsogt-ya/albert-mongolian/}} } ``` ## For More Information Please contact by [email protected]
{"language": "mn"}
fill-mask
bayartsogt/albert-mongolian
[ "transformers", "pytorch", "tf", "safetensors", "albert", "fill-mask", "mn", "arxiv:1904.00962", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1904.00962" ]
[ "mn" ]
TAGS #transformers #pytorch #tf #safetensors #albert #fill-mask #mn #arxiv-1904.00962 #autotrain_compatible #endpoints_compatible #region-us
# ALBERT-Mongolian pretraining repo link ## Model description Here we provide pretrained ALBERT model and trained SentencePiece model for Mongolia text. Training data is the Mongolian wikipedia corpus from Wikipedia Downloads and Mongolian News corpus. ## Evaluation Result: ## Fine-tuning Result on Eduge Dataset: ## Reference 1. ALBERT - official repo 2. WikiExtrator 3. Mongolian BERT 4. ALBERT - Japanese 5. Mongolian Text Classification 6. You's paper ## For More Information Please contact by bayartsogtyadamsuren@URL
[ "# ALBERT-Mongolian\npretraining repo link", "## Model description\nHere we provide pretrained ALBERT model and trained SentencePiece model for Mongolia text. Training data is the Mongolian wikipedia corpus from Wikipedia Downloads and Mongolian News corpus.", "## Evaluation Result:", "## Fine-tuning Result on Eduge Dataset:", "## Reference\n1. ALBERT - official repo\n2. WikiExtrator\n3. Mongolian BERT\n4. ALBERT - Japanese\n5. Mongolian Text Classification\n6. You's paper", "## For More Information\nPlease contact by bayartsogtyadamsuren@URL" ]
[ "TAGS\n#transformers #pytorch #tf #safetensors #albert #fill-mask #mn #arxiv-1904.00962 #autotrain_compatible #endpoints_compatible #region-us \n", "# ALBERT-Mongolian\npretraining repo link", "## Model description\nHere we provide pretrained ALBERT model and trained SentencePiece model for Mongolia text. Training data is the Mongolian wikipedia corpus from Wikipedia Downloads and Mongolian News corpus.", "## Evaluation Result:", "## Fine-tuning Result on Eduge Dataset:", "## Reference\n1. ALBERT - official repo\n2. WikiExtrator\n3. Mongolian BERT\n4. ALBERT - Japanese\n5. Mongolian Text Classification\n6. You's paper", "## For More Information\nPlease contact by bayartsogtyadamsuren@URL" ]
[ 55, 12, 44, 5, 12, 36, 15 ]
[ "passage: TAGS\n#transformers #pytorch #tf #safetensors #albert #fill-mask #mn #arxiv-1904.00962 #autotrain_compatible #endpoints_compatible #region-us \n# ALBERT-Mongolian\npretraining repo link## Model description\nHere we provide pretrained ALBERT model and trained SentencePiece model for Mongolia text. Training data is the Mongolian wikipedia corpus from Wikipedia Downloads and Mongolian News corpus.## Evaluation Result:## Fine-tuning Result on Eduge Dataset:## Reference\n1. ALBERT - official repo\n2. WikiExtrator\n3. Mongolian BERT\n4. ALBERT - Japanese\n5. Mongolian Text Classification\n6. You's paper## For More Information\nPlease contact by bayartsogtyadamsuren@URL" ]
[ 0.036042988300323486, 0.020495785400271416, -0.0018841546261683106, 0.05069993436336517, 0.1334575116634369, -0.023276912048459053, 0.23588824272155762, 0.07094898074865341, 0.08413269370794296, -0.03314704820513725, 0.10511712729930878, -0.08201226592063904, 0.04849732294678688, 0.16120414435863495, 0.08010072261095047, -0.37812569737434387, 0.058005139231681824, -0.005129088647663593, 0.07579468935728073, 0.06776198744773865, 0.11343872547149658, -0.05185646563768387, 0.04545808210968971, -0.001607121666893363, -0.04450174421072006, 0.06438031792640686, -0.051199495792388916, -0.1611671894788742, 0.11720281094312668, 0.03625750169157982, 0.0846954807639122, 0.023057861253619194, 0.04820811375975609, -0.13292180001735687, 0.023497875779867172, -0.07423461973667145, -0.08576659113168716, -0.0035555940121412277, 0.11447646468877792, -0.06595758348703384, 0.13802066445350647, -0.08841156214475632, -0.008227108046412468, -0.007621865253895521, -0.08312492817640305, -0.08433309942483902, 0.0012199782067909837, 0.10793708264827728, 0.15640133619308472, 0.11634834855794907, -0.061754293739795685, 0.09592573344707489, -0.12095070630311966, 0.06291015446186066, 0.1770666390657425, -0.3572312295436859, -0.040645044296979904, 0.1496134251356125, 0.013574481010437012, 0.00813044048845768, -0.17163854837417603, -0.002523232251405716, 0.08024043589830399, -0.014060016721487045, 0.04182847961783409, -0.11601756513118744, -0.18186157941818237, -0.0633637011051178, -0.11328674107789993, 0.08607553690671921, 0.22436287999153137, 0.00672693457454443, -0.046835221350193024, -0.005568210035562515, 0.023712914437055588, -0.025195809081196785, 0.00786392018198967, 0.0615420937538147, 0.006633386947214603, -0.05903998389840126, 0.015811491757631302, 0.00265760556794703, -0.136321023106575, 0.02475493773818016, -0.10921745747327805, 0.13869953155517578, 0.04316246137022972, -0.006894105114042759, -0.08239808678627014, 0.030435170978307724, -0.13271097838878632, -0.11905775964260101, 0.021094029769301414, -0.0562174990773201, -0.07956656813621521, -0.048965178430080414, 0.014951660297811031, -0.12575869262218475, 0.05648059397935867, 0.17373910546302795, -0.024755941703915596, 0.11683434993028641, -0.05839976295828819, 0.06395578384399414, -0.0034498784225434065, 0.17301519215106964, -0.01981847919523716, -0.044557154178619385, 0.019970187917351723, -0.029573073610663414, 0.03954948112368584, 0.03085910715162754, -0.11953277885913849, -0.1121976301074028, 0.0867152139544487, 0.07396664470434189, -0.06539247930049896, 0.08216775953769684, 0.022810375317931175, 0.030514709651470184, 0.023658985272049904, -0.11494369804859161, -0.05163060501217842, -0.040154214948415756, -0.12654350697994232, 0.0089241499081254, -0.01241773460060358, 0.027238817885518074, 0.00022773390810471028, 0.07779181003570557, -0.012305679731070995, 0.10238263010978699, 0.0035573518835008144, -0.07418843358755112, 0.07984766364097595, 0.044907376170158386, 0.07358349859714508, -0.19473814964294434, -0.05777423828840256, -0.023493802174925804, 0.05481022968888283, -0.06463329493999481, 0.01710214838385582, -0.058800023049116135, -0.042074136435985565, 0.04239590838551521, -0.06776025146245956, 0.012721460312604904, -0.06410839408636093, -0.017360832542181015, -0.0159015990793705, 0.11851116269826889, 0.016195571050047874, -0.0016910404665395617, -0.1256515085697174, 0.0281122624874115, -0.14675520360469818, -0.039511896669864655, -0.08464021235704422, 0.11973221600055695, -0.06870618462562561, -0.011005306616425514, 0.01980837620794773, -0.0006774063222110271, -0.006688347086310387, 0.20398657023906708, -0.09425780177116394, -0.023117192089557648, 0.27106910943984985, -0.0652061477303505, -0.03232800215482712, 0.1341852843761444, -0.020803334191441536, 0.0517236664891243, 0.06964816898107529, 0.07405806332826614, 0.001982017420232296, -0.08004266768693924, -0.03259080648422241, 0.04332268610596657, 0.04375129193067551, 0.07839430868625641, 0.120179183781147, 0.06513425707817078, -0.026762155815958977, 0.05074036493897438, 0.05364605784416199, 0.05460856482386589, -0.039539627730846405, -0.04450501874089241, 0.07214200496673584, -0.02765851654112339, 0.15484613180160522, 0.01579868793487549, 0.1384541094303131, -0.0911831483244896, -0.09825616329908371, -0.044657886028289795, 0.059545617550611496, -0.004196573980152607, 0.0397627092897892, -0.049069296568632126, -0.0036618001759052277, -0.04570130631327629, 0.006466908846050501, -0.102165088057518, -0.036492422223091125, 0.036043260246515274, 0.02796083316206932, 0.03427880257368088, 0.023066161200404167, 0.06207021325826645, 0.0023943493142724037, -0.14963886141777039, -0.03726635500788689, 0.03501661866903305, -0.026409663259983063, -0.08003578335046768, -0.17110386490821838, 0.08550501614809036, -0.04243446886539459, 0.10158931463956833, -0.18105007708072662, -0.009084239602088928, -0.06690593808889389, 0.048399150371551514, -0.010706168599426746, 0.07069563120603561, 0.05921478942036629, 0.03505876287817955, -0.07712098956108093, 0.011608103290200233, 0.04569247364997864, -0.011146892793476582, -0.1907787173986435, 0.16415178775787354, -0.07905086129903793, -0.0280485637485981, 0.14025911688804626, -0.08299040049314499, -0.09916947036981583, -0.0051718102768063545, -0.0427020899951458, -0.02987038902938366, -0.06179626286029816, 0.019926322624087334, 0.09880486130714417, -0.02303820475935936, 0.10946498811244965, 0.0003912628744728863, -0.027392003685235977, -0.012945889495313168, -0.14404098689556122, -0.07968390733003616, 0.15540455281734467, -0.028529560193419456, -0.10564343631267548, 0.13804595172405243, 0.05083663761615753, -0.07433561235666275, 0.2002991884946823, 0.04503149539232254, -0.05551757290959358, 0.015621785074472427, 0.027258142828941345, 0.015457249246537685, 0.11873901635408401, -0.06878794729709625, -0.04699694737792015, 0.029112564399838448, 0.01810329593718052, -0.00002490855877113063, -0.08092173933982849, -0.05560322478413582, -0.041685692965984344, -0.033462364226579666, -0.005435720551759005, 0.07798831164836884, -0.04760841652750969, 0.1313193142414093, 0.003987635485827923, -0.028097033500671387, 0.028707005083560944, -0.028915904462337494, -0.10376813262701035, 0.19180354475975037, -0.08620749413967133, -0.3476748466491699, -0.06968037784099579, -0.13865774869918823, -0.017920205369591713, 0.034042708575725555, 0.04209781810641289, -0.21980929374694824, -0.07327746599912643, -0.0324651263654232, 0.05213293805718422, -0.04641560837626457, 0.05910002440214157, -0.006427759770303965, 0.07133641093969345, -0.01618485525250435, -0.007802095729857683, -0.02635837532579899, 0.0013781036250293255, -0.07967256009578705, 0.1603170484304428, -0.1018751859664917, 0.0326031818985939, 0.023841923102736473, 0.0007638427196070552, -0.066482774913311, -0.052148934453725815, 0.1028166115283966, -0.11123979836702347, 0.046593621373176575, 0.03971720114350319, -0.0897270143032074, -0.0029425767716020346, 0.18359923362731934, -0.018248265609145164, -0.01788869872689247, 0.13470493257045746, 0.01898307166993618, -0.07468888908624649, -0.28146272897720337, -0.01944783888757229, -0.047983311116695404, 0.09395010769367218, 0.049250517040491104, 0.05317229405045509, 0.040946077555418015, 0.058466363698244095, -0.03527190163731575, 0.10079257190227509, 0.07575833052396774, 0.09352871775627136, -0.07050777226686478, 0.06282564997673035, 0.03888536989688873, -0.08487449586391449, -0.012003310024738312, 0.020453307777643204, -0.004144053440541029, 0.1981811672449112, 0.07990624010562897, 0.014271629974246025, 0.03365731239318848, 0.023182179778814316, 0.10606706887483597, 0.03755580261349678, -0.011637150309979916, -0.056681063026189804, 0.05805639177560806, -0.04633918032050133, 0.05263233557343483, 0.0834210216999054, -0.055921878665685654, 0.011692470870912075, -0.07897316664457321, 0.009460322558879852, 0.10375353693962097, 0.06279145926237106, 0.0776345506310463, -0.12620708346366882, -0.10773768275976181, 0.08495625853538513, -0.11060044914484024, -0.00946708396077156, 0.058725133538246155, 0.022645538672804832, -0.2059072107076645, 0.14817385375499725, -0.020836621522903442, 0.07764111459255219, -0.001056249369867146, -0.004686790518462658, -0.055372290313243866, -0.08601617813110352, -0.032101355493068695, 0.08415251970291138, -0.4143708050251007, 0.32509490847587585, 0.0016035913722589612, -0.005566518288105726, -0.05802904814481735, -0.01048247329890728, 0.099729023873806, 0.024484999477863312, 0.17064118385314941, -0.019133668392896652, -0.026545964181423187, -0.11480391025543213, -0.1122054010629654, 0.004494675900787115, 0.03864571824669838, -0.04668902978301048, 0.12055103480815887, 0.017501236870884895, -0.019967880100011826, -0.046364083886146545, 0.055163271725177765, -0.1576215624809265, -0.06762772053480148, -0.006701433099806309, -0.0407877154648304, 0.08206184953451157, -0.032332129776477814, -0.1012783870100975, -0.1439545750617981, 0.03893306478857994, -0.09114391356706619, -0.10581136494874954, -0.055062491446733475, 0.12841029465198517, 0.0683654323220253, -0.1272469013929367, -0.028030775487422943, -0.018359970301389694, 0.1249055340886116, -0.022444505244493484, -0.055995069444179535, 0.023341747000813484, -0.11521494388580322, -0.12541824579238892, -0.024208763614296913, 0.1229838952422142, 0.14118066430091858, 0.1599285900592804, 0.004446530714631081, 0.0005877609364688396, 0.045915644615888596, -0.07270311564207077, -0.021450506523251534, -0.04490523040294647, -0.0015365170547738671, 0.01790435053408146, -0.1090962216258049, -0.20737166702747345, -0.1390485316514969, -0.12845465540885925, 0.17516939342021942, 0.25969284772872925, -0.03295522928237915, 0.10819120705127716, 0.09024681150913239, -0.08665937930345535, -0.19868458807468414, 0.05727870389819145, 0.05214069411158562, 0.0595291331410408, -0.004263588227331638, -0.1737135797739029, 0.08572766184806824, 0.03801540285348892, -0.01730496622622013, 0.00980290211737156, -0.13318753242492676, -0.13211850821971893, 0.12565433979034424, 0.013534147292375565, 0.1683695763349533, -0.1781047284603119, -0.007740400731563568, -0.07053355872631073, -0.12017247825860977, 0.022038381546735764, -0.07597555220127106, 0.07083690911531448, -0.0002843731490429491, 0.012944919988512993, 0.00208973023109138, -0.024266138672828674, 0.18170906603336334, 0.025709915906190872, -0.013832937926054, -0.12387488782405853, -0.06678319722414017, 0.058925386518239975, -0.056516747921705246, 0.04554674029350281, 0.011975953355431557, -0.07431235909461975, -0.15451177954673767, -0.04662640020251274, -0.059891872107982635, 0.033197980374097824, -0.07186230272054672, -0.0896649956703186, -0.03734226152300835, 0.16236253082752228, 0.06976578384637833, -0.011164258234202862, 0.2610653340816498, -0.04089467227458954, 0.12760181725025177, -0.0846610814332962, 0.15725219249725342, 0.12308255583047867, 0.12249100208282471, 0.029821500182151794, -0.027910856530070305, -0.0027295283507555723, -0.16341696679592133, -0.012871813029050827, 0.1034453809261322, -0.051346853375434875, 0.12135649472475052, 0.011442623101174831, -0.05245453864336014, 0.12313669174909592, 0.1280326396226883, -0.11803163588047028, 0.0028541507199406624, -0.0016278756083920598, -0.007393197622150183, -0.010465440340340137, -0.05302957445383072, 0.05383662134408951, -0.11617756634950638, -0.053851958364248276, 0.044933609664440155, 0.04380641505122185, -0.000014272281077865046, 0.059260156005620956, 0.023641586303710938, 0.03371059522032738, -0.02647145465016365, 0.13453637063503265, 0.08934076875448227, -0.05801376327872276, 0.036309558898210526, 0.13635016977787018, -0.1188361868262291, -0.009802084416151047, 0.0209706649184227, 0.09354440122842789, -0.08525781333446503, -0.07644836604595184, -0.103586845099926, -0.06835653632879257, -0.023147258907556534, 0.11775310337543488, 0.07198848575353622, -0.05701906979084015, 0.12138065695762634, 0.04552336409687996, -0.06333120912313461, 0.08344453573226929, 0.10863736271858215, 0.014427189715206623, -0.061061833053827286, 0.030429763719439507, 0.018806196749210358, 0.11309898644685745, -0.05598169565200806, -0.031114688143134117, -0.16296379268169403, 0.06340304017066956, -0.17701832950115204, 0.07011512666940689, -0.20665399730205536, -0.031942497938871384, -0.0786982923746109, -0.10676252841949463, -0.04746009409427643, -0.018557526171207428, -0.09120999276638031, 0.047847453504800797, -0.03339095413684845, 0.0858900398015976, 0.004466762766242027, -0.06439323723316193, 0.0543818436563015, -0.03728993237018585, 0.08086707442998886, 0.0950530543923378, 0.014295673929154873, 0.17442966997623444, -0.2149297446012497, 0.056961171329021454, 0.009725780226290226, -0.03364142030477524, 0.040166452527046204, -0.06202041730284691, -0.0450141578912735, -0.030941514298319817, 0.06084661930799484, 0.02286488376557827, -0.01430217269808054, -0.048722121864557266, -0.09626175463199615, -0.09497002512216568, -0.03565339744091034, -0.0505172535777092, 0.09037751704454422, 0.054668691009283066, -0.009906846098601818, 0.12338630855083466, -0.16765828430652618, 0.09518489241600037, -0.024611959233880043, 0.03529144451022148, -0.04720061272382736, -0.06761391460895538, -0.03831733018159866, -0.06321035325527191, 0.06588536500930786, -0.08098344504833221, 0.13252206146717072, -0.010870094411075115, 0.167586550116539, 0.04958953335881233, -0.13590997457504272, 0.025397183373570442, 0.041879747062921524, 0.12253472208976746, 0.07136063277721405, -0.05511164292693138, -0.04271502047777176, 0.04192772135138512, 0.05280511826276779, -0.03922086954116821, 0.19213852286338806, 0.10314542800188065, 0.03240717574954033, 0.11841791123151779, -0.05502578988671303, -0.0658658891916275, -0.059077877551317215, -0.12206688523292542, -0.11485636234283447, 0.027040189132094383, 0.01689913123846054, 0.07760539650917053, 0.24677909910678864, -0.12478011846542358, 0.010500543750822544, -0.04067371413111687, -0.06730050593614578, -0.1049521416425705, -0.13092073798179626, -0.09260065108537674, -0.03092186711728573, 0.04810800030827522, -0.11792193353176117, -0.017907120287418365, 0.07345675677061081, 0.11527230590581894, -0.019965533167123795, 0.19117595255374908, 0.06100253760814667, -0.11235321313142776, 0.12930001318454742, -0.010905656032264233, 0.009381682612001896, -0.01457892544567585, -0.019257668405771255, -0.0688827782869339, -0.045922741293907166, -0.07549825310707092, 0.02031685784459114, 0.0053406707011163235, 0.018532440066337585, -0.026663154363632202, -0.08533243089914322, 0.03364310786128044, 0.041946541517972946, 0.1168043240904808, 0.09861946851015091, 0.026607709005475044, 0.009311052970588207, -0.033247996121644974, 0.22775138914585114, 0.013909848406910896, -0.026535561308264732, -0.14638765156269073, 0.1016622930765152, 0.013914128765463829, -0.024234436452388763, -0.011603593826293945, -0.09192817658185959, -0.0408516563475132, 0.2489389032125473, 0.15111561119556427, 0.1095619797706604, 0.033801086246967316, -0.06354495882987976, 0.027255745604634285, 0.07737615704536438, 0.09010965377092361, 0.07616852968931198, 0.18947672843933105, -0.12008395045995712, -0.055242374539375305, -0.07117106020450592, 0.0006763007841072977, -0.044128794223070145, 0.07181050628423691, 0.04818989336490631, -0.00014004849072080106, -0.10997812449932098, 0.1014983206987381, -0.21839484572410583, -0.15186378359794617, -0.06347237527370453, -0.11775019764900208, -0.11420971155166626, 0.0025375105906277895, 0.008437326177954674, 0.10732942074537277, 0.053862109780311584, 0.014276395551860332, 0.03596874698996544, -0.06213660538196564, 0.07433664053678513, 0.017028827220201492, -0.06937640905380249, 0.10414793342351913, 0.1525757610797882, 0.19913408160209656, -0.011948753148317337, 0.044515229761600494, 0.10344117879867554, 0.022679677233099937, -0.05756138637661934, 0.06199878454208374, 0.10136894136667252, 0.012988077476620674, 0.05416393652558327, 0.15745368599891663, 0.0198411475867033, 0.07545819878578186, 0.10998356342315674, 0.0634128600358963, 0.0618438757956028, -0.01832493580877781, -0.003913420252501965, -0.10939506441354752, 0.15828675031661987, -0.11332544684410095, 0.10271220654249191, 0.23874488472938538, -0.06588705629110336, -0.053543753921985626, -0.04768988490104675, 0.020825030282139778, 0.0006555106374435127, -0.05077453330159187, -0.0657186433672905, -0.1683017611503601, 0.043067704886198044, -0.05643119290471077, 0.02866996079683304, -0.2115592360496521, -0.017164533957839012, -0.10959091782569885, -0.0017722073243930936, -0.09069900214672089, 0.06032007932662964, 0.023194734007120132, 0.03501855581998825, -0.04076259955763817, -0.11010652780532837, 0.03378453478217125, 0.09325063973665237, -0.1486632525920868, -0.13844920694828033 ]
null
null
null
|fold|accuracy| |-|-| | fold 0 | 0.974197247706422 | | fold 1 | 0.9627293577981652 | | fold 2 | 0.9724770642201835 | | fold 3 | 0.9696100917431193 | | fold 4 | 0.9684633027522935 | | OOF Acc | 0.9694954128440367 |
{}
null
bayartsogt/mlub-bert-base-uncased-tr5meaning
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
|fold|accuracy| |-|-| | fold 0 | 0.9730504587155964 | | fold 1 | 0.9690366972477065 | | fold 2 | 0.970756880733945 | | fold 3 | 0.9684633027522935 | | fold 4 | 0.9719036697247706 | | OOF Acc | 0.9706422018348624 |
{}
null
bayartsogt/mlub-bert-large-cased-tr5do30ep25s42
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
|fold|accuracy| |-|-| | fold 0 | 0.9753440366972477 | | fold 1 | 0.9678899082568807 | | fold 2 | 0.9747706422018348 | | fold 3 | 0.9690366972477065 | | fold 4 | 0.9759174311926605 | | OOF Acc | 0.9725917431192661 |
{}
null
bayartsogt/mlub-bert-large-uncased-tr5do20ep25s42
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
null
|fold|accuracy| |-|-| | fold 0 | 0.974197247706422 | | fold 1 | 0.9678899082568807 | | fold 2 | 0.9724770642201835 | | fold 3 | 0.9701834862385321 | | fold 4 | 0.9736238532110092 | | OOF Acc | 0.9716743119266055 | ``` synset_word ะฐะฒ 1.000000 ะฐะผ 0.931507 ะฑะฐะณ 0.980000 ะฑะฐะนั€ 0.943548 ะฑะฐั€ะฐะฐ 0.964789 ะณะฐั€ 0.950210 ะณะพะป 0.938731 ะณาฏะฝ 0.912088 ะทะฐั… 0.946667 ะทัƒัƒ 0.995798 ะทาฏั€ั… 0.918367 ะผำฉะฝะณำฉ 0.973333 ะฝัƒั€ัƒัƒ 0.968750 ะฝาฏะด 1.000000 ะฝาฏาฏั€ 0.987805 ัะฐะปะฑะฐั€ 0.963636 ัะฐั€ 0.996627 ััƒะผ 0.816667 ั‚ัั€ัะณ 0.822581 ั‚าฏาฏั… 0.980237 ั‚ำฉั€ 0.998428 ั…ะธะน 0.993077 ั…ัƒั€ะฐะฐ 0.858268 ั…ัะปะฑัั€ 0.727273 ั…ำฉะฝะดะธะน 1.000000 ัˆะฐั‚ 1.000000 ัะผ 1.000000 ัั€าฏาฏะป 1.000000 dtype: float64 ```
{}
null
bayartsogt/mlub-bert-large-uncased-tr5do30ep25
[ "region:us" ]
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
[]
[ "TAGS\n#region-us \n" ]
[ 6 ]
[ "passage: TAGS\n#region-us \n" ]
[ 0.024608636274933815, -0.026205500587821007, -0.009666500613093376, -0.10395516455173492, 0.08638657629489899, 0.059816278517246246, 0.01882290467619896, 0.020661840215325356, 0.23975107073783875, -0.005599027033895254, 0.1219947561621666, 0.0015615287702530622, -0.037353623658418655, 0.03733762726187706, -0.0035912662278860807, -0.17583473026752472, 0.03876631706953049, -0.018274923786520958, 0.01843859627842903, 0.026470553129911423, -0.07776834815740585, -0.07564429938793182, 0.015296397730708122, -0.10247814655303955, -0.083692267537117, 0.11002834886312485, 0.031466204673051834, -0.019670886918902397, 0.10779199749231339, -0.04243955761194229, 0.18699054419994354, -0.011512263678014278, -0.11213519424200058, -0.2536850869655609, 0.021806683391332626, -0.01765260472893715, -0.08747660368680954, 0.01506110467016697, 0.0665089413523674, -0.09014441072940826, -0.0588928684592247, 0.0795099288225174, -0.01132340170443058, 0.04246443510055542, -0.27593839168548584, -0.12684126198291779, -0.05297930911183357, -0.1421966552734375, 0.08651168644428253, 0.04035491496324539, 0.008764253929257393, 0.15506891906261444, -0.20897391438484192, 0.004104613792151213, 0.08255259692668915, -0.2538507878780365, 0.05591634660959244, 0.17671173810958862, 0.03623908758163452, 0.18037272989749908, 0.0060391901060938835, 0.11029672622680664, 0.0716743916273117, -0.024263937026262283, -0.17590197920799255, -0.08127854019403458, -0.04696211963891983, 0.16642488539218903, -0.06727185100317001, -0.14248386025428772, 0.34701237082481384, 0.00015008423360995948, 0.009657775051891804, 0.16921205818653107, -0.059524230659008026, -0.09972117841243744, 0.07259953022003174, 0.016484731808304787, 0.018492350354790688, 0.1471305936574936, 0.16307872533798218, -0.0458691343665123, -0.13837823271751404, -0.018630273640155792, -0.22798998653888702, 0.17510560154914856, -0.03248048573732376, 0.13137903809547424, -0.27447956800460815, 0.01684025302529335, -0.2570667266845703, 0.0032130838371813297, 0.04178816080093384, -0.06004921346902847, -0.0226522795855999, -0.013265985064208508, -0.08018817007541656, 0.004899587947875261, 0.06192673370242119, 0.1266920566558838, -0.06128726154565811, 0.06128238886594772, -0.09319206327199936, 0.141696035861969, 0.07166698575019836, 0.07868369668722153, 0.13037432730197906, 0.041205424815416336, -0.07187089323997498, -0.21872246265411377, -0.0026476888451725245, -0.06275863200426102, -0.09502086788415909, -0.0020165652967989445, -0.11606067419052124, 0.17244569957256317, -0.030802514404058456, -0.09825427830219269, -0.11208184063434601, 0.09148659557104111, -0.032992321997880936, -0.03437839448451996, -0.03552987426519394, -0.020977836102247238, 0.019381176680326462, 0.04704452306032181, -0.1548958420753479, -0.005131472367793322, 0.07039852440357208, 0.11502562463283539, -0.1346137970685959, -0.003783059772104025, -0.07908964157104492, 0.03039063885807991, 0.07654735445976257, -0.16510222852230072, 0.03158547356724739, -0.1124754324555397, -0.07531405985355377, 0.002912673633545637, -0.015710093080997467, -0.016202643513679504, 0.166526660323143, -0.0020451415330171585, 0.0714716836810112, -0.026345307007431984, -0.05890209600329399, -0.11243434250354767, -0.08489254862070084, 0.05390460044145584, 0.03670717030763626, 0.03266148269176483, -0.2193479984998703, 0.014805203303694725, -0.12762966752052307, 0.1360815018415451, -0.10566820204257965, -0.04705966264009476, -0.022842247039079666, 0.20562705397605896, 0.037286072969436646, 0.08762791007757187, -0.22171171009540558, 0.039756543934345245, -0.05404696613550186, 0.18480908870697021, -0.1502426266670227, -0.0799463614821434, 0.20813211798667908, -0.07964949309825897, -0.10115210711956024, 0.021235812455415726, 0.020391687750816345, 0.026287272572517395, 0.0766737088561058, 0.4564172327518463, -0.09766800701618195, -0.09146861732006073, 0.10178250074386597, 0.17055274546146393, -0.12427149713039398, -0.1827561855316162, 0.06446871906518936, -0.16666454076766968, -0.1973118633031845, 0.0018917324487119913, 0.09222044050693512, 0.038269978016614914, -0.07875611633062363, -0.020746968686580658, 0.06325206160545349, -0.0007678253459744155, 0.09095914661884308, 0.03755716234445572, 0.09034032374620438, -0.08716782182455063, 0.11115926504135132, -0.05017651244997978, 0.004037132486701012, 0.1343354731798172, 0.027325427159667015, -0.03223329409956932, 0.08694463223218918, -0.0485352948307991, 0.05295134335756302, -0.1662379503250122, -0.15068690478801727, 0.03398871049284935, 0.06283251196146011, 0.03186952322721481, 0.1280253529548645, 0.08141885697841644, -0.10732853412628174, 0.022690722718834877, -0.004228927195072174, 0.058398615568876266, 0.03891623765230179, 0.006107209715992212, 0.008764320984482765, 0.0961301177740097, -0.10607069730758667, -0.13589619100093842, -0.07336436957120895, -0.014715781435370445, 0.14371353387832642, -0.0302802175283432, 0.07690227776765823, -0.004240254405885935, 0.00013200697139836848, 0.06930823624134064, 0.08137880265712738, 0.016412746161222458, 0.08971183747053146, -0.05237193778157234, -0.05160155147314072, 0.10863113403320312, -0.13533565402030945, 0.17837053537368774, 0.14053137600421906, -0.20532016456127167, 0.029453208670020103, -0.06838275492191315, 0.03670361638069153, -0.008162540383636951, 0.0975119024515152, -0.08272241055965424, -0.02106042578816414, 0.013134466484189034, 0.0052274600602686405, -0.013007243163883686, 0.017682146281003952, -0.07295988500118256, -0.07787393033504486, -0.10233919322490692, 0.08436838537454605, 0.11562882363796234, -0.10282530635595322, 0.14214380085468292, 0.4384984076023102, 0.11495281755924225, 0.21582984924316406, -0.09581480920314789, -0.0412987545132637, 0.007486371789127588, 0.0001535322517156601, -0.04476691037416458, 0.08031861484050751, -0.15973517298698425, -0.038901735097169876, 0.027348900213837624, 0.07128690183162689, 0.11475157737731934, -0.14959022402763367, -0.09639324247837067, -0.00793045200407505, 0.0022841424215584993, -0.1249532699584961, 0.023905446752905846, -0.03974650055170059, 0.04015624523162842, 0.07232289016246796, -0.021535737439990044, 0.13939237594604492, -0.04166141897439957, -0.0639561116695404, 0.07585346698760986, -0.2017085999250412, -0.23179671168327332, -0.12309670448303223, -0.14680525660514832, 0.04366797208786011, 0.05154111236333847, 0.01726446859538555, -0.17635835707187653, -0.015074856579303741, 0.07706750929355621, 0.07820965349674225, -0.20886357128620148, -0.022814949974417686, -0.004290030337870121, 0.0895976573228836, -0.10227091610431671, -0.0017130117630586028, -0.04419664293527603, -0.10150232166051865, 0.0017003051470965147, 0.07279510796070099, -0.137485533952713, 0.13807645440101624, 0.21589438617229462, 0.07225540280342102, 0.07359948754310608, -0.019093448296189308, 0.09936179965734482, -0.10856141895055771, -0.16549113392829895, 0.08348225057125092, -0.06234746053814888, 0.047262318432331085, 0.17534415423870087, 0.03307317942380905, -0.13904969394207, -0.015682822093367577, -0.0402069091796875, -0.15603256225585938, -0.238995760679245, -0.09178274869918823, -0.1182505264878273, 0.16442428529262543, 0.0009358620154671371, 0.06651917099952698, 0.08258313685655594, -0.022042419761419296, 0.16447891294956207, -0.07379321753978729, -0.07578866183757782, -0.006978808436542749, 0.12375060468912125, -0.056660156697034836, -0.03080669604241848, -0.10566964000463486, -0.008295975625514984, 0.1151021271944046, 0.15304014086723328, 0.12214863300323486, 0.2957419455051422, 0.08268889784812927, 0.026645636186003685, 0.08958091586828232, 0.17622539401054382, 0.09495089203119278, 0.07838419824838638, -0.045413073152303696, -0.014814783819019794, 0.014317171648144722, -0.04022889584302902, 0.010141594335436821, 0.14683100581169128, -0.2679629921913147, -0.006678564939647913, -0.2710230350494385, 0.0965198427438736, -0.10913380235433578, 0.11837165057659149, -0.01015760749578476, 0.10194015502929688, 0.11082887649536133, 0.03233652561903, -0.03858073800802231, 0.16613617539405823, 0.08450309932231903, -0.11277695000171661, 0.001758623169735074, 0.03737903758883476, 0.09715615212917328, -0.02818971499800682, 0.12721189856529236, -0.11048974841833115, -0.1464834064245224, 0.013753619976341724, 0.07152791321277618, -0.15373679995536804, 0.3138748109340668, 0.012069208547472954, -0.13481520116329193, -0.01481647603213787, -0.09957809001207352, -0.006440147757530212, 0.1254177987575531, 0.09333524852991104, 0.07935678958892822, -0.2185502052307129, -0.13339371979236603, 0.05872276425361633, -0.00575496768578887, 0.22408108413219452, -0.034034017473459244, -0.11356475204229355, -0.027013886719942093, 0.04241163283586502, -0.06043251231312752, 0.08524788916110992, 0.023536119610071182, -0.08113526552915573, -0.032957352697849274, 0.05323701351881027, 0.012368366122245789, 0.00524376705288887, 0.09360801428556442, 0.020107939839363098, -0.0009265501867048442, 0.01785753294825554, 0.047885000705718994, -0.0675911232829094, -0.1984109878540039, 0.09357594698667526, -0.05215044692158699, 0.0015536568826064467, -0.08013670891523361, -0.15122665464878082, -0.08837161958217621, -0.16009655594825745, 0.12540200352668762, -0.034406669437885284, 0.12700119614601135, -0.06619787961244583, 0.17341409623622894, -0.07871770113706589, 0.04481020197272301, -0.047349292784929276, 0.050332702696323395, -0.007268077693879604, -0.07756082713603973, 0.16585899889469147, -0.15564003586769104, 0.01809087023139, 0.19572502374649048, -0.018915493041276932, 0.07177707552909851, 0.021322092041373253, -0.0636206790804863, 0.23147478699684143, 0.3014698624610901, 0.008138049393892288, 0.1665448248386383, 0.3018903136253357, -0.07466315478086472, -0.2642788887023926, -0.05505012720823288, -0.2841376066207886, -0.05371501296758652, 0.10716094076633453, -0.22523896396160126, 0.06986407935619354, 0.14383509755134583, -0.06471995264291763, 0.30228954553604126, -0.21825523674488068, 0.012589273042976856, 0.15434536337852478, -0.08868814259767532, 0.5515313148498535, -0.1133413165807724, -0.17677772045135498, -0.008122089318931103, -0.08741296827793121, 0.10602109134197235, -0.0340677872300148, 0.06877441704273224, 0.013465235009789467, 0.04797380417585373, 0.048932258039712906, -0.03111894056200981, 0.22701001167297363, 0.008710170164704323, 0.09015397727489471, -0.07378865778446198, -0.18624304234981537, 0.11639340221881866, -0.04359482601284981, -0.08891059458255768, 0.0849778801202774, -0.05942516401410103, -0.11078983545303345, 0.04663389176130295, -0.07950539886951447, -0.024862350896000862, 0.08423490077257156, -0.04678233340382576, -0.042606171220541, -0.008054176345467567, -0.1618063747882843, -0.0002289071271661669, 0.31360217928886414, -0.07096036523580551, 0.16695955395698547, 0.03677211329340935, 0.00038613268407061696, -0.11027684062719345, 0.030288029462099075, -0.05203165486454964, -0.021576624363660812, 0.09578979015350342, -0.11096979677677155, 0.03204701095819473, 0.14160704612731934, -0.04864364117383957, 0.05846960097551346, 0.09256096184253693, -0.0849417969584465, 0.007583672646433115, 0.17753590643405914, -0.17537221312522888, -0.1273445188999176, -0.006135711446404457, -0.09862716495990753, 0.14055661857128143, 0.04394126310944557, 0.05191568285226822, 0.16669964790344238, 0.03967129811644554, -0.029474308714270592, -0.02817419543862343, -0.1153380498290062, -0.0201893113553524, 0.040153320878744125, 0.00045633706031367183, -0.08791285753250122, 0.2262638509273529, 0.06409153342247009, -0.1328488290309906, -0.051157206296920776, 0.2161225974559784, -0.06805316358804703, -0.04911920800805092, -0.223562553524971, 0.10752306133508682, -0.07112517952919006, -0.0965060144662857, 0.05453834682703018, -0.02270081453025341, 0.005106312222778797, 0.181985542178154, 0.03941008821129799, 0.11070270836353302, 0.03738937899470329, -0.02448922023177147, 0.15798696875572205, -0.142850860953331, -0.14191335439682007, -0.025354057550430298, -0.08757315576076508, -0.13844476640224457, -0.026804137974977493, 0.1617041826248169, -0.09177309274673462, -0.14772607386112213, -0.2621181011199951, 0.10968475043773651, -0.16432365775108337, -0.10192688554525375, -0.03469514101743698, -0.08968492597341537, 0.0696166530251503, 0.030301768332719803, -0.03093348816037178, -0.06706760823726654, -0.18593791127204895, 0.0816768929362297, 0.06349513679742813, 0.045533183962106705, -0.017847947776317596, 0.0067379772663116455, 0.1720137596130371, 0.025955144315958023, 0.10040043294429779, 0.16762186586856842, 0.011397695168852806, 0.2246655523777008, -0.1671202927827835, -0.11496317386627197, 0.1336962729692459, -0.026543032377958298, 0.06762003898620605, 0.16792191565036774, -0.0772583931684494, 0.015526676550507545, -0.028136352077126503, 0.07066910713911057, -0.11003983020782471, -0.105624258518219, 0.007937257178127766, 0.02567129209637642, -0.2755882740020752, -0.005599735304713249, -0.19717298448085785, 0.14788752794265747, 0.02579621411859989, 0.03297143429517746, 0.10257530212402344, 0.10404334217309952, 0.08312062919139862, -0.0017710148822516203, 0.03226327523589134, -0.1176818460226059, 0.02753005363047123, -0.059239376336336136, -0.020663779228925705, 0.017624232918024063, 0.36952024698257446, -0.03603357449173927, -0.046802736818790436, 0.003710439894348383, 0.1307835876941681, -0.02139742486178875, 0.017395347356796265, 0.13209912180900574, 0.12607666850090027, -0.08595693111419678, -0.1504845917224884, 0.04888554662466049, -0.04565655067563057, -0.02836887165904045, 0.1464131623506546, 0.05905961990356445, 0.1050296202301979, 0.0908031314611435, -0.014463032595813274, -0.00318976235575974, 0.012856799177825451, -0.15486004948616028, 0.06223496049642563, -0.010558074340224266, 0.012565906159579754, 0.017934376373887062, 0.15238402783870697, -0.005540105979889631, 0.07739730179309845, -0.09889880567789078, 0.004208535887300968, -0.13498884439468384, -0.07913459837436676, 0.03617347031831741, -0.13393273949623108, 0.04141177982091904, -0.01871878281235695, 0.029611799865961075, 0.30386561155319214, 0.02558239921927452, -0.020639164373278618, 0.12512871623039246, -0.1214587539434433, -0.12050267308950424, -0.001594188273884356, -0.029960084706544876, 0.0791488066315651, -0.02633434161543846, -0.0997740775346756, -0.1001306027173996, -0.15166029334068298, -0.09759195148944855, 0.05182836204767227, -0.04993441700935364, -0.059362251311540604, -0.17634081840515137, -0.05707859992980957, -0.05147340148687363, 0.14025864005088806, -0.12263951450586319, 0.15159130096435547, -0.014490418136119843, 0.004084470681846142, 0.04405883327126503, 0.1950942426919937, -0.03644494712352753, 0.08714226633310318, 0.0154351145029068, 0.1522706001996994, -0.05119588226079941, 0.14720745384693146, -0.10931728035211563, -0.04014137014746666, -0.06710435450077057, 0.21513493359088898, 0.25630924105644226, -0.06136954948306084, -0.008937356993556023, -0.012760217301547527, 0.058654606342315674, 0.1073930487036705, 0.16049085557460785, 0.002326392102986574, 0.2802925705909729, -0.03133585304021835, 0.04815128445625305, 0.02901598811149597, 0.013607407920062542, -0.06336209923028946, 0.03397751972079277, 0.07539387792348862, -0.035039983689785004, -0.1412304788827896, 0.15837742388248444, -0.21980468928813934, 0.18157227337360382, 0.11640069633722305, -0.19996967911720276, -0.013728445395827293, -0.04882071167230606, 0.1689416468143463, -0.0856364443898201, 0.1637246012687683, -0.0903693437576294, -0.2108195722103119, -0.2056000679731369, 0.03867346793413162, -0.34623071551322937, -0.254462867975235, 0.10422009229660034, 0.1488201916217804, 0.04015883058309555, -0.018507536500692368, -0.019967829808592796, -0.018367022275924683, 0.04877542704343796, -0.0067357709631323814, 0.06014643982052803, 0.031397558748722076, -0.02988368645310402, -0.24127542972564697, -0.029804671183228493, 0.023964406922459602, -0.07093082368373871, 0.07464958727359772, -0.06874357163906097, -0.022495782002806664, 0.08059766888618469, -0.03066304884850979, 0.03298592567443848, -0.035373736172914505, -0.16326889395713806, 0.027529051527380943, 0.03900543600320816, 0.036012712866067886, 0.00634160777553916, 0.0008072225609794259, -0.03455270454287529, 0.0644603744149208, -0.16716794669628143, -0.16015739738941193, 0.14140215516090393, -0.06745140254497528, 0.2779497504234314, -0.05812826007604599, -0.0809100940823555, 0.04766704887151718, -0.03426874056458473, 0.1807648241519928, -0.07756473124027252, 0.047254521399736404, 0.12766779959201813, 0.011127962730824947, 0.03121316432952881, -0.3092964291572571, 0.11082969605922699, -0.000795336440205574, -0.006093299947679043, -0.07581598311662674 ]
null
null
transformers
# StructBERT: Un-Official Copy Official Repository Link: https://github.com/alibaba/AliceMind/tree/main/StructBERT **Claimer** * This model card is not produced by [AliceMind Team](https://github.com/alibaba/AliceMind/) ## Reproduce HFHub models: Download model/tokenizer vocab ```bash wget https://raw.githubusercontent.com/alibaba/AliceMind/main/StructBERT/config/large_bert_config.json && mv large_bert_config.json config.json wget https://raw.githubusercontent.com/alibaba/AliceMind/main/StructBERT/config/vocab.txt wget https://alice-open.oss-cn-zhangjiakou.aliyuncs.com/StructBERT/en_model && mv en_model pytorch_model.bin ``` ```python from transformers import AutoConfig, AutoModelForMaskedLM, AutoTokenizer config = AutoConfig.from_pretrained("./config.json") model = AutoModelForMaskedLM.from_pretrained(".", config=config) tokenizer = AutoTokenizer.from_pretrained(".", config=config) model.push_to_hub("structbert-large") tokenizer.push_to_hub("structbert-large") ``` [https://arxiv.org/abs/1908.04577](https://arxiv.org/abs/1908.04577) # StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding ## Introduction We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. ## Pre-trained models |Model | Description | #params | Download | |------------------------|-------------------------------------------|------|------| |structbert.en.large | StructBERT using the BERT-large architecture | 340M | [structbert.en.large](https://alice-open.oss-cn-zhangjiakou.aliyuncs.com/StructBERT/en_model) | |structroberta.en.large | StructRoBERTa continue training from RoBERTa | 355M | Coming soon | |structbert.ch.large | Chinese StructBERT; BERT-large architecture | 330M | [structbert.ch.large](https://alice-open.oss-cn-zhangjiakou.aliyuncs.com/StructBERT/ch_model) | ## Results The results of GLUE & CLUE tasks can be reproduced using the hyperparameters listed in the following "Example usage" section. #### structbert.en.large [GLUE benchmark](https://gluebenchmark.com/leaderboard) |Model| MNLI | QNLIv2 | QQP | SST-2 | MRPC | |--------------------|-------|-------|-------|-------|-------| |structbert.en.large |86.86% |93.04% |91.67% |93.23% |86.51% | #### structbert.ch.large [CLUE benchmark](https://www.cluebenchmarks.com/) |Model | CMNLI | OCNLI | TNEWS | AFQMC | |--------------------|-------|-------|-------|-------| |structbert.ch.large |84.47% |81.28% |68.67% |76.11% | ## Example usage #### Requirements and Installation * [PyTorch](https://pytorch.org/) version >= 1.0.1 * Install other libraries via ``` pip install -r requirements.txt ``` * For faster training install NVIDIA's [apex](https://github.com/NVIDIA/apex) library #### Finetune MNLI ``` python run_classifier_multi_task.py \ --task_name MNLI \ --do_train \ --do_eval \ --do_test \ --amp_type O1 \ --lr_decay_factor 1 \ --dropout 0.1 \ --do_lower_case \ --detach_index -1 \ --core_encoder bert \ --data_dir path_to_glue_data \ --vocab_file config/vocab.txt \ --bert_config_file config/large_bert_config.json \ --init_checkpoint path_to_pretrained_model \ --max_seq_length 128 \ --train_batch_size 32 \ --learning_rate 2e-5 \ --num_train_epochs 3 \ --fast_train \ --gradient_accumulation_steps 1 \ --output_dir path_to_output_dir ``` ## Citation If you use our work, please cite: ``` @article{wang2019structbert, title={Structbert: Incorporating language structures into pre-training for deep language understanding}, author={Wang, Wei and Bi, Bin and Yan, Ming and Wu, Chen and Bao, Zuyi and Xia, Jiangnan and Peng, Liwei and Si, Luo}, journal={arXiv preprint arXiv:1908.04577}, year={2019} } ```
{}
fill-mask
bayartsogt/structbert-large
[ "transformers", "pytorch", "bert", "fill-mask", "arxiv:1908.04577", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1908.04577" ]
[]
TAGS #transformers #pytorch #bert #fill-mask #arxiv-1908.04577 #autotrain_compatible #endpoints_compatible #region-us
StructBERT: Un-Official Copy ============================ Official Repository Link: URL Claimer * This model card is not produced by AliceMind Team Reproduce HFHub models: ----------------------- Download model/tokenizer vocab URL StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding =============================================================================================== Introduction ------------ We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. Pre-trained models ------------------ Results ------- The results of GLUE & CLUE tasks can be reproduced using the hyperparameters listed in the following "Example usage" section. #### URL GLUE benchmark #### URL CLUE benchmark Example usage ------------- #### Requirements and Installation * PyTorch version >= 1.0.1 * Install other libraries via * For faster training install NVIDIA's apex library #### Finetune MNLI If you use our work, please cite:
[ "#### URL\n\n\nGLUE benchmark", "#### URL\n\n\nCLUE benchmark\n\n\n\nExample usage\n-------------", "#### Requirements and Installation\n\n\n* PyTorch version >= 1.0.1\n* Install other libraries via\n* For faster training install NVIDIA's apex library", "#### Finetune MNLI\n\n\nIf you use our work, please cite:" ]
[ "TAGS\n#transformers #pytorch #bert #fill-mask #arxiv-1908.04577 #autotrain_compatible #endpoints_compatible #region-us \n", "#### URL\n\n\nGLUE benchmark", "#### URL\n\n\nCLUE benchmark\n\n\n\nExample usage\n-------------", "#### Requirements and Installation\n\n\n* PyTorch version >= 1.0.1\n* Install other libraries via\n* For faster training install NVIDIA's apex library", "#### Finetune MNLI\n\n\nIf you use our work, please cite:" ]
[ 44, 6, 11, 36, 16 ]
[ "passage: TAGS\n#transformers #pytorch #bert #fill-mask #arxiv-1908.04577 #autotrain_compatible #endpoints_compatible #region-us \n#### URL\n\n\nGLUE benchmark#### URL\n\n\nCLUE benchmark\n\n\n\nExample usage\n-------------#### Requirements and Installation\n\n\n* PyTorch version >= 1.0.1\n* Install other libraries via\n* For faster training install NVIDIA's apex library#### Finetune MNLI\n\n\nIf you use our work, please cite:" ]
[ -0.042748548090457916, 0.15704379975795746, -0.0016536741750314832, 0.12826275825500488, 0.12493099272251129, 0.07056711614131927, 0.0520964153110981, 0.1499430537223816, 0.08936437964439392, -0.0028121881186962128, 0.19695255160331726, 0.22007960081100464, 0.04100190848112106, 0.14242789149284363, 0.038566019386053085, -0.1538761556148529, -0.005891074892133474, 0.07803242653608322, 0.05205497518181801, 0.13640868663787842, 0.07559634745121002, -0.12120332568883896, 0.09440559893846512, -0.01117134839296341, -0.15790508687496185, -0.020782742649316788, 0.03475704789161682, -0.053614385426044464, 0.12728610634803772, 0.009786674752831459, 0.09179902076721191, -0.0018737622303888202, 0.045201774686574936, -0.1462777853012085, 0.027137869969010353, 0.03245522454380989, -0.017797084525227547, 0.07006696611642838, 0.057949893176555634, -0.017671311274170876, 0.0014896491775289178, 0.02956041879951954, -0.004116090480238199, 0.038319025188684464, -0.0632425919175148, -0.18776102364063263, -0.03302712365984917, -0.027358192950487137, -0.06578435748815536, 0.034693602472543716, 0.011895322240889072, 0.23104111850261688, -0.10563389956951141, 0.06391748040914536, 0.2246338278055191, -0.29658815264701843, -0.09005293995141983, 0.07223910838365555, 0.062040213495492935, -0.01729678176343441, -0.05711570009589195, 0.07151695340871811, 0.025131618604063988, 0.05904674157500267, 0.16468220949172974, -0.02265992760658264, -0.0780683383345604, 0.028322329744696617, -0.10188417881727219, -0.03271077573299408, 0.05203746259212494, -0.003359000664204359, 0.0305049829185009, -0.07819285243749619, -0.13502539694309235, -0.06801947206258774, -0.03059060126543045, 0.01732068508863449, -0.007741529028862715, -0.035914625972509384, -0.019768735393881798, -0.060422662645578384, -0.05011290684342384, -0.11691335588693619, -0.02660641074180603, 0.15456704795360565, 0.06642791628837585, 0.060589008033275604, 0.09313438832759857, 0.14659713208675385, -0.007261873688548803, -0.12057799100875854, 0.005466611124575138, -0.12494829297065735, -0.01008719950914383, 0.07188461720943451, -0.01577567309141159, 0.009917457588016987, 0.06131758540868759, 0.16317816078662872, 0.1276572346687317, 0.030098313465714455, 0.07196930050849915, 0.05731312930583954, -0.12127310782670975, 0.06851259618997574, -0.07050394266843796, -0.09514825791120529, 0.0461159311234951, 0.09507890045642853, 0.07369116693735123, -0.007252849172800779, -0.08227279782295227, -0.10006514936685562, 0.012893709354102612, 0.038961123675107956, 0.04544535651803017, 0.040945880115032196, -0.0970662459731102, -0.03865578770637512, 0.07438009977340698, -0.07955868542194366, 0.0208075400441885, 0.0052663483656942844, -0.07585041970014572, -0.10142962634563446, 0.05030788481235504, -0.04322274029254913, -0.0465068519115448, -0.019504673779010773, -0.03347424417734146, -0.018344661220908165, -0.1860715001821518, -0.032872188836336136, 0.03623250499367714, -0.0665804073214531, 0.02876698598265648, -0.16747142374515533, -0.17596904933452606, 0.07454352080821991, 0.11082626134157181, -0.03289850428700447, -0.050661537796258926, -0.0034166621044278145, -0.016803104430437088, 0.05437608063220978, -0.06351485848426819, 0.1869240403175354, -0.06944602727890015, 0.12856806814670563, 0.104265958070755, 0.07845199853181839, -0.022376585751771927, 0.006589815020561218, -0.060295943170785904, -0.0016095489263534546, -0.05595232918858528, 0.009266496635973454, -0.08308421820402145, -0.02143852412700653, -0.05163343623280525, -0.07992731034755707, -0.11726365238428116, -0.0145428366959095, 0.08068142086267471, 0.11084055155515671, -0.09957502037286758, -0.015091919340193272, 0.12968583405017853, -0.019296061247587204, -0.1301836371421814, 0.17801284790039062, -0.03665792569518089, -0.07139622420072556, 0.020769409835338593, 0.10586181282997131, 0.0690818801522255, -0.21923038363456726, -0.004155924078077078, 0.08729520440101624, -0.11020459979772568, -0.12415866553783417, 0.12732382118701935, 0.10895601660013199, -0.0013413208071142435, 0.07417819648981094, -0.047939151525497437, 0.13585951924324036, -0.04880255088210106, -0.08303356915712357, -0.08322575688362122, -0.10881929099559784, 0.0921844020485878, 0.08544027805328369, 0.008510863408446312, -0.06716182827949524, -0.14086921513080597, -0.11610440909862518, 0.0764707401394844, -0.05212290585041046, -0.006624678149819374, -0.09099967032670975, 0.15719188749790192, -0.08190516382455826, 0.006500082556158304, -0.10604765266180038, -0.05434638634324074, 0.06352538615465164, -0.04754875972867012, -0.021515710279345512, -0.01965070329606533, 0.07357364892959595, 0.04656184837222099, -0.044846631586551666, 0.006639257539063692, -0.057799700647592545, -0.056957826018333435, -0.013020825572311878, -0.027520576491951942, -0.03666037321090698, -0.020039886236190796, 0.0071490746922791, -0.1411554217338562, 0.04231938347220421, -0.041230302304029465, 0.09992845356464386, -0.05330633744597435, -0.01847670041024685, -0.013214249163866043, 0.0005906174774281681, -0.031092055141925812, -0.060643214732408524, 0.11063123494386673, 0.016638578847050667, -0.03580145165324211, 0.06903954595327377, -0.030907873064279556, 0.212127223610878, 0.18159380555152893, -0.1572970300912857, 0.009585755877196789, 0.05756201967597008, -0.061654988676309586, -0.03459484875202179, -0.021641109138727188, 0.02126675471663475, 0.10534593462944031, 0.030645377933979034, 0.13610343635082245, -0.08460257947444916, -0.02434420958161354, 0.045765381306409836, 0.03269355371594429, 0.01701420359313488, 0.07253775745630264, 0.18530666828155518, -0.03513333201408386, 0.09754838794469833, 0.09991847723722458, -0.09070936590433121, 0.12478258460760117, 0.030563917011022568, -0.04763041436672211, 0.02195223607122898, -0.01812431588768959, -0.02264256216585636, 0.1749686449766159, -0.05757404863834381, 0.018606916069984436, 0.08725675195455551, -0.10356439650058746, 0.055766887962818146, -0.179326131939888, -0.034823015332221985, -0.004292489029467106, -0.02771058678627014, -0.034816354513168335, 0.03741271421313286, -0.0026043823454529047, 0.043456073850393295, 0.02223014645278454, -0.03075408563017845, 0.05614970624446869, 0.002416439587250352, -0.038905948400497437, 0.14363962411880493, -0.07566342502832413, -0.3014009892940521, -0.1310696005821228, -0.05618058145046234, -0.01179652102291584, 0.03055041842162609, 0.017170840874314308, -0.03659895807504654, -0.015098132193088531, 0.05649556592106819, 0.01922837272286415, 0.0013635765062645078, -0.03705567866563797, -0.058399710804224014, 0.04596773535013199, 0.005886411760002375, -0.1272595226764679, 0.01160760410130024, 0.020179567858576775, 0.0010726876789703965, 0.08158879727125168, -0.004363854415714741, 0.13763564825057983, 0.03675520420074463, 0.00045070095802657306, -0.002663121558725834, -0.0016422861954197288, 0.18705524504184723, -0.10294298082590103, -0.04342101141810417, 0.16731928288936615, 0.023267284035682678, 0.038386110216379166, 0.022399723529815674, 0.0650549903512001, -0.058826737105846405, -0.015586660243570805, -0.008028721436858177, -0.04694202169775963, -0.16034282743930817, -0.03068486973643303, -0.09538833051919937, -0.036201246082782745, 0.11083376407623291, 0.03502541035413742, -0.06841466575860977, 0.10979185253381729, 0.016792725771665573, -0.013504495844244957, -0.019315768033266068, 0.07175622135400772, 0.15259557962417603, -0.008101298473775387, 0.0923398956656456, -0.03892117366194725, -0.0381292849779129, 0.031827960163354874, 0.11693290621042252, 0.12257378548383713, -0.07823079824447632, 0.15422458946704865, 0.03648664057254791, 0.18335886299610138, 0.09150388836860657, 0.06847188621759415, 0.0468016155064106, -0.020705370232462883, -0.02007564716041088, -0.033030204474925995, -0.13992241024971008, 0.02778431959450245, 0.0638691708445549, 0.0358787402510643, 0.022548973560333252, -0.027303418144583702, 0.08971403539180756, 0.19146892428398132, -0.04779357835650444, -0.3735569417476654, -0.026033250615000725, -0.023781491443514824, -0.05373930186033249, -0.10842689871788025, 0.037797585129737854, 0.03839128091931343, -0.09641186147928238, 0.04164458438754082, -0.11627926677465439, 0.09014514833688736, 0.007820679806172848, -0.01804531365633011, 0.05590245500206947, 0.12392077594995499, 0.06300141662359238, 0.13016974925994873, -0.27594542503356934, 0.1695019155740738, 0.038682833313941956, 0.11067695915699005, -0.07184236496686935, -0.006806612946093082, 0.05279745161533356, 0.003458865452557802, 0.14074429869651794, -0.04957285523414612, 0.11235766857862473, -0.05101070925593376, -0.1225910559296608, 0.06333622336387634, 0.03284414857625961, 0.037569522857666016, 0.02698625810444355, -0.027461456134915352, -0.022955769672989845, -0.010395853780210018, -0.02172059193253517, -0.10591524094343185, -0.133466437458992, 0.009927131235599518, 0.002407298656180501, -0.0320456400513649, -0.06509900838136673, -0.09203142672777176, -0.10514549165964127, 0.14084307849407196, -0.11521700769662857, -0.07903041690587997, -0.07177190482616425, 0.06037118285894394, 0.11320247501134872, -0.04555944725871086, 0.08278127759695053, -0.07514572888612747, 0.08936848491430283, -0.059270892292261124, -0.1072198748588562, 0.001474378746934235, -0.06863818317651749, -0.0879257470369339, -0.04291842132806778, 0.09606441110372543, -0.05628542974591255, 0.0902305468916893, -0.016324251890182495, 0.023660486564040184, -0.13252560794353485, -0.05104932188987732, 0.026460733264684677, -0.006446254905313253, 0.08087196201086044, -0.05307026952505112, -0.07791499048471451, 0.12472938001155853, -0.06144273653626442, -0.0009761254186742008, 0.16235876083374023, 0.20596076548099518, -0.10809525102376938, 0.057572271674871445, 0.10307783633470535, -0.03144017979502678, -0.2653043568134308, -0.07930801063776016, 0.032377827912569046, 0.004343667533248663, -0.06957276165485382, -0.18532094359397888, 0.031549859791994095, 0.044542599469423294, -0.04405452683568001, 0.18686451017856598, -0.2053992599248886, -0.09903771430253983, 0.0486438050866127, 0.12845492362976074, 0.08398732542991638, -0.11189556866884232, 0.008422697894275188, -0.00813661701977253, -0.21246542036533356, 0.14824195206165314, 0.03594498708844185, 0.12559249997138977, -0.07138822227716446, 0.002394743962213397, -0.014639069326221943, -0.11076093465089798, 0.10305316746234894, -0.1655365377664566, 0.03558496758341789, -0.015230362303555012, 0.049211692065000534, 0.047646161168813705, -0.018716176971793175, 0.024403639137744904, -0.07109106332063675, -0.000989418593235314, -0.06511074304580688, -0.07636583596467972, -0.08258529752492905, 0.09735508263111115, 0.009654670022428036, -0.06508073210716248, -0.022259969264268875, 0.024063410237431526, 0.005451173521578312, -0.00814161915332079, 0.04666735231876373, 0.011320861056447029, 0.008900335058569908, 0.08150922507047653, 0.061538733541965485, -0.1180589571595192, -0.12759310007095337, -0.028751444071531296, -0.0247309859842062, 0.07748255878686905, -0.0902114063501358, -0.022983234375715256, 0.137526735663414, 0.008598677814006805, -0.017230231314897537, 0.08018721640110016, -0.04246985912322998, 0.016982099041342735, 0.08783116936683655, -0.1874983310699463, -0.06736669689416885, -0.027941666543483734, -0.13377687335014343, -0.009868817403912544, -0.11224284023046494, 0.014128660783171654, -0.07462847977876663, -0.04473218321800232, 0.03149206563830376, 0.005128031130880117, -0.039335090667009354, 0.1246083676815033, 0.12379402667284012, 0.06706029176712036, -0.10037639737129211, 0.0990087017416954, -0.03743039071559906, -0.13686737418174744, -0.04561571776866913, 0.09027344733476639, -0.08598750084638596, -0.09367962926626205, -0.012806637212634087, 0.04659118503332138, 0.015051773749291897, -0.0752052441239357, -0.15399476885795593, 0.00441052857786417, 0.03242791071534157, 0.019404945895075798, 0.1238541454076767, 0.0013660918921232224, -0.0034731540363281965, -0.006215306930243969, -0.07375949621200562, 0.12223462760448456, -0.028323248028755188, 0.04310034215450287, -0.14518305659294128, 0.001986090559512377, 0.05550539866089821, 0.1163870170712471, -0.07713315635919571, 0.04960213601589203, -0.053427405655384064, 0.07424528896808624, -0.016412846744060516, 0.03381817415356636, -0.022484416142106056, 0.01115473359823227, -0.05000627785921097, -0.026113029569387436, -0.09239384531974792, 0.020494505763053894, -0.09611894190311432, -0.018949061632156372, 0.027464615181088448, -0.015988443046808243, -0.05449596792459488, -0.04787639528512955, 0.056283850222826004, -0.04694550111889839, 0.013500352390110493, 0.05641896650195122, -0.03998526930809021, 0.05536974221467972, -0.13282129168510437, -0.08907701075077057, 0.01203092373907566, 0.09438668936491013, -0.027304016053676605, -0.026417605578899384, 0.06741771101951599, 0.0024790007155388594, 0.030311809852719307, -0.027405209839344025, 0.1672569364309311, -0.13537098467350006, -0.025540823116898537, 0.02008364163339138, -0.07096372544765472, -0.02181864157319069, 0.023417523130774498, 0.06886545568704605, 0.07915135473012924, 0.12696313858032227, -0.05446258932352066, 0.03896855190396309, -0.13358727097511292, 0.04010559245944023, -0.03844548761844635, -0.1225276067852974, -0.05775781348347664, -0.018989788368344307, 0.07474866509437561, -0.0030322689563035965, 0.2776436507701874, 0.012613755650818348, -0.02472243644297123, -0.012852124869823456, -0.07547187060117722, -0.039622120559215546, -0.008356704376637936, 0.1874241828918457, 0.03122362308204174, 0.03143414109945297, 0.04304856061935425, 0.07809016108512878, 0.11052553355693817, 0.06884313374757767, 0.052697427570819855, 0.2690947353839874, 0.15359722077846527, 0.08937375247478485, 0.02280518040060997, -0.07653898745775223, -0.13780158758163452, 0.029982782900333405, -0.03822847828269005, 0.11261090636253357, -0.07190044969320297, 0.1561288982629776, 0.069535031914711, -0.048125527799129486, 0.026368185877799988, -0.04171144217252731, -0.04700123891234398, -0.12206406891345978, -0.0020907826256006956, -0.07664874941110611, -0.09859194606542587, 0.07280801236629486, -0.021789895370602608, -0.040648769587278366, 0.07990691810846329, 0.015786129981279373, -0.0685109943151474, 0.13209670782089233, 0.11363846808671951, -0.013274909928441048, 0.03497115895152092, 0.036233045160770416, -0.03630516678094864, -0.05032290518283844, -0.06445807963609695, -0.03695393353700638, -0.04171011596918106, 0.06882888078689575, -0.025780562311410904, -0.08484964817762375, 0.030489526689052582, -0.004476465750485659, -0.047725822776556015, -0.028343455865979195, 0.04915992170572281, 0.01802828349173069, 0.09934857487678528, -0.031636711210012436, 0.007471034303307533, -0.02327844873070717, 0.16481226682662964, -0.003808327717706561, -0.0983949825167656, -0.12191920727491379, 0.21676084399223328, -0.035243336111307144, -0.01152689941227436, 0.0038605823647230864, -0.0020382683724164963, -0.039284270256757736, 0.30933380126953125, 0.20809082686901093, -0.0821032002568245, -0.01728130131959915, 0.09024810791015625, -0.005431005265563726, -0.02497003972530365, 0.1674962192773819, 0.08074519038200378, 0.16602621972560883, -0.12789595127105713, -0.06898684054613113, -0.05089998245239258, -0.0105774300172925, -0.08517691493034363, 0.04877804219722748, -0.00385097274556756, 0.01249010767787695, -0.1002989411354065, -0.0058294422924518585, -0.02377750724554062, 0.00780673511326313, 0.037862349301576614, -0.13992939889431, -0.08678609132766724, -0.03935588151216507, -0.004090951755642891, -0.10690601915121078, 0.038340047001838684, -0.06527227908372879, 0.02941594086587429, -0.020212603732943535, -0.005236804485321045, -0.15525425970554352, 0.00737590529024601, 0.07555711269378662, 0.02373356930911541, 0.27154451608657837, -0.06415589153766632, 0.11133415997028351, 0.14070972800254822, -0.02746160328388214, -0.134812593460083, 0.04655778408050537, 0.05147106945514679, -0.0743299275636673, 0.03774665296077728, 0.04106760397553444, 0.009641822427511215, -0.021777348592877388, -0.020176494494080544, 0.05965472757816315, 0.04135438799858093, -0.0937320813536644, -0.041904646903276443, -0.07828488945960999, 0.028447778895497322, -0.06320041418075562, 0.09617099910974503, 0.11537735909223557, 0.011505123227834702, -0.047651637345552444, -0.13447003066539764, 0.08104897290468216, 0.03924688696861267, 0.05505729466676712, -0.055061135441064835, -0.17543582618236542, -0.030791625380516052, 0.07034575939178467, -0.04822622612118721, -0.20702335238456726, -0.06692364066839218, -0.0428517647087574, -0.04504914954304695, -0.04904906451702118, 0.0312383733689785, 0.05185173079371452, 0.04013728350400925, -0.022176047787070274, -0.21086210012435913, -0.04346868395805359, 0.02938404306769371, -0.16225381195545197, -0.10356950014829636 ]
null
null
fairseq
# tts_transformer-mn-mbspeech [Transformer](https://arxiv.org/abs/1809.08895) text-to-speech model from fairseq S^2 ([paper](https://arxiv.org/abs/2109.06912)/[code](https://github.com/pytorch/fairseq/tree/main/examples/speech_synthesis)): - Mongolian - Single-speaker male voice - Trained on [MBSpeech](https://github.com/tugstugi/mongolian-nlp/blob/master/datasets/MBSpeech-1.0-csv.zip)
{"language": "mn", "library_name": "fairseq", "tags": ["fairseq", "audio", "text-to-speech"], "datasets": ["mbspeech"], "task": "text-to-speech", "widget": [{"text": "\u043c\u0438\u043d\u0438\u0439 \u043d\u044d\u0440\u0438\u0439\u0433 \u0431\u0430\u044f\u0440\u0446\u043e\u0433\u0442 \u0433\u044d\u0434\u044d\u0433", "example_title": "Say my name!"}, {"text": "\u0431\u0438 \u043c\u043e\u043d\u0433\u043e\u043b \u0443\u043b\u0441\u044b\u043d \u043d\u0438\u0439\u0441\u043b\u044d\u043b, \u0443\u043b\u0430\u0430\u043d\u0431\u0430\u0430\u0442\u0430\u0440 \u0445\u043e\u0442\u043e\u0434 \u0430\u043c\u044c\u0434\u0430\u0440\u0434\u0430\u0433", "example_title": "Where I am from?"}, {"text": "\u044d\u043d\u044d\u0445\u04af\u04af \u04e9\u0433\u04e9\u0433\u0434\u043b\u0438\u0439\u0433 \u043d\u044d\u044d\u043b\u0442\u0442\u044d\u0439 \u0431\u043e\u043b\u0433\u043e\u0441\u043e\u043d, \u0431\u043e\u043b\u043e\u0440 \u0441\u043e\u043e\u0444\u0442\u044b\u043d\u0445\u043e\u043d\u0434 \u0431\u0430\u044f\u0440\u043b\u0430\u043b\u0430\u0430", "example_title": "Thank you!"}, {"text": "\u044d\u043d\u044d\u0445\u04af\u04af \u0430\u0436\u043b\u044b\u043d \u0438\u0445\u044d\u043d\u0445 \u0445\u044d\u0441\u0433\u0438\u0439\u0433, \u0442\u04e9\u0433\u04e9\u043b\u0434\u04e9\u0440 \u0430\u0445 \u0445\u0438\u0439\u0441\u044d\u043d \u0431\u043e\u043b\u043d\u043e", "example_title": "Shout out to original creater"}]}
text-to-speech
bayartsogt/tts_transformer-mn-mbspeech
[ "fairseq", "audio", "text-to-speech", "mn", "dataset:mbspeech", "arxiv:1809.08895", "arxiv:2109.06912", "region:us" ]
2022-03-02T23:29:05+00:00
[ "1809.08895", "2109.06912" ]
[ "mn" ]
TAGS #fairseq #audio #text-to-speech #mn #dataset-mbspeech #arxiv-1809.08895 #arxiv-2109.06912 #region-us
# tts_transformer-mn-mbspeech Transformer text-to-speech model from fairseq S^2 (paper/code): - Mongolian - Single-speaker male voice - Trained on MBSpeech
[ "# tts_transformer-mn-mbspeech\nTransformer text-to-speech model from fairseq S^2 (paper/code):\n- Mongolian\n- Single-speaker male voice\n- Trained on MBSpeech" ]
[ "TAGS\n#fairseq #audio #text-to-speech #mn #dataset-mbspeech #arxiv-1809.08895 #arxiv-2109.06912 #region-us \n", "# tts_transformer-mn-mbspeech\nTransformer text-to-speech model from fairseq S^2 (paper/code):\n- Mongolian\n- Single-speaker male voice\n- Trained on MBSpeech" ]
[ 46, 52 ]
[ "passage: TAGS\n#fairseq #audio #text-to-speech #mn #dataset-mbspeech #arxiv-1809.08895 #arxiv-2109.06912 #region-us \n# tts_transformer-mn-mbspeech\nTransformer text-to-speech model from fairseq S^2 (paper/code):\n- Mongolian\n- Single-speaker male voice\n- Trained on MBSpeech" ]
[ -0.04972884804010391, 0.06193527206778526, -0.003819467034190893, -0.05382200703024864, 0.11825018376111984, -0.09411830455064774, 0.2016710340976715, 0.05148587375879288, -0.03506341949105263, 0.0046702418476343155, 0.0278953704982996, 0.09629633277654648, 0.05082865431904793, 0.057231828570365906, 0.014302350580692291, -0.254777193069458, 0.040537286549806595, -0.0025233232881873846, 0.0016678852261975408, 0.05492859706282616, 0.14370176196098328, -0.015791842713952065, -0.007515421137213707, 0.010966670699417591, -0.08154036849737167, 0.06793708354234695, 0.022105971351265907, -0.0918654277920723, 0.001709420932456851, 0.05335196480154991, -0.05989731848239899, 0.08700774610042572, 0.006355166435241699, -0.07748087495565414, 0.032199639827013016, -0.09649588912725449, 0.025006061419844627, 0.012603136710822582, -0.003970062825828791, -0.003675432177260518, 0.11560894548892975, 0.016696590930223465, -0.052813928574323654, 0.06342987716197968, -0.05408032611012459, -0.09166984260082245, 0.05415353551506996, 0.03152607008814812, 0.06689371168613434, 0.08990049362182617, -0.07506375014781952, 0.08046569675207138, -0.05913286283612251, 0.09330053627490997, 0.11736276745796204, -0.2510312497615814, 0.013632087036967278, -0.028387079015374184, 0.02946842461824417, 0.10241863131523132, -0.10984839498996735, -0.00030421101837418973, 0.01802179403603077, -0.02999652922153473, -0.11681828647851944, -0.15568652749061584, -0.10741570591926575, -0.010664517991244793, -0.09995962679386139, 0.11252336949110031, 0.3588670492172241, 0.0586298368871212, -0.02519720420241356, -0.01910298317670822, -0.02257734164595604, 0.04830503836274147, 0.012709510512650013, 0.02331576682627201, -0.04978887364268303, 0.01568256877362728, -0.121611088514328, -0.07885324209928513, -0.11760792881250381, -0.01504878792911768, -0.13559602200984955, 0.1903931349515915, -0.051911577582359314, -0.0308686550706625, -0.05671088024973869, -0.04711490496993065, -0.043556198477745056, -0.04682079330086708, 0.034680090844631195, -0.036861907690763474, -0.055002059787511826, 0.024482756853103638, 0.04740385338664055, -0.3029785454273224, 0.10948698967695236, -0.06792810559272766, -0.026490574702620506, 0.09140941500663757, -0.06458782404661179, 0.05073559656739235, 0.016838740557432175, 0.010197234340012074, -0.02285272441804409, -0.04808242619037628, 0.004239040892571211, -0.002798778470605612, 0.0399942547082901, -0.04931878671050072, -0.13888467848300934, -0.07460165023803711, -0.08325047791004181, 0.07244864851236343, -0.032063618302345276, 0.062238823622465134, 0.06305521726608276, -0.04620718955993652, 0.11053848266601562, -0.07531827688217163, -0.09318409860134125, 0.010952945798635483, -0.027589157223701477, 0.05651313066482544, -0.022266633808612823, 0.050465770065784454, -0.004693657159805298, -0.09251274168491364, 0.005743480287492275, 0.07470003515481949, 0.03189501911401749, -0.11264072358608246, 0.06489621102809906, 0.08134757727384567, 0.02344495616853237, -0.16086852550506592, -0.08181697130203247, -0.022586682811379433, -0.10102825611829758, 0.07329780608415604, -0.0438249409198761, -0.1182495504617691, -0.0581011027097702, 0.04057683050632477, -0.10322711616754532, -0.02751227654516697, -0.04342721402645111, 0.028201913461089134, -0.005284039303660393, 0.1192636787891388, -0.018272195011377335, 0.07518327236175537, -0.012272762134671211, -0.020851243287324905, -0.14588844776153564, 0.14575807750225067, -0.03243175894021988, 0.0072832610458135605, 0.001161088002845645, 0.014329059049487114, -0.08454622328281403, 0.050250496715307236, -0.060539085417985916, 0.12355532497167587, -0.16111259162425995, -0.1045004352927208, 0.1906358301639557, -0.057495955377817154, -0.02144520729780197, 0.1010187566280365, 0.04819408431649208, 0.11176412552595139, 0.13767080008983612, 0.3042106330394745, 0.05147562175989151, -0.0966935083270073, -0.058888159692287445, 0.12675543129444122, -0.023945633322000504, -0.04089559242129326, 0.13886673748493195, -0.046115487813949585, 0.07033389806747437, -0.00862357672303915, 0.18587660789489746, 0.05175750330090523, -0.07374116778373718, -0.01518766675144434, 0.08476626873016357, -0.04572402685880661, 0.06628260016441345, -0.050777386873960495, 0.06665537506341934, -0.0459492988884449, 0.0004618086095433682, 0.1319427490234375, 0.12389231473207474, -0.0754992738366127, 0.07231184840202332, -0.15371224284172058, -0.07711105048656464, -0.12197574973106384, 0.010448324494063854, -0.15500415861606598, 0.16482236981391907, -0.0759701132774353, -0.054048724472522736, 0.22652092576026917, 0.0962183028459549, 0.03646625578403473, -0.02317032590508461, -0.13720370829105377, 0.05072391405701637, 0.1493673473596573, 0.07453668862581253, -0.04444025456905365, -0.20933696627616882, 0.16475963592529297, -0.08932536840438843, 0.06988205015659332, -0.061917051672935486, -0.03314726799726486, 0.04257579892873764, 0.021342333406209946, 0.030301958322525024, 0.043308813124895096, 0.13834987580776215, 0.0488944947719574, -0.005208726041018963, 0.026622705161571503, 0.006799920927733183, 0.0629364401102066, -0.18358291685581207, 0.28258249163627625, -0.19440516829490662, 0.02991047129034996, 0.12713412940502167, -0.10362089425325394, -0.04712419584393501, 0.03936239704489708, 0.02331663854420185, -0.04022809863090515, 0.07970979064702988, -0.09924289584159851, 0.08477336913347244, -0.09556392580270767, 0.09525663405656815, -0.031324148178100586, 0.015713093802332878, -0.0031415147241204977, -0.15765085816383362, -0.002853743964806199, 0.16195926070213318, -0.12991926074028015, -0.2562514543533325, 0.1022767573595047, 0.04877002164721489, -0.04303011670708656, 0.3432590663433075, -0.026902420446276665, -0.03205639123916626, -0.03927149623632431, 0.03386278823018074, -0.020457351580262184, 0.040980130434036255, -0.206990048289299, -0.020008759573101997, 0.034380778670310974, 0.09734851866960526, 0.1260804980993271, -0.024974975734949112, -0.03242369368672371, -0.041877053678035736, -0.08631771057844162, -0.18694937229156494, 0.13714134693145752, -0.04231861233711243, 0.09472712129354477, -0.05629140883684158, 0.07039140164852142, 0.06210074573755264, -0.05692256987094879, -0.14715614914894104, 0.05653291568160057, -0.2285941243171692, -0.19316338002681732, -0.09945084899663925, -0.10848867893218994, -0.04593444988131523, 0.07167180627584457, 0.11006011068820953, -0.18942783772945404, 0.0006310802418738604, -0.04544714838266373, 0.11448321491479874, -0.008653736673295498, 0.03161924332380295, 0.03238730505108833, 0.0035257067065685987, 0.013344953767955303, -0.05622578039765358, -0.0006768322200514376, 0.024066245183348656, 0.06148466095328331, 0.08974307775497437, -0.1520082950592041, 0.03138861805200577, 0.20994672179222107, 0.08753405511379242, -0.03724303096532822, -0.019787538796663284, 0.11417106539011002, -0.07253653556108475, -0.04333958029747009, 0.116937555372715, 0.013537813909351826, -0.0003517316945362836, 0.11209198832511902, -0.030331052839756012, 0.02429964579641819, 0.08967011421918869, -0.015581296756863594, -0.08503220975399017, -0.21279755234718323, -0.021013883873820305, -0.08524321764707565, 0.06479277461767197, -0.22226864099502563, -0.002008317271247506, 0.012237109243869781, -0.06356995552778244, -0.03187903016805649, -0.0753113254904747, 0.12020862847566605, 0.015567596070468426, 0.12623751163482666, -0.06612583994865417, 0.053233783692121506, -0.1382957100868225, -0.017397606745362282, 0.09918741881847382, -0.052581481635570526, 0.08420608192682266, 0.136088564991951, 0.14352667331695557, 0.028361070901155472, -0.10082480311393738, 0.10850002616643906, 0.04201801121234894, 0.05437722057104111, -0.02266881987452507, -0.011103374883532524, -0.08562162518501282, 0.0158087108284235, 0.020471937954425812, 0.21330410242080688, -0.002665627282112837, 0.05805773288011551, 0.049826111644506454, 0.0750424861907959, -0.060787711292505264, 0.11167380213737488, -0.004386514890938997, -0.07628533989191055, 0.04786306619644165, -0.06537691503763199, 0.07089479267597198, 0.03239021077752113, 0.1619819551706314, -0.02863300032913685, 0.11389821767807007, 0.0969371423125267, 0.028637295588850975, 0.055714476853609085, 0.03425350412726402, -0.12915058434009552, 0.01415732316672802, -0.050819188356399536, 0.035859301686286926, -0.2193964719772339, 0.2125677764415741, 0.05062863975763321, 0.042605698108673096, 0.027014171704649925, -0.02746681310236454, 0.0810193419456482, 0.037199292331933975, 0.08202248811721802, 0.028200307860970497, -0.09995542466640472, -0.12345106154680252, -0.08432453125715256, -0.03271346911787987, 0.12291758507490158, 0.08828476071357727, 0.001184475957415998, 0.07887659221887589, -0.03155325725674629, 0.03931744396686554, -0.08332004398107529, -0.1297404170036316, -0.01676461659371853, 0.0592510849237442, 0.15142211318016052, 0.07541569322347641, 0.050694745033979416, -0.0476490780711174, -0.10568251460790634, -0.03351324796676636, -0.11334580183029175, -0.08944254368543625, -0.03236754611134529, -0.08190368115901947, 0.0762072205543518, -0.11866612732410431, -0.10434425622224808, 0.002989848842844367, 0.023726608604192734, 0.03507128730416298, -0.024858593940734863, 0.06132597103714943, -0.028835996985435486, -0.026807861402630806, 0.00789504311978817, 0.20805983245372772, 0.06097737327218056, 0.13068325817584991, -0.025910887867212296, -0.07948479801416397, 0.009590999223291874, -0.0854327529668808, 0.037369504570961, -0.04059701785445213, -0.1307908147573471, 0.11614862084388733, 0.0037011937238276005, -0.12033260613679886, -0.11379385739564896, -0.14787650108337402, 0.17348457872867584, 0.23018582165241241, 0.0011372731532901525, 0.08427990227937698, 0.268851101398468, -0.02676408737897873, -0.21550345420837402, -0.03429793193936348, -0.011481870897114277, 0.03138067573308945, -0.023809535428881645, -0.13856066763401031, 0.03714996576309204, -0.07193250954151154, 0.014949527569115162, -0.03539949655532837, -0.16667763888835907, -0.14892610907554626, 0.11758771538734436, -0.12407334893941879, 0.16327551007270813, -0.0830768570303917, -0.09003981202840805, -0.12453336268663406, -0.013990985229611397, 0.03800950571894646, -0.23022247850894928, 0.1443430632352829, 0.11942284554243088, 0.13437460362911224, 0.007302508223801851, 0.02943466417491436, 0.16409537196159363, 0.08260694146156311, -0.04605252295732498, -0.03126038610935211, -0.009111911989748478, 0.06696794927120209, 0.0771319791674614, 0.02848874405026436, -0.009111469611525536, -0.0168902687728405, -0.10800445079803467, -0.0703500434756279, -0.05586519464850426, -0.0003863362653646618, 0.08364801108837128, -0.07369589060544968, -0.02457205019891262, -0.02120916359126568, 0.02796652540564537, -0.03924861550331116, 0.15199610590934753, -0.1916140913963318, 0.0251529011875391, 0.08770604431629181, 0.20130231976509094, -0.0814809799194336, 0.07341916114091873, -0.02271377108991146, -0.07809380441904068, 0.032931599766016006, -0.12390095740556717, 0.02085135318338871, 0.06074124202132225, 0.013102342374622822, 0.1708107888698578, -0.020596737042069435, -0.07209756970405579, 0.20813913643360138, 0.05589895695447922, -0.05305906757712364, -0.030546866357326508, -0.01914859376847744, -0.05074082687497139, 0.03575511649250984, 0.015227183699607849, 0.13986530900001526, -0.005893367808312178, 0.058438170701265335, 0.003128509037196636, -0.021894922479987144, -0.10738951712846756, 0.184207484126091, 0.05837326869368553, 0.014719177968800068, -0.06983204185962677, 0.09453323483467102, -0.058674369007349014, -0.05513300746679306, -0.04499634727835655, 0.04273783415555954, -0.06337068974971771, -0.032594114542007446, -0.1584169864654541, 0.0662631019949913, -0.04495269060134888, -0.09985603392124176, -0.05222838371992111, -0.12739568948745728, -0.007345694117248058, 0.160238116979599, 0.0019936426542699337, 0.023526404052972794, -0.003768318798393011, 0.07666821777820587, -0.003620392642915249, 0.01932695135474205, 0.09842844307422638, -0.03941944241523743, -0.10672231018543243, 0.15716244280338287, -0.016615640372037888, 0.1384028047323227, -0.05939989537000656, -0.05041259527206421, -0.050696998834609985, 0.06930439919233322, -0.11777530610561371, -0.003555821953341365, -0.18312892317771912, -0.05951924994587898, -0.015316091477870941, -0.07283513993024826, -0.061086833477020264, -0.009023577906191349, -0.0793309286236763, 0.06305219978094101, 0.030437733978033066, 0.07033856213092804, 0.0012400004779919982, -0.011316092684864998, 0.009676109068095684, -0.02440342679619789, 0.09724433720111847, 0.239492267370224, -0.09534400701522827, 0.14703316986560822, -0.13829967379570007, -0.0015320319216698408, 0.11645317077636719, 0.030500411987304688, 0.002959667472168803, -0.006888201460242271, -0.09631145745515823, 0.07081473618745804, 0.077300064265728, 0.010559805668890476, 0.06689897179603577, 0.023684337735176086, 0.034243930131196976, -0.23750372231006622, 0.024977482855319977, -0.09563751518726349, 0.03239115700125694, 0.15865013003349304, 0.12700776755809784, 0.13739046454429626, -0.12284263223409653, 0.034842826426029205, -0.03182158246636391, 0.030685894191265106, -0.0005232273833826184, -0.07271174341440201, 0.04110342264175415, -0.006544910371303558, 0.06258727610111237, -0.11603289842605591, 0.09800447523593903, -0.12590457499027252, -0.05839982628822327, 0.01776018738746643, -0.06934770196676254, -0.01605360023677349, 0.0017624336760491133, 0.12355001270771027, 0.10019601881504059, -0.018195828422904015, -0.0599970668554306, 0.021452616900205612, 0.010880588553845882, 0.08495242148637772, 0.01166176050901413, 0.07112099230289459, 0.06546328216791153, 0.10949883610010147, 0.039882954210042953, 0.010698872618377209, -0.09535231441259384, 0.04245658963918686, -0.10537470877170563, -0.07201841473579407, -0.03853154554963112, 0.138533815741539, 0.19024768471717834, -0.023400362581014633, 0.0021870725322514772, 0.009286217391490936, -0.04983610287308693, -0.14377182722091675, -0.04629359394311905, -0.08664871007204056, -0.04720451310276985, 0.012714265845716, -0.08431746065616608, 0.08134668320417404, 0.05761624500155449, 0.0815102830529213, 0.028851134702563286, 0.1319241225719452, -0.13125307857990265, -0.16840672492980957, 0.0628746747970581, -0.08463409543037415, -0.03497694060206413, -0.05652439966797829, -0.004255864769220352, 0.18394050002098083, -0.008341338485479355, 0.0057656303979456425, 0.033175259828567505, -0.04968641325831413, -0.04368165135383606, -0.20059813559055328, -0.052916619926691055, -0.03203430771827698, 0.10160005837678909, 0.11601488292217255, 0.06802131235599518, 0.08470802754163742, -0.09168317914009094, -0.0009638936026021838, 0.1067146509885788, -0.04115144908428192, -0.17358934879302979, -0.11546776443719864, -0.041424307972192764, -0.06169280782341957, 0.09230902045965195, -0.07206740230321884, -0.0409369170665741, -0.0768832415342331, 0.07583044469356537, 0.24147427082061768, -0.07558790594339371, 0.06311003863811493, -0.06242183968424797, 0.01929156668484211, -0.031010471284389496, -0.009053151123225689, 0.08384942263364792, 0.09035254269838333, 0.06556475162506104, -0.022091593593358994, -0.14537742733955383, -0.00011656885908450931, 0.013749142177402973, 0.021450906991958618, -0.006334635429084301, -0.039255302399396896, -0.013256696984171867, 0.1202995553612709, -0.20711657404899597, -0.15087059140205383, -0.2472085952758789, -0.06728611141443253, 0.005287447478622198, 0.05834032595157623, -0.011409305967390537, 0.17529304325580597, -0.040105339139699936, -0.03143277019262314, 0.0013532857410609722, -0.07642611861228943, 0.03543864190578461, -0.03758752718567848, 0.05402497202157974, 0.05663899704813957, -0.0951174646615982, -0.0933038741350174, -0.018688542768359184, 0.16377490758895874, 0.022094257175922394, 0.07958563417196274, 0.05578811466693878, 0.2149164378643036, 0.048261478543281555, -0.05616005137562752, 0.030325956642627716, 0.1668316125869751, -0.03554106876254082, 0.23795348405838013, 0.05122339352965355, 0.002557766158133745, 0.0308447927236557, 0.004670997615903616, -0.10935825109481812, -0.04449107125401497, 0.05850309878587723, -0.09445557743310928, 0.08420276641845703, 0.05751708149909973, -0.03263548016548157, -0.019328486174345016, 0.041893843561410904, -0.05878949165344238, -0.013283446431159973, -0.09067162126302719, -0.09038535505533218, -0.18263377249240875, -0.02822830155491829, 0.0035364364739507437, 0.052347708493471146, -0.19209244847297668, -0.05275467783212662, -0.14471103250980377, 0.009077333845198154, 0.037671491503715515, 0.07820555567741394, 0.019830988720059395, 0.00559460511431098, 0.007970675826072693, -0.02569739893078804, 0.04029844328761101, 0.15322354435920715, -0.09279639273881912, -0.13242441415786743 ]