sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
listlengths
1
1.84k
created_at
stringlengths
25
25
arxiv
listlengths
0
201
languages
listlengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
listlengths
0
722
processed_texts
listlengths
1
723
tokens_length
listlengths
1
723
input_texts
listlengths
1
61
embeddings
listlengths
768
768
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1325309784191459329/XJXVbxEi_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Anshul Kundaje</div> <div style="text-align: center; font-size: 14px;">@anshulkundaje</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Anshul Kundaje. | Data | Anshul Kundaje | | --- | --- | | Tweets downloaded | 3245 | | Retweets | 2396 | | Short tweets | 72 | | Tweets kept | 777 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1zmfoelh/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @anshulkundaje's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3unb3kxb) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3unb3kxb/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/anshulkundaje') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/anshulkundaje/1621976127507/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/anshulkundaje
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Anshul Kundaje @anshulkundaje I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Anshul Kundaje. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @anshulkundaje's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1321608907882663941/BV4i0pBm_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Anson Tong 🤖 AI Bot </div> <div style="font-size: 15px">@ansonjtong bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@ansonjtong's tweets](https://twitter.com/ansonjtong). | Data | Quantity | | --- | --- | | Tweets downloaded | 900 | | Retweets | 413 | | Short tweets | 42 | | Tweets kept | 445 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/c6jwbmq3/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ansonjtong's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3d5w7wzh) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3d5w7wzh/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ansonjtong') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/ansonjtong/1616693149848/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/ansonjtong
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Anson Tong AI Bot @ansonjtong bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @ansonjtong's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ansonjtong's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1477498953524518912/yvJkd9VL_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">ANTICARBON</div> <div style="text-align: center; font-size: 14px;">@anticarbons</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from ANTICARBON. | Data | ANTICARBON | | --- | --- | | Tweets downloaded | 2518 | | Retweets | 427 | | Short tweets | 352 | | Tweets kept | 1739 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/s9q99sc5/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @anticarbons's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1k8boybi) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1k8boybi/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/anticarbons') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/anticarbons/1642719091326/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/anticarbons
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT ANTICARBON @anticarbons I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from ANTICARBON. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @anticarbons's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1375086529270398977/w8zAorR0_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Dawn 😼 🤖 AI Bot </div> <div style="font-size: 15px">@antifashgremlin bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@antifashgremlin's tweets](https://twitter.com/antifashgremlin). | Data | Quantity | | --- | --- | | Tweets downloaded | 3231 | | Retweets | 258 | | Short tweets | 599 | | Tweets kept | 2374 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2qlnfyby/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @antifashgremlin's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1dj10xj4) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1dj10xj4/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/antifashgremlin') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/antifashgremlin/1616827903052/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/antifashgremlin
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Dawn AI Bot @antifashgremlin bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @antifashgremlin's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @antifashgremlin's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1429221119647752192/XKX0DgWA_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Sarcastic Venom</div> <div style="text-align: center; font-size: 14px;">@antiihope</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Sarcastic Venom. | Data | Sarcastic Venom | | --- | --- | | Tweets downloaded | 3245 | | Retweets | 36 | | Short tweets | 877 | | Tweets kept | 2332 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2ervvg9p/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @antiihope's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2wbdnrdn) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2wbdnrdn/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/antiihope') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/antiihope/1629824412403/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/antiihope
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Sarcastic Venom @antiihope I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Sarcastic Venom. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @antiihope's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1065944627268730880/z7DXpekv_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Antoine Bordes 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@antoinebordes bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@antoinebordes's tweets](https://twitter.com/antoinebordes). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>257</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>195</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>8</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>54</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/161sn1dw/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @antoinebordes's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/3kw8rzi0) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/3kw8rzi0/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/antoinebordes'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/antoinebordes
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Antoine Bordes AI Bot </div> <div style="font-size: 15px; color: #657786">@antoinebordes bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @antoinebordes's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>257</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>195</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>8</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>54</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @antoinebordes's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/antoinebordes'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @antoinebordes's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>257</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>195</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>8</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>54</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @antoinebordes's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/antoinebordes'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @antoinebordes's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>257</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>195</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>8</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>54</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @antoinebordes's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/antoinebordes'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 429, 76, 9, 168, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1378303317433024513/KAxy7ESG_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">antto 🤖 AI Bot </div> <div style="font-size: 15px">@anttoretu bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@anttoretu's tweets](https://twitter.com/anttoretu). | Data | Quantity | | --- | --- | | Tweets downloaded | 3236 | | Retweets | 456 | | Short tweets | 1319 | | Tweets kept | 1461 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/rdz4tooo/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @anttoretu's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3t776sk6) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3t776sk6/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/anttoretu') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/anttoretu/1617913015894/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/anttoretu
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
antto AI Bot @anttoretu bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @anttoretu's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @anttoretu's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1370281893271281666/v6-WAWCk_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">antyzer 🤖 AI Bot </div> <div style="font-size: 15px">@antyzer_ bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@antyzer_'s tweets](https://twitter.com/antyzer_). | Data | Quantity | | --- | --- | | Tweets downloaded | 3117 | | Retweets | 1213 | | Short tweets | 635 | | Tweets kept | 1269 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/230q2cin/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @antyzer_'s tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/317k4hqh) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/317k4hqh/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/antyzer_') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/antyzer_/1616722940871/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/antyzer_
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
antyzer AI Bot @antyzer\_ bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @antyzer\_'s tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @antyzer\_'s tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1322782820964667392/dcigipzG_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">cool kid anushk 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@anushkmittal bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@anushkmittal's tweets](https://twitter.com/anushkmittal). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3204</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>728</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>321</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2155</td> </tr> </tbody> </table> [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1qa9h984/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @anushkmittal's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1dp13cdl) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1dp13cdl/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/anushkmittal'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/anushkmittal/1607746679770/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/anushkmittal
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">cool kid anushk AI Bot </div> <div style="font-size: 15px; color: #657786">@anushkmittal bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @anushkmittal's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3204</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>728</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>321</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2155</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @anushkmittal's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/anushkmittal'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @anushkmittal's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3204</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>728</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>321</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2155</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @anushkmittal's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/anushkmittal'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @anushkmittal's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3204</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>728</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>321</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2155</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @anushkmittal's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/anushkmittal'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 432, 77, 9, 169, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1305932428607922177/Rh6HaRlW_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">YugenSSBU</div> <div style="text-align: center; font-size: 14px;">@anvers1158</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from YugenSSBU. | Data | YugenSSBU | | --- | --- | | Tweets downloaded | 257 | | Retweets | 16 | | Short tweets | 19 | | Tweets kept | 222 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2qmptuy4/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @anvers1158's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2erdua3k) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2erdua3k/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/anvers1158') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/anvers1158/1639530531829/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/anvers1158
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT YugenSSBU @anvers1158 I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from YugenSSBU. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @anvers1158's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/923274881197895680/AbHcStkl_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Alexandria Ocasio-Cortez</div> <div style="text-align: center; font-size: 14px;">@aoc</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Alexandria Ocasio-Cortez. | Data | Alexandria Ocasio-Cortez | | --- | --- | | Tweets downloaded | 3221 | | Retweets | 1253 | | Short tweets | 126 | | Tweets kept | 1842 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3i05suuv/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @aoc's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1gjmi5b8) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1gjmi5b8/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/aoc') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/aoc/1658528812949/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/aoc
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Alexandria Ocasio-Cortez @aoc I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Alexandria Ocasio-Cortez. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @aoc's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1361559454882365441/7sIpFR-Z_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">gabi 🤖 AI Bot </div> <div style="font-size: 15px">@appleddragon bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@appleddragon's tweets](https://twitter.com/appleddragon). | Data | Quantity | | --- | --- | | Tweets downloaded | 3172 | | Retweets | 644 | | Short tweets | 813 | | Tweets kept | 1715 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/i14p6r6k/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @appleddragon's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/26jav9ze) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/26jav9ze/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/appleddragon') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/appleddragon/1614103802939/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/appleddragon
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
gabi AI Bot @appleddragon bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @appleddragon's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @appleddragon's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1006876146443669505/w9tyOPGm_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Antonin Raffin 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@araffin2 bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@araffin2's tweets](https://twitter.com/araffin2). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>446</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>173</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>5</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>268</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/3qawuhc2/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @araffin2's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/aczxeidd) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/aczxeidd/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/araffin2'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/araffin2/1602238408015/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/araffin2
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Antonin Raffin AI Bot </div> <div style="font-size: 15px; color: #657786">@araffin2 bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @araffin2's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>446</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>173</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>5</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>268</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @araffin2's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/araffin2'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @araffin2's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>446</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>173</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>5</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>268</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @araffin2's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/araffin2'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @araffin2's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>446</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>173</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>5</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>268</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @araffin2's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/araffin2'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 428, 75, 9, 167, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1236224512737390592/nYMKnkqe_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">𝔸𝕣𝕖𝕫𝕟𝕠 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@arezno bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@arezno's tweets](https://twitter.com/arezno). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3213</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>1080</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>340</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>1793</td> </tr> </tbody> </table> [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/jet2tw15/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @arezno's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/r7cnhzlr) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/r7cnhzlr/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/arezno'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/arezno/1608197180736/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/arezno
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">𝔸𝕣𝕖𝕫𝕟𝕠 AI Bot </div> <div style="font-size: 15px; color: #657786">@arezno bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @arezno's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3213</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>1080</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>340</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>1793</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @arezno's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/arezno'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @arezno's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3213</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>1080</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>340</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>1793</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @arezno's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/arezno'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @arezno's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3213</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>1080</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>340</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>1793</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @arezno's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/arezno'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 428, 74, 9, 166, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1813281667/arrl-logo-ylo2_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">ARRL 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@arrl bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@arrl's tweets](https://twitter.com/arrl). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3210</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>173</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>37</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>3000</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/2zoukgkm/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @arrl's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/2oucppvx) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/2oucppvx/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/arrl'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/arrl/1603319674811/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/arrl
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">ARRL AI Bot </div> <div style="font-size: 15px; color: #657786">@arrl bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @arrl's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3210</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>173</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>37</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>3000</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @arrl's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/arrl'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @arrl's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3210</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>173</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>37</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>3000</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @arrl's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/arrl'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @arrl's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3210</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>173</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>37</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>3000</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @arrl's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/arrl'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 427, 74, 9, 166, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1479498403251896320/uDVlO62z_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">الرياضية - عاجل</div> <div style="text-align: center; font-size: 14px;">@arryadia_brk</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from الرياضية - عاجل. | Data | الرياضية - عاجل | | --- | --- | | Tweets downloaded | 1548 | | Retweets | 11 | | Short tweets | 33 | | Tweets kept | 1504 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/24udtdhw/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @arryadia_brk's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2e36ahiu) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2e36ahiu/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/arryadia_brk') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/arryadia_brk/1643119471683/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/arryadia_brk
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT الرياضية - عاجل @arryadia\_brk I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from الرياضية - عاجل. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @arryadia\_brk's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1383564179471175680/xALN4Z-R_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Aster, internet archaeologist 🤖 AI Bot </div> <div style="font-size: 15px">@arsonatdennys bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@arsonatdennys's tweets](https://twitter.com/arsonatdennys). | Data | Quantity | | --- | --- | | Tweets downloaded | 3207 | | Retweets | 1561 | | Short tweets | 223 | | Tweets kept | 1423 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2rggt8p2/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @arsonatdennys's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/350x43du) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/350x43du/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/arsonatdennys') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/arsonatdennys/1620019020010/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/arsonatdennys
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Aster, internet archaeologist AI Bot @arsonatdennys bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @arsonatdennys's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @arsonatdennys's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1342836590998134786/tDwNDfFs_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">frostington ambassady the third (5’2”) 🤖 AI Bot </div> <div style="font-size: 15px">@arsondoer bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@arsondoer's tweets](https://twitter.com/arsondoer). | Data | Quantity | | --- | --- | | Tweets downloaded | 3200 | | Retweets | 270 | | Short tweets | 799 | | Tweets kept | 2131 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3mhuavj6/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @arsondoer's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2fz88vjc) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2fz88vjc/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/arsondoer') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/arsondoer/1616645630695/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/arsondoer
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
frostington ambassady the third (5’2”) AI Bot @arsondoer bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @arsondoer's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @arsondoer's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1183085761123573765/Eq-EK4l5_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">ArtificialStupidity</div> <div style="text-align: center; font-size: 14px;">@artificialstup5</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from ArtificialStupidity. | Data | ArtificialStupidity | | --- | --- | | Tweets downloaded | 397 | | Retweets | 1 | | Short tweets | 99 | | Tweets kept | 297 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/y4u5glx8/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @artificialstup5's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/nvkezn5h) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/nvkezn5h/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/artificialstup5') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/artificialstup5/1625398017410/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/artificialstup5
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT ArtificialStupidity @artificialstup5 I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from ArtificialStupidity. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @artificialstup5's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1264848279411531776/fJ3OZJEx_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Artorra 🤖 AI Bot </div> <div style="font-size: 15px">@artorrattv bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@artorrattv's tweets](https://twitter.com/artorrattv). | Data | Quantity | | --- | --- | | Tweets downloaded | 216 | | Retweets | 2 | | Short tweets | 13 | | Tweets kept | 201 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1eeu79ju/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @artorrattv's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1hazm0dg) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1hazm0dg/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/artorrattv') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/artorrattv
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Artorra AI Bot @artorrattv bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @artorrattv's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @artorrattv's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1406785434093604864/pNV1y7vJ_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Starcross</div> <div style="text-align: center; font-size: 14px;">@artstarcross</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Starcross. | Data | Starcross | | --- | --- | | Tweets downloaded | 1846 | | Retweets | 217 | | Short tweets | 67 | | Tweets kept | 1562 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/177l3jal/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @artstarcross's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2w1qo4hm) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2w1qo4hm/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/artstarcross') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/artstarcross/1627659166884/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/artstarcross
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Starcross @artstarcross I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Starcross. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @artstarcross's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1374537552292687879/Sy7M0aFk_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1400341059842891782/nJw_YYUy_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">👑 Prince Reinhard Ascart 👑 DEBUT TBA(COMMS OPEN) & Kicchin (Most Powerful VTweeter)</div> <div style="text-align: center; font-size: 14px;">@ascartprince-kicchinnezumi</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from 👑 Prince Reinhard Ascart 👑 DEBUT TBA(COMMS OPEN) & Kicchin (Most Powerful VTweeter). | Data | 👑 Prince Reinhard Ascart 👑 DEBUT TBA(COMMS OPEN) | Kicchin (Most Powerful VTweeter) | | --- | --- | --- | | Tweets downloaded | 3240 | 3247 | | Retweets | 672 | 644 | | Short tweets | 1223 | 1223 | | Tweets kept | 1345 | 1380 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1voh8kfv/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ascartprince-kicchinnezumi's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/y5knw4f6) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/y5knw4f6/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ascartprince-kicchinnezumi') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/ascartprince-kicchinnezumi/1623135392213/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/ascartprince-kicchinnezumi
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Prince Reinhard Ascart DEBUT TBA(COMMS OPEN) & Kicchin (Most Powerful VTweeter) @ascartprince-kicchinnezumi I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Prince Reinhard Ascart DEBUT TBA(COMMS OPEN) & Kicchin (Most Powerful VTweeter). Data: Tweets downloaded, Prince Reinhard Ascart DEBUT TBA(COMMS OPEN): 3240, Kicchin (Most Powerful VTweeter): 3247 Data: Retweets, Prince Reinhard Ascart DEBUT TBA(COMMS OPEN): 672, Kicchin (Most Powerful VTweeter): 644 Data: Short tweets, Prince Reinhard Ascart DEBUT TBA(COMMS OPEN): 1223, Kicchin (Most Powerful VTweeter): 1223 Data: Tweets kept, Prince Reinhard Ascart DEBUT TBA(COMMS OPEN): 1345, Kicchin (Most Powerful VTweeter): 1380 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ascartprince-kicchinnezumi's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1205345506433675264/l5Rq68pX_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Arsenio Dev 🤖 AI Bot </div> <div style="font-size: 15px">@ascii211 bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@ascii211's tweets](https://twitter.com/ascii211). | Data | Quantity | | --- | --- | | Tweets downloaded | 3249 | | Retweets | 98 | | Short tweets | 257 | | Tweets kept | 2894 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/f1o7qkfr/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ascii211's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/39q62gze) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/39q62gze/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ascii211') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/ascii211/1617764322125/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/ascii211
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Arsenio Dev AI Bot @ascii211 bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @ascii211's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ascii211's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('http://pbs.twimg.com/profile_images/1235512936547966977/_YotVKfT_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Asım Cesim 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@asimcesim bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@asimcesim's tweets](https://twitter.com/asimcesim). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>865</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>638</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>16</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>211</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/1ooeyfob/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @asimcesim's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/2gujryp4) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/2gujryp4/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/asimcesim'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://res.cloudinary.com/huggingtweets/image/upload/v1600040134/asimcesim.jpg", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/asimcesim
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Asım Cesim AI Bot </div> <div style="font-size: 15px; color: #657786">@asimcesim bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @asimcesim's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>865</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>638</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>16</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>211</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @asimcesim's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/asimcesim'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @asimcesim's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>865</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>638</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>16</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>211</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @asimcesim's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/asimcesim'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @asimcesim's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>865</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>638</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>16</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>211</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @asimcesim's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/asimcesim'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 429, 76, 9, 168, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/875394454449815552/FAzOLgVh_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">A Small Fiction 🤖 AI Bot </div> <div style="font-size: 15px">@asmallfiction bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@asmallfiction's tweets](https://twitter.com/asmallfiction). | Data | Quantity | | --- | --- | | Tweets downloaded | 2034 | | Retweets | 197 | | Short tweets | 75 | | Tweets kept | 1762 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/7bib97vd/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @asmallfiction's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3blkqco2) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3blkqco2/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/asmallfiction') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/asmallfiction/1616770285259/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/asmallfiction
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
A Small Fiction AI Bot @asmallfiction bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @asmallfiction's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @asmallfiction's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1326889766240067585/DLmTeuFh_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">a softer scp foundation 🤖 AI Bot </div> <div style="font-size: 15px">@asofterscp bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@asofterscp's tweets](https://twitter.com/asofterscp). | Data | Quantity | | --- | --- | | Tweets downloaded | 3250 | | Retweets | 0 | | Short tweets | 0 | | Tweets kept | 3250 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2jj05g3q/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @asofterscp's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/32zww8ef) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/32zww8ef/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/asofterscp') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/asofterscp/1617358172358/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/asofterscp
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
a softer scp foundation AI Bot @asofterscp bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @asofterscp's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @asofterscp's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1393094522008080385/1urtPrKy_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1422647641507311617/_phzOoGk_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1366057669027639300/ulWJe-9i_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">gnome 👼🏻 & Ass 🇮🇹🏳️‍⚧️🍕🎩 & tyler</div> <div style="text-align: center; font-size: 14px;">@ass420weed-gnomeszs-tyler01010101</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from gnome 👼🏻 & Ass 🇮🇹🏳️‍⚧️🍕🎩 & tyler. | Data | gnome 👼🏻 | Ass 🇮🇹🏳️‍⚧️🍕🎩 | tyler | | --- | --- | --- | --- | | Tweets downloaded | 3220 | 3193 | 3238 | | Retweets | 1079 | 1470 | 110 | | Short tweets | 438 | 703 | 1066 | | Tweets kept | 1703 | 1020 | 2062 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2dn68r1g/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ass420weed-gnomeszs-tyler01010101's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/163tmae6) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/163tmae6/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ass420weed-gnomeszs-tyler01010101') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/ass420weed-gnomeszs-tyler01010101/1628062907982/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/ass420weed-gnomeszs-tyler01010101
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG gnome & Ass 🇮🇹️‍️ & tyler @ass420weed-gnomeszs-tyler01010101 I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from gnome & Ass 🇮🇹️‍️ & tyler. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ass420weed-gnomeszs-tyler01010101's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1323522646152282120/STwG1Xk3_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Atheistic One 🤖 AI Bot </div> <div style="font-size: 15px">@atheistic_1 bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@atheistic_1's tweets](https://twitter.com/atheistic_1). | Data | Quantity | | --- | --- | | Tweets downloaded | 3247 | | Retweets | 179 | | Short tweets | 275 | | Tweets kept | 2793 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2gyocq1j/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atheistic_1's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/l5vjnai7) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/l5vjnai7/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atheistic_1') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atheistic_1/1616797786127/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atheistic_1
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Atheistic One AI Bot @atheistic\_1 bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @atheistic\_1's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atheistic\_1's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1042510623962275840/1Iw_Mvud_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Sébastien Chopin 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@atinux bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@atinux's tweets](https://twitter.com/atinux). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3204</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>792</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>302</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2110</td> </tr> </tbody> </table> [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/6nzh2yez/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atinux's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/15g6ru92) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/15g6ru92/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/atinux'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atinux/1605286256394/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atinux
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Sébastien Chopin AI Bot </div> <div style="font-size: 15px; color: #657786">@atinux bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @atinux's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3204</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>792</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>302</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2110</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @atinux's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/atinux'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @atinux's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3204</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>792</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>302</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2110</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @atinux's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/atinux'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @atinux's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3204</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>792</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>302</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2110</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @atinux's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/atinux'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 429, 74, 9, 166, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1377989668189405192/II6ZfJPK_400x400.png&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Atlassian</div> <div style="text-align: center; font-size: 14px;">@atlassian</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Atlassian. | Data | Atlassian | | --- | --- | | Tweets downloaded | 3249 | | Retweets | 824 | | Short tweets | 58 | | Tweets kept | 2367 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2i1f4hr0/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atlassian's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/olb55vh0) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/olb55vh0/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atlassian') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atlassian/1623889197185/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atlassian
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Atlassian @atlassian I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Atlassian. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atlassian's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1351987572747202560/v_vDGtnX_400x400.png')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">AtomicNicos | @d_overcon co-organiser 🤖 AI Bot </div> <div style="font-size: 15px">@atomicnicos bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@atomicnicos's tweets](https://twitter.com/atomicnicos). | Data | Quantity | | --- | --- | | Tweets downloaded | 3249 | | Retweets | 221 | | Short tweets | 452 | | Tweets kept | 2576 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3mnuo591/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atomicnicos's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/n30kmifp) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/n30kmifp/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atomicnicos') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atomicnicos/1616352732946/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atomicnicos
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AtomicNicos | @d\_overcon co-organiser AI Bot @atomicnicos bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @atomicnicos's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atomicnicos's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1238289351861137408/k-sNIYOh_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">site specific carnivorous occurrence 🤖 AI Bot </div> <div style="font-size: 15px">@atomicthumbs bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@atomicthumbs's tweets](https://twitter.com/atomicthumbs). | Data | Quantity | | --- | --- | | Tweets downloaded | 3203 | | Retweets | 1686 | | Short tweets | 287 | | Tweets kept | 1230 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2oe2pbns/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atomicthumbs's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3jgkhbgt) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3jgkhbgt/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atomicthumbs') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atomicthumbs/1614107568559/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atomicthumbs
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
site specific carnivorous occurrence AI Bot @atomicthumbs bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @atomicthumbs's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atomicthumbs's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1363925362699337731/cAnWI7wL_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Atreyu Joker 🤖 AI Bot </div> <div style="font-size: 15px">@atreyupilled bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@atreyupilled's tweets](https://twitter.com/atreyupilled). | Data | Quantity | | --- | --- | | Tweets downloaded | 3233 | | Retweets | 232 | | Short tweets | 500 | | Tweets kept | 2501 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1nd8rl11/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atreyupilled's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1gt351ez) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1gt351ez/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atreyupilled') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atreyupilled/1614099692275/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atreyupilled
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Atreyu Joker AI Bot @atreyupilled bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @atreyupilled's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atreyupilled's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1323206624765448197/eqBniY_E_400x400.png&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">AtticScientist</div> <div style="text-align: center; font-size: 14px;">@atticscientist</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from AtticScientist. | Data | AtticScientist | | --- | --- | | Tweets downloaded | 3250 | | Retweets | 1 | | Short tweets | 8 | | Tweets kept | 3241 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3pvpbxir/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @atticscientist's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/ktckeg7n) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/ktckeg7n/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/atticscientist') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/atticscientist/1625552752637/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/atticscientist
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT AtticScientist @atticscientist I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from AtticScientist. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @atticscientist's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1137529671690706949/nKjy1T77_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Himboclese,,,, of ACAB 🤖 AI Bot </div> <div style="font-size: 15px">@august77lng bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@august77lng's tweets](https://twitter.com/august77lng). | Data | Quantity | | --- | --- | | Tweets downloaded | 3151 | | Retweets | 2357 | | Short tweets | 114 | | Tweets kept | 680 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3ly74f65/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @august77lng's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3ck3jotc) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3ck3jotc/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/august77lng') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/august77lng/1614115355848/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/august77lng
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Himboclese,,,, of ACAB AI Bot @august77lng bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @august77lng's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @august77lng's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1400755065644802052/ZaFgV9cj_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Haruka</div> <div style="text-align: center; font-size: 14px;">@aumgensokyo</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Haruka. | Data | Haruka | | --- | --- | | Tweets downloaded | 3227 | | Retweets | 632 | | Short tweets | 494 | | Tweets kept | 2101 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/lpttj7aa/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @aumgensokyo's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/mxwc1h4k) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/mxwc1h4k/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/aumgensokyo') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/aumgensokyo/1624891912864/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/aumgensokyo
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Haruka @aumgensokyo I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Haruka. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @aumgensokyo's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1298871483310190593/Bev51BER_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Austen Allred 🤖 AI Bot </div> <div style="font-size: 15px">@austen bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@austen's tweets](https://twitter.com/austen). | Data | Quantity | | --- | --- | | Tweets downloaded | 3249 | | Retweets | 112 | | Short tweets | 465 | | Tweets kept | 2672 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/wyv57q5p/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @austen's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2jltqo1x) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2jltqo1x/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/austen') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/austen/1617725908050/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/austen
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Austen Allred AI Bot @austen bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @austen's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @austen's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1356365325663666177/fgKMKMGD_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">April 🤖 AI Bot </div> <div style="font-size: 15px">@autogynefiles bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@autogynefiles's tweets](https://twitter.com/autogynefiles). | Data | Quantity | | --- | --- | | Tweets downloaded | 3202 | | Retweets | 901 | | Short tweets | 360 | | Tweets kept | 1941 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2jw5booy/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @autogynefiles's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1b1l87sd) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1b1l87sd/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/autogynefiles') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/autogynefiles
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
April AI Bot @autogynefiles bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @autogynefiles's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @autogynefiles's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1357641829710241792/93qv43Iv_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">mika 🤖 AI Bot </div> <div style="font-size: 15px">@autophagian bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@autophagian's tweets](https://twitter.com/autophagian). | Data | Quantity | | --- | --- | | Tweets downloaded | 3237 | | Retweets | 586 | | Short tweets | 380 | | Tweets kept | 2271 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/16w10v81/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @autophagian's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3dalb9ir) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3dalb9ir/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/autophagian') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/autophagian/1616669932241/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/autophagian
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
mika AI Bot @autophagian bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @autophagian's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @autophagian's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1192531689060200448/S9KoiehJ_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1294927107605356544/CVXTlp9y_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1468895545007775746/NIWzzmye_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Speedcafe.com & Formula One World & Autosport</div> <div style="text-align: center; font-size: 14px;">@autosport-formulaoneworld-speedcafe</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Speedcafe.com & Formula One World & Autosport. | Data | Speedcafe.com | Formula One World | Autosport | | --- | --- | --- | --- | | Tweets downloaded | 3250 | 3247 | 3250 | | Retweets | 0 | 2778 | 52 | | Short tweets | 3 | 178 | 15 | | Tweets kept | 3247 | 291 | 3183 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/kcn72bl0/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @autosport-formulaoneworld-speedcafe's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2fq703qs) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2fq703qs/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/autosport-formulaoneworld-speedcafe') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/autosport-formulaoneworld-speedcafe/1642217065882/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/autosport-formulaoneworld-speedcafe
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG URL & Formula One World & Autosport @autosport-formulaoneworld-speedcafe I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from URL & Formula One World & Autosport. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @autosport-formulaoneworld-speedcafe's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1046984086026371073/zN5I5g0G_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">deep dish pizza eater 🤖 AI Bot </div> <div style="font-size: 15px">@avantredguard bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@avantredguard's tweets](https://twitter.com/avantredguard). | Data | Quantity | | --- | --- | | Tweets downloaded | 3173 | | Retweets | 851 | | Short tweets | 386 | | Tweets kept | 1936 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2ybevoux/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @avantredguard's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/10kbxe0u) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/10kbxe0u/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/avantredguard') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/avantredguard/1614096845665/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/avantredguard
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
deep dish pizza eater AI Bot @avantredguard bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @avantredguard's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @avantredguard's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1368753714568327168/oh6BSjqX_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">AverageVermontSmasher</div> <div style="text-align: center; font-size: 14px;">@averagesmasher</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from AverageVermontSmasher. | Data | AverageVermontSmasher | | --- | --- | | Tweets downloaded | 41 | | Retweets | 0 | | Short tweets | 2 | | Tweets kept | 39 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/auyr340s/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @averagesmasher's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2qnfjchi) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2qnfjchi/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/averagesmasher') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/averagesmasher/1625924846625/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/averagesmasher
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT AverageVermontSmasher @averagesmasher I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from AverageVermontSmasher. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @averagesmasher's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/847818629840228354/VXyQHfn0_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1427457256958930948/J2FGNejT_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1354274870264266753/9D_FgIsC_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">wint & Chet & ac</div> <div style="text-align: center; font-size: 14px;">@avgmeat-dril-methwaffles</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from wint & Chet & ac. | Data | wint | Chet | ac | | --- | --- | --- | --- | | Tweets downloaded | 3189 | 2471 | 3167 | | Retweets | 468 | 748 | 209 | | Short tweets | 310 | 299 | 816 | | Tweets kept | 2411 | 1424 | 2142 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1gv4gxjf/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @avgmeat-dril-methwaffles's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3dg2j508) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3dg2j508/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/avgmeat-dril-methwaffles') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/avgmeat-dril-methwaffles/1631963152302/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/avgmeat-dril-methwaffles
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG wint & Chet & ac @avgmeat-dril-methwaffles I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from wint & Chet & ac. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @avgmeat-dril-methwaffles's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/847818629840228354/VXyQHfn0_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1255544995190321152/2ZTJAJ7K_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1354274870264266753/9D_FgIsC_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">wint & megan & ac</div> <div style="text-align: center; font-size: 14px;">@avgmeat-dril-slitthroatz</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from wint & megan & ac. | Data | wint | megan | ac | | --- | --- | --- | --- | | Tweets downloaded | 3186 | 3175 | 3167 | | Retweets | 466 | 190 | 209 | | Short tweets | 310 | 920 | 816 | | Tweets kept | 2410 | 2065 | 2142 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1p8qo7jh/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @avgmeat-dril-slitthroatz's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/29pi8slb) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/29pi8slb/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/avgmeat-dril-slitthroatz') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/avgmeat-dril-slitthroatz
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG wint & megan & ac @avgmeat-dril-slitthroatz I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from wint & megan & ac. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @avgmeat-dril-slitthroatz's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1072899793054121990/QPK0Qv46_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Avril Lavigne 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@avrillavigne bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@avrillavigne's tweets](https://twitter.com/avrillavigne). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3112</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>766</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>306</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2040</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/f4t9xd39/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @avrillavigne's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/2rkl8zq3) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/2rkl8zq3/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/avrillavigne'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/avrillavigne/1601265702542/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/avrillavigne
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Avril Lavigne AI Bot </div> <div style="font-size: 15px; color: #657786">@avrillavigne bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @avrillavigne's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3112</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>766</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>306</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2040</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @avrillavigne's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/avrillavigne'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @avrillavigne's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3112</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>766</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>306</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2040</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @avrillavigne's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/avrillavigne'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @avrillavigne's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3112</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>766</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>306</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2040</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @avrillavigne's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/avrillavigne'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 431, 76, 9, 168, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1336485377461219329/VqCT-VZ0_400x400.png')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">A Wondering of Eye 🤖 AI Bot </div> <div style="font-size: 15px">@awanderingi bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@awanderingi's tweets](https://twitter.com/awanderingi). | Data | Quantity | | --- | --- | | Tweets downloaded | 1641 | | Retweets | 202 | | Short tweets | 166 | | Tweets kept | 1273 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3pnmx279/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @awanderingi's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/19ne4lje) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/19ne4lje/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/awanderingi') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/awanderingi/1616615246454/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/awanderingi
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
A Wondering of Eye AI Bot @awanderingi bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @awanderingi's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @awanderingi's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1374488661673185281/l071mOH1_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Chris 🤖 AI Bot </div> <div style="font-size: 15px">@awaythrow8 bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@awaythrow8's tweets](https://twitter.com/awaythrow8). | Data | Quantity | | --- | --- | | Tweets downloaded | 3060 | | Retweets | 230 | | Short tweets | 469 | | Tweets kept | 2361 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/18vas1gq/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @awaythrow8's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2ju36rrz) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2ju36rrz/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/awaythrow8') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/awaythrow8/1617904264202/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/awaythrow8
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Chris AI Bot @awaythrow8 bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @awaythrow8's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @awaythrow8's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1402029332516773888/oJJ69stf_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Axel! ♠️</div> <div style="text-align: center; font-size: 14px;">@axel_hugsky</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Axel! ♠️. | Data | Axel! ♠️ | | --- | --- | | Tweets downloaded | 3244 | | Retweets | 529 | | Short tweets | 1491 | | Tweets kept | 1224 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2ox7p0bd/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @axel_hugsky's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/rrwwxdal) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/rrwwxdal/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/axel_hugsky') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/axel_hugsky/1624473007421/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/axel_hugsky
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Axel! ️ @axel\_hugsky I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Axel! ️. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @axel\_hugsky's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1362332889204015104/Rk5Iloej_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">lillie! 🤖 AI Bot </div> <div style="font-size: 15px">@axialcatwalk bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@axialcatwalk's tweets](https://twitter.com/axialcatwalk). | Data | Quantity | | --- | --- | | Tweets downloaded | 1153 | | Retweets | 244 | | Short tweets | 197 | | Tweets kept | 712 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/thgvv1za/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @axialcatwalk's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3mvj3lqy) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3mvj3lqy/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/axialcatwalk') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/axialcatwalk/1614103612518/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/axialcatwalk
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
lillie! AI Bot @axialcatwalk bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @axialcatwalk's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @axialcatwalk's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1216804050991108097/kOy4RwPD_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Axiom of Rock House</div> <div style="text-align: center; font-size: 14px;">@axiaofficial</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Axiom of Rock House. | Data | Axiom of Rock House | | --- | --- | | Tweets downloaded | 1455 | | Retweets | 432 | | Short tweets | 146 | | Tweets kept | 877 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1gs7ydag/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @axiaofficial's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/wa3qdf22) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/wa3qdf22/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/axiaofficial') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/axiaofficial/1627065097228/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/axiaofficial
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Axiom of Rock House @axiaofficial I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Axiom of Rock House. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @axiaofficial's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
null
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1334136134234849280/XgE0O39a_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1480842681182220288/ywam5sXK_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1480168235417083905/Kp8uyXIy_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">池澤あやか / いけあや & ちょまど🎀💻エンジニア兼漫画家 & 職業「戸倉彩」👩‍💻とくあや</div> <div style="text-align: center; font-size: 14px;">@ayatokura-chomado-ikeay</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from 池澤あやか / いけあや & ちょまど🎀💻エンジニア兼漫画家 & 職業「戸倉彩」👩‍💻とくあや. | Data | 池澤あやか / いけあや | ちょまど🎀💻エンジニア兼漫画家 | 職業「戸倉彩」👩‍💻とくあや | | --- | --- | --- | --- | | Tweets downloaded | 3250 | 3245 | 3249 | | Retweets | 224 | 717 | 1266 | | Short tweets | 2813 | 867 | 1036 | | Tweets kept | 213 | 1661 | 947 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2rhguk5h/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ayatokura-chomado-ikeay's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/34bxjwb8) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/34bxjwb8/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ayatokura-chomado-ikeay') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/ayatokura-chomado-ikeay/1642462957980/predictions.png", "widget": [{"text": "My dream is"}]}
null
huggingtweets/ayatokura-chomado-ikeay
[ "huggingtweets", "en", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #huggingtweets #en #region-us
AI CYBORG 池澤あやか / いけあや & ちょまどエンジニア兼漫画家 & 職業「戸倉彩」‍とくあや @ayatokura-chomado-ikeay I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from 池澤あやか / いけあや & ちょまどエンジニア兼漫画家 & 職業「戸倉彩」‍とくあや. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ayatokura-chomado-ikeay's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#huggingtweets #en #region-us \n" ]
[ 13 ]
[ "passage: TAGS\n#huggingtweets #en #region-us \n" ]
[ 0.06862699240446091, -0.10726692527532578, -0.010401245206594467, -0.06321337819099426, 0.08621513843536377, 0.10232581198215485, 0.027164114639163017, 0.0985046774148941, 0.13242372870445251, 0.0191008560359478, 0.14544545114040375, -0.02377847209572792, -0.059302784502506256, 0.05649275705218315, 0.013633874244987965, -0.17248544096946716, 0.03670460358262062, -0.015605193562805653, 0.026972385123372078, 0.09363088756799698, 0.021145451813936234, 0.020441683009266853, 0.0902106985449791, -0.08021455258131027, -0.1398237943649292, 0.10563203692436218, 0.07052353769540787, 0.034161631017923355, 0.12517565488815308, -0.03278198093175888, 0.135564923286438, -0.03551725298166275, -0.11858534812927246, -0.20654892921447754, 0.05404523015022278, 0.03882395476102829, -0.061763372272253036, 0.03536899387836456, 0.07869033515453339, -0.15818791091442108, 0.11689841002225876, 0.05036948621273041, 0.036260005086660385, 0.13469965755939484, -0.22866027057170868, -0.07904011756181717, -0.04602409526705742, -0.07707278430461884, 0.10253201425075531, 0.00002397169555479195, 0.009154452942311764, 0.2598714828491211, -0.1161927580833435, 0.01967962086200714, 0.12730732560157776, -0.17017430067062378, 0.01806800253689289, 0.019835561513900757, 0.08412177860736847, 0.09157158434391022, 0.011509888805449009, 0.10852907598018646, 0.0707852765917778, -0.0101486686617136, -0.1360270082950592, -0.04950163885951042, -0.009859689511358738, 0.0853719413280487, -0.019690826535224915, -0.05046107620000839, 0.17664311826229095, 0.058999452739953995, 0.010314999148249626, 0.05766073241829872, -0.05329941213130951, 0.02751901187002659, 0.006196039728820324, -0.020993784070014954, 0.018363341689109802, 0.14473530650138855, 0.01760878413915634, -0.05249418318271637, -0.13410590589046478, 0.07058989256620407, -0.18945349752902985, 0.05454850196838379, -0.03698793053627014, 0.06096231937408447, -0.258131206035614, -0.01568523794412613, -0.14322851598262787, 0.012979496270418167, 0.09651564061641693, -0.0732380747795105, 0.03764953464269638, -0.012448606081306934, -0.021792607381939888, 0.11723798513412476, 0.10804640501737595, 0.1609608232975006, -0.04320541396737099, -0.00869857519865036, -0.0065362900495529175, 0.07786694169044495, 0.09295008331537247, -0.07654742151498795, 0.07032539695501328, -0.0564110204577446, -0.04492797330021858, -0.21906612813472748, 0.0008765582460910082, -0.030310221016407013, 0.04600467532873154, -0.01984437182545662, -0.13926945626735687, 0.0724203959107399, 0.028251374140381813, 0.005772269796580076, -0.10059063136577606, 0.07909262925386429, -0.02333223633468151, -0.002073736162856221, -0.04910872504115105, -0.043945688754320145, 0.01002762746065855, 0.056697506457567215, -0.10170026123523712, 0.03527959808707237, 0.07433190196752548, 0.07395917177200317, -0.16713260114192963, -0.013432089239358902, 0.04245100915431976, 0.042995184659957886, 0.062102507799863815, -0.17763984203338623, 0.0445265956223011, -0.15618294477462769, -0.0354858934879303, -0.00870739296078682, -0.05047069117426872, 0.01355440728366375, 0.06636203825473785, -0.027222590520977974, 0.11233346909284592, 0.01111317053437233, 0.005379479844123125, -0.12045124173164368, -0.06185131147503853, 0.06667035073041916, -0.02369607612490654, 0.0511794276535511, -0.14944906532764435, 0.007654519751667976, -0.180730938911438, 0.07720775902271271, 0.016102097928524017, 0.026931453496217728, -0.026627028360962868, 0.3114912509918213, 0.030569342896342278, 0.10634830594062805, -0.22846999764442444, 0.05719701200723648, -0.023560456931591034, 0.28067547082901, -0.1065937876701355, -0.12447673082351685, 0.18407796323299408, -0.06522929668426514, -0.12309088557958603, 0.12086527794599533, 0.03488042950630188, 0.10873927175998688, 0.09767720103263855, 0.3835752606391907, -0.10195262730121613, -0.08146942406892776, 0.03636458143591881, 0.1353151649236679, -0.25350436568260193, 0.02267700806260109, 0.025416342541575432, -0.08121002465486526, -0.1325274407863617, -0.01190396212041378, 0.2367665320634842, 0.14369091391563416, -0.11477484554052353, -0.02520707994699478, 0.020064817741513252, -0.016537215560674667, 0.05091886967420578, 0.017726575955748558, -0.0013406252255663276, -0.21194274723529816, 0.02760516293346882, -0.124215267598629, 0.03498653694987297, 0.13977846503257751, 0.04585346579551697, -0.07557849586009979, -0.040234941989183426, 0.08057703077793121, -0.01661347784101963, -0.02294166386127472, -0.16728036105632782, -0.06481286138296127, 0.1498730629682541, 0.1340765357017517, 0.013609949499368668, 0.09578457474708557, -0.12554091215133667, 0.002669427078217268, 0.013310762122273445, 0.13174262642860413, -0.007498446851968765, -0.04374158754944801, -0.022994134575128555, 0.1897602677345276, -0.09339086711406708, 0.07846151292324066, -0.05010027438402176, 0.001938516623340547, 0.10550153255462646, 0.060221847146749496, 0.0357060432434082, -0.013110004365444183, 0.03018314018845558, 0.005840349942445755, 0.018294237554073334, -0.06945645809173584, 0.12700411677360535, -0.011421116068959236, -0.11599461734294891, 0.22505104541778564, -0.12537546455860138, 0.1077398806810379, 0.1641158014535904, -0.15259717404842377, -0.09176535904407501, -0.05451177433133125, -0.05600990727543831, -0.04779776558279991, 0.12334655225276947, -0.07418269664049149, -0.04411538317799568, -0.044391438364982605, -0.0011415277840569615, 0.017120715230703354, -0.07650631666183472, 0.019346725195646286, -0.08435862511396408, -0.19952528178691864, 0.0862463191151619, -0.048986468464136124, -0.15294228494167328, 0.12348992377519608, 0.3562917411327362, 0.1481546014547348, 0.23727938532829285, -0.044758010655641556, 0.017626842483878136, 0.026004822924733162, -0.0384499691426754, -0.045703496783971786, 0.004267206881195307, -0.1251041442155838, -0.0365486666560173, -0.0015835420927032828, 0.004274635575711727, 0.10203591734170914, -0.036351218819618225, -0.14834646880626678, 0.02251344546675682, 0.0014692102558910847, -0.08886273205280304, 0.10843174904584885, -0.027543172240257263, 0.07448233664035797, 0.032922543585300446, 0.03858240693807602, 0.05848072096705437, 0.002110567642375827, -0.11813180893659592, 0.006854616105556488, -0.22355535626411438, -0.2711923122406006, -0.09768662601709366, -0.08925486356019974, 0.03989091515541077, 0.041187938302755356, 0.051699478179216385, -0.18302172422409058, 0.03948478400707245, 0.003198040649294853, 0.10697315633296967, -0.06025030463933945, 0.04016250744462013, -0.06218542903661728, 0.039252810180187225, -0.026844710111618042, -0.036029912531375885, -0.0012391663622111082, -0.0970999225974083, -0.10773636400699615, 0.11318950355052948, -0.12849527597427368, 0.08555713295936584, 0.14430534839630127, 0.03159665688872337, 0.02274978905916214, -0.06426727026700974, 0.08118022978305817, -0.10766161978244781, -0.02468705177307129, -0.020854998379945755, -0.012126539833843708, 0.08575579524040222, 0.08519800752401352, 0.02189561165869236, -0.15146233141422272, 0.0591089241206646, 0.048329032957553864, -0.13319815695285797, -0.13935089111328125, -0.09409863501787186, 5.991143439132429e-7, 0.2797071039676666, -0.002339477650821209, 0.023659970611333847, 0.14668741822242737, -0.005060214549303055, 0.07289005070924759, -0.20045752823352814, -0.05620472878217697, 0.018704742193222046, -0.004834607243537903, -0.022842248901724815, -0.008816422894597054, -0.08761768788099289, 0.006385814398527145, 0.2184516340494156, -0.004971520509570837, -0.12653324007987976, 0.13448448479175568, 0.0007297596894204617, 0.0038821459747850895, 0.1352946162223816, 0.08407466113567352, 0.06298195570707321, -0.07155273109674454, -0.03599351644515991, -0.043885089457035065, 0.03730371966958046, 0.0007575029158033431, 0.01575464755296707, 0.03105960786342621, -0.22486138343811035, -0.016705667600035667, -0.2129937708377838, 0.10879084467887878, -0.04748593270778656, 0.13807302713394165, 0.0026444492395967245, 0.04441913589835167, 0.10906891524791718, -0.005030614789575338, -0.011882180348038673, 0.10863572359085083, 0.1383846551179886, -0.03977278992533684, 0.07618652284145355, 0.10392951220273972, 0.08540554344654083, 0.10903793573379517, 0.06167486310005188, -0.08492720872163773, -0.021127475425601006, -0.053188424557447433, 0.04624804109334946, -0.20438914000988007, 0.16149352490901947, -0.038125935941934586, -0.14207349717617035, 0.04278627038002014, -0.1275794953107834, 0.047217957675457, 0.02532968670129776, 0.05271764472126961, 0.09215202182531357, 0.04994546249508858, -0.08136153966188431, -0.047041911631822586, -0.009494416415691376, 0.20800645649433136, -0.04930320754647255, -0.10087979584932327, -0.015097087249159813, 0.02010302059352398, -0.0005327145336195827, 0.11119779944419861, 0.02447434328496456, -0.04610208794474602, -0.08690282702445984, 0.03303453326225281, -0.03661355376243591, 0.08763010054826736, 0.09737902879714966, -0.02698802947998047, -0.05718262121081352, 0.1235092505812645, 0.06895491480827332, -0.1251404732465744, -0.14452208578586578, 0.015589582733809948, -0.007731362245976925, -0.003907877020537853, -0.03874189779162407, -0.17162851989269257, -0.09909867495298386, -0.1906745284795761, 0.12603247165679932, -0.02163488231599331, 0.030842140316963196, -0.054747339338064194, 0.230157732963562, -0.009354943409562111, 0.009969666600227356, -0.03108477219939232, -0.013541288673877716, -0.012568399310112, -0.05987691134214401, 0.17130033671855927, -0.1589484065771103, -0.04290361702442169, 0.08715757727622986, 0.03195212036371231, 0.022129757329821587, -0.07987111806869507, -0.03223046660423279, 0.13554437458515167, 0.35851117968559265, 0.0012283556861802936, 0.2099764049053192, 0.2042219638824463, -0.05234047770500183, -0.21901710331439972, -0.09460500627756119, -0.2348603904247284, -0.07078184932470322, 0.18186907470226288, -0.07664406299591064, 0.0643012747168541, 0.12657569348812103, 0.02973685972392559, 0.20995554327964783, -0.12457278370857239, -0.044523198157548904, 0.04818897321820259, -0.08845916390419006, 0.6114599704742432, -0.061698999255895615, -0.10802759975194931, -0.01825915277004242, -0.15282532572746277, 0.2387581467628479, -0.05154171586036682, 0.013269291259348392, 0.03874128684401512, 0.1005491390824318, 0.05056579411029816, 0.05772935599088669, 0.19292889535427094, 0.02319646254181862, 0.058188073337078094, -0.13999897241592407, -0.1260925531387329, 0.10026954114437103, -0.0229759830981493, -0.08925508707761765, -0.017890285700559616, -0.02936561405658722, -0.19818171858787537, 0.05141949653625488, -0.1515510380268097, 0.05321922153234482, 0.056100163608789444, -0.02688506431877613, -0.039986640214920044, -0.002744022523984313, -0.09706974029541016, -0.04741370305418968, 0.13259218633174896, -0.08547285199165344, 0.21339218318462372, -0.08619081228971481, 0.062479544430971146, -0.1132776141166687, 0.09236282855272293, -0.05286029726266861, -0.06702502816915512, 0.020908309146761894, -0.17462939023971558, 0.04489823803305626, 0.05220690742135048, -0.02032669261097908, 0.0322212353348732, 0.03294215351343155, -0.09376110136508942, 0.04418664798140526, 0.21215519309043884, -0.20752321183681488, -0.10239655524492264, -0.032411038875579834, -0.18671981990337372, 0.19688937067985535, 0.007372637744992971, 0.09758103638887405, 0.1276085525751114, -0.02415306493639946, 0.05030597373843193, 0.02030378393828869, -0.13016556203365326, -0.019506316632032394, 0.02471683733165264, -0.04843040183186531, -0.07417872548103333, 0.16410969197750092, 0.003125022863969207, -0.17150603234767914, -0.0407160185277462, 0.33249592781066895, -0.06882669776678085, -0.021546127274632454, -0.15638038516044617, 0.06429695338010788, -0.03442022576928139, 0.031490225344896317, 0.08549392223358154, 0.021496908739209175, -0.01903107762336731, 0.243308424949646, -0.0012956787832081318, 0.20592927932739258, 0.08698756247758865, 0.02289220504462719, 0.06535287946462631, -0.1015409454703331, 0.003708842908963561, -0.019963471218943596, -0.06298170238733292, -0.025758573785424232, -0.03440312668681145, 0.12883484363555908, -0.08708789944648743, -0.08515651524066925, -0.18808633089065552, -0.00678415410220623, -0.045131731778383255, -0.14115594327449799, -0.0497581847012043, -0.09431345015764236, 0.08569512516260147, -0.01951061561703682, -0.0172411035746336, -0.13612228631973267, -0.0707104429602623, 0.06029095500707626, 0.08022084087133408, 0.056907493621110916, -0.02142960950732231, -0.005597047973424196, 0.202048197388649, 0.0028164649847894907, 0.19390766322612762, 0.20113031566143036, -0.061333067715168, 0.1175730898976326, -0.2240629494190216, -0.10820194333791733, 0.09987909346818924, -0.1013496071100235, 0.02835783362388611, 0.2528587877750397, -0.06944312155246735, -0.07183004170656204, 0.023082008585333824, 0.0819302424788475, 0.0695723444223404, -0.06962034851312637, 0.13945920765399933, 0.11123763024806976, -0.24406595528125763, -0.029789475724101067, -0.22912776470184326, 0.10466330498456955, 0.026690829545259476, 0.05581321194767952, 0.030828367918729782, 0.025830436497926712, 0.02255583181977272, -0.011340123601257801, 0.06775032728910446, -0.11715590953826904, 0.07594534754753113, -0.006308197975158691, -0.0009284214465878904, 0.020968755707144737, 0.15689221024513245, -0.06443887948989868, -0.07048701494932175, 0.06289470195770264, 0.19063851237297058, -0.0735047310590744, -0.017401263117790222, 0.032150279730558395, 0.07059439271688461, -0.10505253821611404, -0.21839775145053864, 0.07302705198526382, -0.03587287291884422, -0.19983458518981934, 0.04817315936088562, -0.014539380557835102, 0.08539435267448425, 0.02739058807492256, 0.003347607795149088, 0.06624280661344528, 0.08827538043260574, -0.12586817145347595, -0.0005465050344355404, -0.06126665323972702, 0.040976397693157196, 0.1078856810927391, 0.16836385428905487, 0.012725654058158398, 0.0595620721578598, -0.091275155544281, 0.034217145293951035, -0.04657742381095886, -0.0708092749118805, 0.03863679990172386, -0.17856280505657196, 0.032593801617622375, -0.003704255213961005, -0.02675822377204895, 0.14164531230926514, 0.08683508634567261, -0.023918727412819862, 0.011150016449391842, -0.09630842506885529, -0.10818011313676834, 0.0868176817893982, -0.0019381127785891294, 0.03816099464893341, 0.013883480802178383, -0.11632172018289566, -0.13214722275733948, 0.006640009116381407, -0.10724195092916489, 0.014073909260332584, -0.015860673040151596, -0.027241721749305725, -0.2226807177066803, -0.07590645551681519, -0.021995285525918007, 0.0819355845451355, -0.15479689836502075, -0.07021456956863403, 0.03772281855344772, 0.07989605516195297, 0.010158415883779526, 0.18902714550495148, 0.04430154338479042, 0.06323259323835373, -0.021539712324738503, -0.03570494055747986, -0.1826917976140976, 0.08993129432201385, -0.10861438512802124, -0.0066994610242545605, -0.11356137692928314, 0.20738758146762848, 0.22770558297634125, -0.08759643882513046, -0.011138317175209522, -0.12471133470535278, 0.05767860263586044, 0.08991259336471558, 0.13322731852531433, 0.04955065995454788, 0.2215200960636139, -0.053830891847610474, -0.00016504255472682416, 0.07554452121257782, 0.005791462026536465, -0.03607539087533951, -0.0017499765381217003, 0.008505070582032204, 0.0027193266432732344, -0.05850288271903992, 0.12590549886226654, -0.22327828407287598, 0.09907186031341553, 0.011895917356014252, -0.216516375541687, 0.031215285882353783, -0.07893261313438416, 0.06044485419988632, -0.01862451434135437, 0.17547211050987244, -0.0153840696439147, -0.13527122139930725, -0.1813722848892212, 0.07340814918279648, -0.3564576506614685, -0.16425473988056183, 0.04304486885666847, -0.09049394726753235, 0.024793129414319992, -0.04938378185033798, -0.12449127435684204, -0.007948409765958786, 0.016860339790582657, -0.004252702929079533, -0.008621524088084698, 0.03531380742788315, 0.07221032679080963, -0.19206926226615906, 0.0369514636695385, 0.04588461294770241, -0.18126672506332397, 0.06043267995119095, -0.12337397038936615, -0.036451902240514755, 0.07750041782855988, 0.0004883371293544769, 0.037772078067064285, 0.06528574228286743, -0.12008415907621384, 0.058698076754808426, -0.008846146054565907, -0.0022655015345662832, 0.03388579562306404, -0.040226709097623825, -0.0697655975818634, 0.037853848189115524, -0.16796614229679108, -0.0534391887485981, 0.07193274796009064, -0.07208003848791122, 0.17316791415214539, -0.08974345773458481, -0.052828285843133926, 0.02319551631808281, -0.11347067356109619, 0.09840789437294006, -0.11054719239473343, 0.1294768750667572, 0.13222336769104004, 0.03777482360601425, 0.04262825846672058, -0.22420136630535126, 0.17175815999507904, 0.10596811026334763, 0.03310242295265198, -0.06821910291910172 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1385798100430868486/QusggrjY_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Azul Crescent</div> <div style="text-align: center; font-size: 14px;">@azulcrescent</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Azul Crescent. | Data | Azul Crescent | | --- | --- | | Tweets downloaded | 1039 | | Retweets | 9 | | Short tweets | 92 | | Tweets kept | 938 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3hna2b57/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @azulcrescent's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/23ccs0az) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/23ccs0az/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/azulcrescent') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/azulcrescent/1621537893204/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/azulcrescent
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Azul Crescent @azulcrescent I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Azul Crescent. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @azulcrescent's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/637912443281272832/dAsVKNNU_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Azzam Ameen</div> <div style="text-align: center; font-size: 14px;">@azzamameen</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Azzam Ameen. | Data | Azzam Ameen | | --- | --- | | Tweets downloaded | 3247 | | Retweets | 2828 | | Short tweets | 18 | | Tweets kept | 401 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/4gs8vmx7/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @azzamameen's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1yezh3jh) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1yezh3jh/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/azzamameen') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/azzamameen/1622350298746/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/azzamameen
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Azzam Ameen @azzamameen I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Azzam Ameen. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @azzamameen's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1327803375144955904/CPYJCjKe_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Anupam Gupta 🤖 AI Bot </div> <div style="font-size: 15px">@b50 bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@b50's tweets](https://twitter.com/b50). | Data | Quantity | | --- | --- | | Tweets downloaded | 3243 | | Retweets | 438 | | Short tweets | 283 | | Tweets kept | 2522 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3sfdctj9/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @b50's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3jrmsda8) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3jrmsda8/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/b50') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/b50/1617343613656/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/b50
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Anupam Gupta AI Bot @b50 bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @b50's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @b50's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1313658315767910400/bCaV9qVB_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">BadBunny 💢</div> <div style="text-align: center; font-size: 14px;">@badbunnytwitch</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from BadBunny 💢. | Data | BadBunny 💢 | | --- | --- | | Tweets downloaded | 3246 | | Retweets | 177 | | Short tweets | 1018 | | Tweets kept | 2051 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/jrtmk7ym/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @badbunnytwitch's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/r2t5349l) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/r2t5349l/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/badbunnytwitch') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/badbunnytwitch/1643762099951/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/badbunnytwitch
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT BadBunny @badbunnytwitch I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from BadBunny . Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @badbunnytwitch's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1351393562764709889/ZS3zxvTT_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">c a l (b r b)</div> <div style="text-align: center; font-size: 14px;">@badsleepwelll</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from c a l (b r b). | Data | c a l (b r b) | | --- | --- | | Tweets downloaded | 3216 | | Retweets | 692 | | Short tweets | 533 | | Tweets kept | 1991 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3cl0ip5y/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @badsleepwelll's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/cz16cnsk) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/cz16cnsk/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/badsleepwelll') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/badsleepwelll/1621366332895/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/badsleepwelll
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT c a l (b r b) @badsleepwelll I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from c a l (b r b). Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @badsleepwelll's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/938437222948265985/NhKG-UQd_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Baidu Inc. 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@baidu_inc bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@baidu_inc's tweets](https://twitter.com/baidu_inc). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>1364</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>254</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>7</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>1103</td> </tr> </tbody> </table> [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/7tblrg78/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @baidu_inc's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/23h9pyfk) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/23h9pyfk/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/baidu_inc'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/baidu_inc/1609715579938/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/baidu_inc
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Baidu Inc. AI Bot </div> <div style="font-size: 15px; color: #657786">@baidu_inc bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @baidu_inc's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>1364</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>254</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>7</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>1103</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @baidu_inc's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/baidu_inc'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @baidu_inc's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>1364</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>254</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>7</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>1103</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @baidu_inc's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/baidu_inc'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @baidu_inc's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>1364</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>254</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>7</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>1103</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @baidu_inc's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/baidu_inc'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 431, 76, 9, 168, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1406974882919813128/LOUb2m4R_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Balaji</div> <div style="text-align: center; font-size: 14px;">@balajis</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Balaji. | Data | Balaji | | --- | --- | | Tweets downloaded | 3243 | | Retweets | 849 | | Short tweets | 54 | | Tweets kept | 2340 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/bioobb8j/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @balajis's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1iql7y69) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1iql7y69/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/balajis') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/balajis/1668967580599/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/balajis
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Balaji @balajis I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Balaji. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @balajis's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1372581144873472004/jc7AWhim_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">maddie 🤖 AI Bot </div> <div style="font-size: 15px">@balanchinarinaa bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@balanchinarinaa's tweets](https://twitter.com/balanchinarinaa). | Data | Quantity | | --- | --- | | Tweets downloaded | 3187 | | Retweets | 358 | | Short tweets | 232 | | Tweets kept | 2597 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2u50n6tw/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @balanchinarinaa's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3klfmyxg) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3klfmyxg/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/balanchinarinaa') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/balanchinarinaa/1616741385918/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/balanchinarinaa
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
maddie AI Bot @balanchinarinaa bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @balanchinarinaa's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @balanchinarinaa's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1379189650556911619/EiZklugS_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1441278016164818955/T-PDXXvg_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1438447321604313089/5_lZmeyb_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Wait Forgot & Balco - Special Boperative & Liyrex</div> <div style="text-align: center; font-size: 14px;">@balcobops-liyrex_irl-waitforgot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Wait Forgot & Balco - Special Boperative & Liyrex. | Data | Wait Forgot | Balco - Special Boperative | Liyrex | | --- | --- | --- | --- | | Tweets downloaded | 3194 | 1171 | 3189 | | Retweets | 1294 | 129 | 1587 | | Short tweets | 285 | 122 | 279 | | Tweets kept | 1615 | 920 | 1323 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/371suxoa/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @balcobops-liyrex_irl-waitforgot's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/bj54dpp8) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/bj54dpp8/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/balcobops-liyrex_irl-waitforgot') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/balcobops-liyrex_irl-waitforgot/1632888280266/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/balcobops-liyrex_irl-waitforgot
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Wait Forgot & Balco - Special Boperative & Liyrex @balcobops-liyrex\_irl-waitforgot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Wait Forgot & Balco - Special Boperative & Liyrex. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @balcobops-liyrex\_irl-waitforgot's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1349590876931072001/kmtebQA4_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Banjodogg 🤖 AI Bot </div> <div style="font-size: 15px">@banjocatt bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@banjocatt's tweets](https://twitter.com/banjocatt). | Data | Quantity | | --- | --- | | Tweets downloaded | 1384 | | Retweets | 27 | | Short tweets | 150 | | Tweets kept | 1207 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3u0ms4j5/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @banjocatt's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2me74qbw) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2me74qbw/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/banjocatt') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/banjocatt
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Banjodogg AI Bot @banjocatt bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @banjocatt's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @banjocatt's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/988775660163252226/XpgonN0X_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1329647526807543809/2SGvnHYV_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/988775660163252226/XpgonN0X_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Bill Gates & Barack Obama</div> <div style="text-align: center; font-size: 14px;">@barackobama-billgates</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Bill Gates & Barack Obama. | Data | Bill Gates | Barack Obama | | --- | --- | --- | | Tweets downloaded | 3207 | 3214 | | Retweets | 186 | 353 | | Short tweets | 6 | 25 | | Tweets kept | 3015 | 2836 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/7wwgduen/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @barackobama-billgates's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1afk1edu) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1afk1edu/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/barackobama-billgates') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/barackobama-billgates/1620428621185/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/barackobama-billgates
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Bill Gates & Barack Obama @barackobama-billgates I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Bill Gates & Barack Obama. Data: Tweets downloaded, Bill Gates: 3207, Barack Obama: 3214 Data: Retweets, Bill Gates: 186, Barack Obama: 353 Data: Short tweets, Bill Gates: 6, Barack Obama: 25 Data: Tweets kept, Bill Gates: 3015, Barack Obama: 2836 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @barackobama-billgates's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1383184766959120385/MM9DHPWC_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1329647526807543809/2SGvnHYV_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Elon Musk & Barack Obama</div> <div style="text-align: center; font-size: 14px;">@barackobama-elonmusk</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Elon Musk & Barack Obama. | Data | Elon Musk | Barack Obama | | --- | --- | --- | | Tweets downloaded | 1021 | 3234 | | Retweets | 52 | 353 | | Short tweets | 268 | 25 | | Tweets kept | 701 | 2856 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/osj70w7o/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @barackobama-elonmusk's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/21huy5w1) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/21huy5w1/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/barackobama-elonmusk') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/barackobama-elonmusk
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Elon Musk & Barack Obama @barackobama-elonmusk I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Elon Musk & Barack Obama. Data: Tweets downloaded, Elon Musk: 1021, Barack Obama: 3234 Data: Retweets, Elon Musk: 52, Barack Obama: 353 Data: Short tweets, Elon Musk: 268, Barack Obama: 25 Data: Tweets kept, Elon Musk: 701, Barack Obama: 2856 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @barackobama-elonmusk's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1263334523187671040/_hCX0Xmb_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1396877840763719684/88N2DjSH_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1329647526807543809/2SGvnHYV_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">University of OnFoeNem 🇭🇹 & Karlous & Barack Obama</div> <div style="text-align: center; font-size: 14px;">@barackobama-karlousm-uofofn</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from University of OnFoeNem 🇭🇹 & Karlous & Barack Obama. | Data | University of OnFoeNem 🇭🇹 | Karlous | Barack Obama | | --- | --- | --- | --- | | Tweets downloaded | 3153 | 3207 | 3249 | | Retweets | 457 | 1736 | 352 | | Short tweets | 576 | 175 | 25 | | Tweets kept | 2120 | 1296 | 2872 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3ghcjele/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @barackobama-karlousm-uofofn's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/24cklty4) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/24cklty4/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/barackobama-karlousm-uofofn') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/barackobama-karlousm-uofofn/1625032946203/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/barackobama-karlousm-uofofn
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG University of OnFoeNem 🇭🇹 & Karlous & Barack Obama @barackobama-karlousm-uofofn I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from University of OnFoeNem 🇭🇹 & Karlous & Barack Obama. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @barackobama-karlousm-uofofn's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1329647526807543809/2SGvnHYV_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Barack Obama</div> <div style="text-align: center; font-size: 14px;">@barackobama</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Barack Obama. | Data | Barack Obama | | --- | --- | | Tweets downloaded | 3250 | | Retweets | 319 | | Short tweets | 17 | | Tweets kept | 2914 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/tjijp9lf/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @barackobama's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2jfeof85) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2jfeof85/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/barackobama') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/barackobama
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
AI BOT Barack Obama @barackobama I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Barack Obama. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @barackobama's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 58 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 0.01490766927599907, -0.029707439243793488, -0.005551214329898357, 0.019200731068849564, 0.13538123667240143, 0.031960584223270416, 0.07854003459215164, 0.15176640450954437, -0.03478027135133743, 0.020825274288654327, 0.17780765891075134, 0.13369141519069672, -0.025615064427256584, 0.10110418498516083, -0.04495277628302574, -0.27562591433525085, 0.05076512694358826, 0.04804648086428642, -0.03660755604505539, 0.13914404809474945, 0.08130049705505371, -0.04002818092703819, 0.09946437180042267, -0.02348330244421959, -0.18448196351528168, 0.03603345900774002, 0.03874115273356438, -0.10113959014415741, 0.1245182529091835, 0.042210761457681656, 0.097016341984272, 0.02144498936831951, -0.08173582702875137, -0.09949471801519394, 0.04076211526989937, 0.0524214468896389, -0.06747427582740784, 0.07371947914361954, 0.06548287719488144, -0.09104428440332413, 0.15018822252750397, 0.032346583902835846, -0.015613866969943047, 0.05642572045326233, -0.17452795803546906, -0.05158021301031113, -0.038718461990356445, 0.007089211139827967, 0.026932736858725548, 0.0765126645565033, -0.03151388466358185, 0.1801837533712387, -0.09969375282526016, 0.08080597966909409, 0.19449612498283386, -0.3136744201183319, -0.008555608801543713, 0.09032562375068665, 0.10929754376411438, 0.04874800145626068, -0.027296993881464005, 0.0915885642170906, 0.06193246319890022, 0.01960965059697628, 0.03833269700407982, -0.04861060902476311, -0.10319411009550095, 0.05647158622741699, -0.09057577699422836, -0.06054073944687843, 0.21635963022708893, -0.04008340463042259, 0.06982671469449997, -0.06175680458545685, -0.10815948247909546, -0.04076464846730232, -0.0019450898980721831, 0.007722846698015928, -0.04665110260248184, 0.07648039609193802, -0.018055662512779236, -0.07031384110450745, -0.16104750335216522, 0.004024024587124586, -0.19634990394115448, 0.15040342509746552, -0.007091521751135588, 0.03826533630490303, -0.18292304873466492, 0.11239926517009735, -0.00463539082556963, -0.09352986514568329, 0.05768691375851631, -0.09079932421445847, 0.06515228003263474, 0.012864203192293644, -0.07714717835187912, -0.01476553175598383, 0.07529455423355103, 0.14459288120269775, -0.03202451393008232, -0.014661530964076519, 0.005731707438826561, 0.08028950542211533, 0.0649947077035904, 0.0529506653547287, -0.05044480413198471, -0.04604479670524597, 0.023471293970942497, -0.12279047816991806, 0.005323461256921291, -0.08722265064716339, -0.11850544810295105, -0.06324991583824158, 0.02913680486381054, 0.0409526452422142, 0.05413410812616348, 0.11646701395511627, -0.03819722682237625, 0.003125704126432538, 0.0790749341249466, -0.04754907265305519, 0.014819944277405739, -0.019563961774110794, 0.019939353689551353, 0.10169259458780289, -0.004264532588422298, 0.02739645354449749, -0.08098349720239639, 0.053036682307720184, -0.10909318923950195, -0.01619906909763813, -0.01401914469897747, -0.08360559493303299, 0.03495157137513161, -0.13044656813144684, 0.003459199797362089, -0.1794617623090744, -0.08515556156635284, 0.004473025444895029, -0.02669714391231537, -0.029448989778757095, -0.07044726610183716, -0.000012087725735909771, -0.03782980516552925, 0.09037511795759201, -0.049331095069646835, 0.02187872678041458, -0.05999310687184334, 0.10105309635400772, -0.06390658020973206, 0.10026112198829651, -0.14347022771835327, 0.062101077288389206, -0.13453714549541473, -0.007185438182204962, -0.056844428181648254, 0.06125464290380478, 0.003428247757256031, 0.14510339498519897, -0.0019743351731449366, -0.024058373644948006, -0.11893454194068909, 0.07660667598247528, -0.013276024721562862, 0.20594745874404907, -0.08867699652910233, -0.12363981455564499, 0.18977078795433044, -0.05639764666557312, -0.13151882588863373, 0.12418028712272644, 0.014429234899580479, 0.0632154569029808, 0.06213442608714104, 0.22679246962070465, 0.019381923601031303, -0.012931653298437595, 0.022296493873000145, 0.09503678232431412, -0.14555561542510986, -0.009661003015935421, 0.00841028243303299, -0.0034040703903883696, -0.08606491982936859, 0.03532338887453079, 0.11260384321212769, 0.09308808296918869, -0.06410733610391617, -0.02130906470119953, -0.04973118007183075, -0.002825164934620261, 0.08984078466892242, -0.004783526994287968, 0.10731708258390427, -0.1087048277258873, -0.0696249008178711, -0.030250849202275276, -0.0012943379115313292, 0.019239740446209908, 0.047808386385440826, -0.02344650961458683, 0.12562713027000427, -0.007734235376119614, 0.03894360736012459, -0.1425095796585083, -0.08136747777462006, -0.0352344736456871, 0.1538124978542328, 0.043824564665555954, 0.12064811587333679, 0.05824866518378258, -0.05751828849315643, -0.01182011142373085, -0.0009749550954438746, 0.1436033844947815, -0.01271585002541542, -0.0852496474981308, -0.05857346951961517, 0.08234002441167831, -0.07005385309457779, 0.025747623294591904, -0.044775284826755524, 0.034807976335287094, 0.07559926062822342, 0.12458331137895584, -0.012196341529488564, 0.03711283579468727, -0.0114806042984128, -0.00011358146002748981, -0.0896436795592308, -0.017758425325155258, 0.08426769077777863, -0.01031328085809946, -0.06891145557165146, 0.2548580765724182, -0.20196770131587982, 0.2299574315547943, 0.2296006977558136, -0.24926680326461792, -0.031021486967802048, -0.033769410103559494, -0.053823258727788925, 0.017695831134915352, 0.037970591336488724, -0.06758655607700348, 0.04836089164018631, -0.04798879474401474, 0.14743368327617645, -0.039952509105205536, -0.067154660820961, 0.011684614233672619, -0.06448059529066086, -0.05935867503285408, 0.06313008069992065, 0.042497556656599045, -0.12196247279644012, 0.19907300174236298, 0.256646990776062, 0.05731065943837166, 0.19537882506847382, 0.018505526706576347, -0.008647909387946129, 0.05532584711909294, -0.06024787575006485, -0.05939318984746933, -0.047756556421518326, -0.17080695927143097, -0.04192587733268738, 0.08264796435832977, 0.03818074241280556, 0.10142038762569427, -0.10767655074596405, -0.07215311378240585, 0.009001186117529869, 0.015298610553145409, -0.0071457610465586185, 0.13598360121250153, 0.0518062487244606, 0.13914446532726288, -0.005930832587182522, 0.025881925597786903, 0.06802397966384888, 0.02908273972570896, -0.08634071797132492, 0.1456945538520813, -0.1443340927362442, -0.3761582672595978, -0.1337488889694214, -0.09692523628473282, -0.01728254184126854, 0.04180695861577988, 0.12187701463699341, -0.12770286202430725, 0.0015871572541072965, -0.038519375026226044, 0.11093265563249588, -0.08619660884141922, 0.05064225569367409, -0.09312444925308228, 0.013121986761689186, -0.05371590703725815, -0.09208092838525772, -0.03968694806098938, -0.025510141626000404, -0.09566790610551834, 0.17000380158424377, -0.06683114916086197, 0.06858081370592117, 0.18506819009780884, 0.00719145592302084, 0.0289844311773777, -0.048802170902490616, 0.19252337515354156, -0.11087619513273239, 0.03966445475816727, 0.16040337085723877, 0.017642250284552574, 0.0870000347495079, 0.1063152626156807, -0.012242939323186874, -0.0729413852095604, 0.045604314655065536, 0.00838512647897005, -0.1192161962389946, -0.1628645360469818, -0.12800830602645874, -0.09412717819213867, 0.11832976341247559, 0.04896165430545807, 0.06800009310245514, 0.15884871780872345, 0.07476164400577545, -0.026399539783596992, -0.014989175833761692, -0.030628371983766556, 0.06698833405971527, 0.1810096949338913, -0.03410731256008148, 0.14575767517089844, -0.058808114379644394, -0.11332095414400101, 0.13757702708244324, 0.02930915355682373, 0.041703663766384125, 0.02182352915406227, -0.00909164547920227, 0.002967006294056773, 0.1275966912508011, 0.12905162572860718, 0.09085892140865326, -0.009210709482431412, -0.02600487507879734, -0.043374914675951004, -0.013961076736450195, -0.004321942571550608, 0.0544881634414196, 0.04976500943303108, -0.17056837677955627, -0.05252770707011223, -0.15036118030548096, 0.10635104030370712, 0.0658026859164238, 0.1097705140709877, -0.1943172961473465, -0.002813685918226838, 0.07878942042589188, -0.032336506992578506, -0.10688291490077972, 0.0711626410484314, 0.08084181696176529, -0.10172833502292633, 0.061861153692007065, -0.006532561499625444, 0.10406243056058884, 0.023742493242025375, 0.09878388792276382, -0.05866028368473053, -0.06148262321949005, -0.016518592834472656, 0.0939461812376976, -0.2926773428916931, 0.18179792165756226, -0.02907240390777588, -0.09844435751438141, -0.07134795933961868, -0.027840720489621162, 0.02770385891199112, 0.08060616999864578, 0.07026281952857971, 0.03338824585080147, 0.004331306088715792, -0.096750907599926, -0.025082198902964592, 0.01906721480190754, 0.12366755306720734, -0.05625581368803978, -0.01350511983036995, -0.041342347860336304, 0.02173074521124363, 0.00881099235266447, 0.06707418709993362, 0.021282022818922997, -0.16175130009651184, 0.07081238180398941, 0.03874411806464195, 0.03632965683937073, 0.02809896133840084, -0.02992626093327999, -0.1548270583152771, 0.1699383705854416, 0.02149343490600586, -0.06355534493923187, -0.12458667904138565, -0.04415702819824219, 0.06474760919809341, -0.04559854418039322, 0.039427343755960464, -0.0629853680729866, -0.001072446582838893, -0.07396374642848969, -0.19678819179534912, 0.1352648138999939, -0.06877507269382477, -0.07956159859895706, -0.03609750047326088, 0.17941854894161224, -0.07844837009906769, 0.018584884703159332, 0.007694936357438564, 0.047691553831100464, -0.14645716547966003, -0.11046357452869415, 0.06859224289655685, -0.04709484800696373, 0.036693572998046875, 0.01622764952480793, -0.02127574197947979, 0.005549534223973751, -0.012371892109513283, -0.0025305403396487236, 0.27376097440719604, 0.22965525090694427, -0.0802733525633812, 0.17710179090499878, 0.0630764365196228, -0.0507618673145771, -0.326980859041214, -0.09341348707675934, -0.14117297530174255, -0.009823985397815704, 0.023037875071167946, -0.13106577098369598, 0.05048951879143715, 0.029267780482769012, -0.014271927997469902, 0.11641538888216019, -0.20954635739326477, -0.09825599938631058, 0.09716890752315521, -0.06865759193897247, 0.4034286141395569, -0.1301247775554657, -0.07252515107393265, -0.03537401184439659, -0.14370368421077728, 0.1951674520969391, -0.06888531148433685, 0.09445398300886154, -0.02521556057035923, 0.13606955111026764, 0.05328543111681938, -0.02234726957976818, 0.11178411543369293, -0.01972258649766445, 0.0011309145484119654, -0.13865596055984497, -0.0604546032845974, 0.08001494407653809, -0.007717274595052004, 0.009300438687205315, -0.10272051393985748, 0.02493043802678585, -0.1687590479850769, -0.0011172585655003786, -0.11747178435325623, 0.09020119160413742, 0.021906770765781403, -0.06782494485378265, -0.050800591707229614, -0.0397600457072258, -0.009170935489237309, -0.01781470514833927, 0.14266571402549744, -0.05449723079800606, 0.1958875209093094, 0.07874593138694763, 0.06835102289915085, -0.1398908495903015, 0.05150968208909035, -0.027329890057444572, -0.0722222700715065, 0.07193705439567566, -0.1763947606086731, 0.04665200039744377, 0.08747349679470062, -0.03685600683093071, 0.04361190274357796, 0.09054525941610336, 0.0016616116045042872, 0.015259911306202412, 0.17542323470115662, -0.26249444484710693, -0.009265073575079441, -0.05851370841264725, -0.07754334062337875, 0.09399785101413727, 0.05022705718874931, 0.17302869260311127, 0.01883954182267189, -0.042071253061294556, 0.01570643112063408, 0.01509455218911171, -0.059037789702415466, 0.045513592660427094, 0.02684464305639267, 0.008570663630962372, -0.1221366673707962, 0.05156322568655014, 0.023951835930347443, -0.13511396944522858, 0.028608962893486023, 0.18224509060382843, -0.11025255918502808, -0.12385176122188568, -0.05918734520673752, 0.014604943804442883, -0.1147887259721756, 0.022507676854729652, -0.0024671610444784164, -0.09835006296634674, 0.062000639736652374, 0.13879112899303436, 0.06231173500418663, 0.12207856774330139, -0.016749046742916107, -0.009889025241136551, -0.02596008963882923, -0.03651546314358711, 0.030208537355065346, 0.03168511390686035, -0.0861181765794754, 0.11346250027418137, -0.02512839064002037, 0.1417820155620575, -0.10664749890565872, -0.05201685056090355, -0.14300444722175598, -0.032158151268959045, -0.08575893938541412, -0.11310356855392456, -0.07854323089122772, -0.07035478204488754, 0.009079275652766228, -0.06144321709871292, -0.04742341861128807, -0.07850945740938187, -0.1071249470114708, 0.0065527972765266895, -0.04037778079509735, 0.04593488946557045, -0.08211231976747513, -0.0025119397323578596, 0.11637649685144424, -0.022535819560289383, 0.16392095386981964, 0.11117493361234665, -0.09679865092039108, 0.07616551965475082, -0.12686336040496826, -0.10472826659679413, 0.09312791377305984, -0.011926734820008278, 0.03812364861369133, 0.11054535210132599, 0.005196572281420231, 0.026585882529616356, 0.05407439172267914, 0.06366553902626038, 0.03967318683862686, -0.10743927955627441, 0.0886315256357193, -0.020753242075443268, -0.1556011289358139, -0.045647621154785156, -0.0805148333311081, 0.03534896299242973, 0.01974404975771904, 0.10271792113780975, -0.034986238926649094, 0.0667109340429306, -0.08902619034051895, 0.024180470034480095, -0.000016180018064915203, -0.17815378308296204, -0.01795545034110546, -0.044853758066892624, 0.03495967015624046, 0.017773056402802467, 0.24157072603702545, 0.06085042282938957, -0.06760483235120773, 0.04657021537423134, 0.12547647953033447, 0.015325317159295082, 0.0012204478261992335, 0.15906299650669098, 0.09267544746398926, -0.07720284163951874, -0.1355191469192505, 0.06737678498029709, 0.019282789900898933, -0.045628830790519714, 0.09535399079322815, 0.009011821821331978, 0.021824736148118973, 0.0726994201540947, -0.01307250652462244, 0.0016560814110562205, -0.10145208984613419, -0.11841893196105957, -0.038609180599451065, 0.056700821965932846, -0.03823903948068619, 0.08270184695720673, 0.1624554693698883, -0.01529943011701107, 0.038163430988788605, -0.03936288133263588, -0.01488962396979332, -0.12554575502872467, -0.1426725685596466, -0.06450272351503372, -0.16302600502967834, -0.007763232104480267, -0.08333444595336914, 0.06433288753032684, 0.07949802279472351, 0.059826187789440155, -0.04269853234291077, 0.07651268690824509, 0.041022781282663345, -0.10034794360399246, 0.07510355859994888, -0.02059151791036129, 0.05323530361056328, -0.021385202184319496, -0.019656827673316002, -0.10213424265384674, 0.05417940020561218, -0.020723329856991768, 0.0416044183075428, -0.05111391469836235, 0.009476713836193085, -0.17171601951122284, -0.1145220696926117, -0.06269708275794983, 0.05981360003352165, -0.05057463422417641, 0.05945824086666107, 0.01168898493051529, 0.011022492311894894, 0.016294511035084724, 0.24577881395816803, -0.06521986424922943, -0.04237864539027214, -0.039796773344278336, 0.1401647925376892, -0.010960960760712624, 0.07910837978124619, -0.04971875622868538, -0.026874123141169548, -0.11801496148109436, 0.31808507442474365, 0.3330231308937073, -0.09461544454097748, 0.042824193835258484, -0.011730561032891273, 0.03255566582083702, 0.12393459677696228, 0.12263859808444977, 0.1047649085521698, 0.23149316012859344, -0.07330518960952759, -0.04662523791193962, -0.0235530287027359, -0.011487046256661415, -0.07061693072319031, 0.09144455194473267, 0.024042300879955292, -0.06604029983282089, -0.04112962633371353, 0.061292614787817, -0.1968756765127182, 0.09409318119287491, -0.08292321115732193, -0.2009436935186386, -0.043561115860939026, 0.03980926424264908, 0.11663878709077835, -0.0058410209603607655, 0.12327833473682404, 0.0021221975330263376, -0.06928802281618118, 0.0007765711052343249, 0.021609993651509285, -0.21508292853832245, 0.024919532239437103, 0.07233009487390518, -0.12181923538446426, -0.003256353549659252, -0.03614840656518936, -0.0016090048011392355, 0.08110456168651581, 0.041294027119874954, -0.03950197249650955, 0.026108399033546448, 0.004547862336039543, -0.03431681916117668, -0.023192770779132843, 0.04429425671696663, 0.04074658825993538, -0.15583069622516632, 0.0852603018283844, -0.15913386642932892, 0.045099467039108276, -0.033806633204221725, 0.0012128398520871997, -0.010232296772301197, -0.0017561162821948528, -0.040829166769981384, 0.08149882405996323, 0.07333235442638397, -0.022076698020100594, -0.0061705452390015125, -0.06384896486997604, -0.055786989629268646, -0.027778856456279755, -0.08865642547607422, -0.09847963601350784, -0.12417391687631607, -0.10101866722106934, 0.09827395528554916, -0.022276965901255608, -0.17602279782295227, 0.008451612666249275, -0.09465660154819489, 0.06178250536322594, -0.14477479457855225, 0.11634210497140884, 0.08996126055717468, 0.0033104352187365294, 0.006666821893304586, -0.021679235622286797, 0.07054676860570908, 0.1143367812037468, -0.09097250550985336, -0.06963858008384705 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1362428223611670531/c08rJGpf_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">𓅅 zack 𓅅 🤖 AI Bot </div> <div style="font-size: 15px">@barzoople bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@barzoople's tweets](https://twitter.com/barzoople). | Data | Quantity | | --- | --- | | Tweets downloaded | 1787 | | Retweets | 162 | | Short tweets | 225 | | Tweets kept | 1400 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/32bz3nyh/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @barzoople's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/ezszw7hb) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/ezszw7hb/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/barzoople') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/barzoople/1614101635102/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/barzoople
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
𓅅 zack 𓅅 AI Bot @barzoople bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @barzoople's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @barzoople's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1333015562704785412/cYcTBrHF_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Basedgamerboi 🤖 AI Bot </div> <div style="font-size: 15px">@basedgamerboi bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@basedgamerboi's tweets](https://twitter.com/basedgamerboi). | Data | Quantity | | --- | --- | | Tweets downloaded | 3186 | | Retweets | 1305 | | Short tweets | 603 | | Tweets kept | 1278 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/387726ql/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @basedgamerboi's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/36dqi3qi) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/36dqi3qi/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/basedgamerboi') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/basedgamerboi/1618147489025/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/basedgamerboi
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Basedgamerboi AI Bot @basedgamerboi bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @basedgamerboi's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @basedgamerboi's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1372264552029679619/liCuKg_U_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Mel Andrews 🤖 AI Bot </div> <div style="font-size: 15px">@bayesianboy bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bayesianboy's tweets](https://twitter.com/bayesianboy). | Data | Quantity | | --- | --- | | Tweets downloaded | 3244 | | Retweets | 216 | | Short tweets | 589 | | Tweets kept | 2439 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3shc5anf/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bayesianboy's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3rrpzuts) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3rrpzuts/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bayesianboy') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bayesianboy/1616641444174/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bayesianboy
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Mel Andrews AI Bot @bayesianboy bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @bayesianboy's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bayesianboy's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1358510866371661830/rxzOoe9A_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1073647682487410688/2yrbD4RY_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1334065878917390338/V6Eh8ZJn_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Miss Gbadamosi ♠ & ♠Jenny Summers♠ & ♠️MsWhite♠️</div> <div style="text-align: center; font-size: 14px;">@bbcqos-fitslut63-kellyg_official</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Miss Gbadamosi ♠ & ♠Jenny Summers♠ & ♠️MsWhite♠️. | Data | Miss Gbadamosi ♠ | ♠Jenny Summers♠ | ♠️MsWhite♠️ | | --- | --- | --- | --- | | Tweets downloaded | 480 | 882 | 3063 | | Retweets | 117 | 55 | 1391 | | Short tweets | 154 | 483 | 230 | | Tweets kept | 209 | 344 | 1442 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3rzzq99i/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bbcqos-fitslut63-kellyg_official's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/xd2e2hom) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/xd2e2hom/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bbcqos-fitslut63-kellyg_official') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bbcqos-fitslut63-kellyg_official/1629198375751/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bbcqos-fitslut63-kellyg_official
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Miss Gbadamosi & Jenny Summers & ️MsWhite️ @bbcqos-fitslut63-kellyg\_official I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Miss Gbadamosi & Jenny Summers & ️MsWhite️. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bbcqos-fitslut63-kellyg\_official's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1073647682487410688/2yrbD4RY_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">♠Jenny Summers♠</div> <div style="text-align: center; font-size: 14px;">@bbcqos</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from ♠Jenny Summers♠. | Data | ♠Jenny Summers♠ | | --- | --- | | Tweets downloaded | 882 | | Retweets | 55 | | Short tweets | 483 | | Tweets kept | 344 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2uwts9v5/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bbcqos's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1shy0ous) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1shy0ous/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bbcqos') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bbcqos/1629197549972/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bbcqos
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Jenny Summers @bbcqos I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Jenny Summers. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bbcqos's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1310015308833452032/VWjqd52U_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Benjamin Dreyer 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@bcdreyer bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bcdreyer's tweets](https://twitter.com/bcdreyer). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3213</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>594</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>410</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2209</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/2r28be7y/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bcdreyer's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/1uihefwp) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/1uihefwp/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/bcdreyer'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets) <!--- random size file -->
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bcdreyer/1601263745684/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bcdreyer
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Benjamin Dreyer AI Bot </div> <div style="font-size: 15px; color: #657786">@bcdreyer bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @bcdreyer's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3213</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>594</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>410</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2209</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @bcdreyer's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/bcdreyer'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @bcdreyer's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3213</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>594</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>410</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2209</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @bcdreyer's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/bcdreyer'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @bcdreyer's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3213</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>594</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>410</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2209</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @bcdreyer's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/bcdreyer'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 430, 75, 9, 167, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1440017111531855879/A4p6F07H_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1440481469231558659/ZjEcoltA_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1435265846436409346/yAV2qzDs_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">6529 & Beanie & Loopify 🧙‍♂️</div> <div style="text-align: center; font-size: 14px;">@beaniemaxi-loopifyyy-punk6529</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from 6529 & Beanie & Loopify 🧙‍♂️. | Data | 6529 | Beanie | Loopify 🧙‍♂️ | | --- | --- | --- | --- | | Tweets downloaded | 3249 | 3250 | 3249 | | Retweets | 939 | 391 | 179 | | Short tweets | 525 | 559 | 1194 | | Tweets kept | 1785 | 2300 | 1876 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1ejmosjg/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beaniemaxi-loopifyyy-punk6529's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/15k8d8xn) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/15k8d8xn/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beaniemaxi-loopifyyy-punk6529') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beaniemaxi-loopifyyy-punk6529
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG 6529 & Beanie & Loopify ‍️ @beaniemaxi-loopifyyy-punk6529 I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from 6529 & Beanie & Loopify ‍️. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beaniemaxi-loopifyyy-punk6529's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1087417311361789958/Y1NvpnuX_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Dave Davies 🤖 AI Bot </div> <div style="font-size: 15px">@beanstalkim bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@beanstalkim's tweets](https://twitter.com/beanstalkim). | Data | Quantity | | --- | --- | | Tweets downloaded | 3214 | | Retweets | 121 | | Short tweets | 269 | | Tweets kept | 2824 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/62fffrcn/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beanstalkim's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3ulhjijc) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3ulhjijc/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beanstalkim') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/beanstalkim/1619112801821/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beanstalkim
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Dave Davies AI Bot @beanstalkim bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @beanstalkim's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beanstalkim's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1342942616195895303/13ml_Pdw_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Bee Boileau 🤖 AI Bot </div> <div style="font-size: 15px">@beeboileau bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@beeboileau's tweets](https://twitter.com/beeboileau). | Data | Quantity | | --- | --- | | Tweets downloaded | 238 | | Retweets | 26 | | Short tweets | 18 | | Tweets kept | 194 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/gjcxmczw/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beeboileau's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2antyc8f) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2antyc8f/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beeboileau') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/beeboileau/1616684430388/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beeboileau
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Bee Boileau AI Bot @beeboileau bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @beeboileau's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beeboileau's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/498860581072273408/q5v6iWVw_400x400.jpeg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Bee Movie Script</div> <div style="text-align: center; font-size: 14px;">@beemoviescript</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Bee Movie Script. | Data | Bee Movie Script | | --- | --- | | Tweets downloaded | 1427 | | Retweets | 0 | | Short tweets | 169 | | Tweets kept | 1258 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/291me6fz/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beemoviescript's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3gtdvdf3) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3gtdvdf3/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beemoviescript') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beemoviescript
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Bee Movie Script @beemoviescript I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Bee Movie Script. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beemoviescript's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1429115399975497731/JZdA725e_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1434240567001636864/BkVzkg7C_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1434228331315187712/IrO7AP6L_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">✨たち Tommy’s an Orbit 🌙 たち✨ & Goose & c!tubbo + glatt</div> <div style="text-align: center; font-size: 14px;">@beesforbo-cafe_orbitinnit-weebbutt</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from ✨たち Tommy’s an Orbit 🌙 たち✨ & Goose & c!tubbo + glatt. | Data | ✨たち Tommy’s an Orbit 🌙 たち✨ | Goose | c!tubbo + glatt | | --- | --- | --- | --- | | Tweets downloaded | 2241 | 3243 | 3242 | | Retweets | 1335 | 511 | 108 | | Short tweets | 323 | 512 | 1198 | | Tweets kept | 583 | 2220 | 1936 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/p0uk28zi/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beesforbo-cafe_orbitinnit-weebbutt's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/310986pt) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/310986pt/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beesforbo-cafe_orbitinnit-weebbutt') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/beesforbo-cafe_orbitinnit-weebbutt/1630941920455/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beesforbo-cafe_orbitinnit-weebbutt
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG たち Tommy’s an Orbit たち & Goose & c!tubbo + glatt @beesforbo-cafe\_orbitinnit-weebbutt I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from たち Tommy’s an Orbit たち & Goose & c!tubbo + glatt. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beesforbo-cafe\_orbitinnit-weebbutt's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1323398842268684289/_OAtj-z-_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Cate 🤖 AI Bot </div> <div style="font-size: 15px">@beetleboxes bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@beetleboxes's tweets](https://twitter.com/beetleboxes). | Data | Quantity | | --- | --- | | Tweets downloaded | 2817 | | Retweets | 883 | | Short tweets | 404 | | Tweets kept | 1530 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/16fa37go/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beetleboxes's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2sjz20j3) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2sjz20j3/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beetleboxes') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/beetleboxes/1616645225701/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beetleboxes
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Cate AI Bot @beetleboxes bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @beetleboxes's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beetleboxes's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1404753773939990533/2Ol60_sO_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Mom</div> <div style="text-align: center; font-size: 14px;">@behemilf</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Mom. | Data | Mom | | --- | --- | | Tweets downloaded | 3241 | | Retweets | 858 | | Short tweets | 346 | | Tweets kept | 2037 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/34zvujdl/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @behemilf's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/ss8n55dy) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/ss8n55dy/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/behemilf') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/behemilf
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Mom @behemilf I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Mom. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @behemilf's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1348756593052176385/TjNU6-T__400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Evan (master saucier) 🤖 AI Bot </div> <div style="font-size: 15px">@beingandslime bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@beingandslime's tweets](https://twitter.com/beingandslime). | Data | Quantity | | --- | --- | | Tweets downloaded | 3245 | | Retweets | 55 | | Short tweets | 473 | | Tweets kept | 2717 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2hj6ebde/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beingandslime's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2vtowykv) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2vtowykv/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beingandslime') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/beingandslime/1616648200015/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beingandslime
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Evan (master saucier) AI Bot @beingandslime bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @beingandslime's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beingandslime's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1365987365525815304/uxdWurnN_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Anarcho-Moses 🐍 🤖 AI Bot </div> <div style="font-size: 15px">@ben_r_hoffman bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@ben_r_hoffman's tweets](https://twitter.com/ben_r_hoffman). | Data | Quantity | | --- | --- | | Tweets downloaded | 3247 | | Retweets | 107 | | Short tweets | 264 | | Tweets kept | 2876 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/vlvpdufz/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @ben_r_hoffman's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2nf4hyti) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2nf4hyti/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/ben_r_hoffman') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/ben_r_hoffman/1618455389168/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/ben_r_hoffman
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Anarcho-Moses AI Bot @ben\_r\_hoffman bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @ben\_r\_hoffman's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @ben\_r\_hoffman's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1045023385816686592/7wIqU8ZY_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Ben Chestnut</div> <div style="text-align: center; font-size: 14px;">@benchestnut</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Ben Chestnut. | Data | Ben Chestnut | | --- | --- | | Tweets downloaded | 3229 | | Retweets | 943 | | Short tweets | 124 | | Tweets kept | 2162 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/lyrugs4q/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @benchestnut's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2u96gtbs) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2u96gtbs/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/benchestnut') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/benchestnut/1626453250687/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benchestnut
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Ben Chestnut @benchestnut I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Ben Chestnut. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @benchestnut's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<link rel="stylesheet" href="https://unpkg.com/@tailwindcss/[email protected]/dist/typography.min.css"> <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('http://pbs.twimg.com/profile_images/845396356579508224/1ruiOPLI_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Benedict Evans 🤖 AI Bot </div> <div style="font-size: 15px; color: #657786">@benedictevans bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@benedictevans's tweets](https://twitter.com/benedictevans). <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3236</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>662</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>174</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2400</td> </tr> </tbody> </table> [Explore the data](https://app.wandb.ai/wandb/huggingtweets/runs/3an7r50z/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @benedictevans's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://app.wandb.ai/wandb/huggingtweets/runs/1k099xll) for full transparency and reproducibility. At the end of training, [the final model](https://app.wandb.ai/wandb/huggingtweets/runs/1k099xll/artifacts) is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/benedictevans'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/borisdayma) <section class='prose'> For more details, visit the project repository. </section> [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo_share.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benedictevans
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
<link rel="stylesheet" href="URL <style> @media (prefers-color-scheme: dark) { .prose { color: #E2E8F0 !important; } .prose h2, .prose h3, .prose a, .prose thead { color: #F7FAFC !important; } } </style> <section class='prose'> <div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('URL </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Benedict Evans AI Bot </div> <div style="font-size: 15px; color: #657786">@benedictevans bot</div> </div> I was made with huggingtweets. Create your own bot based on your favorite user with the demo! ## How does it work? The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. ## Training data The model was trained on @benedictevans's tweets. <table style='border-width:0'> <thead style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #CBD5E0'> <th style='border-width:0'>Data</th> <th style='border-width:0'>Quantity</th> </tr> </thead> <tbody style='border-width:0'> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Tweets downloaded</td> <td style='border-width:0'>3236</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Retweets</td> <td style='border-width:0'>662</td> </tr> <tr style='border-width:0 0 1px 0; border-color: #E2E8F0'> <td style='border-width:0'>Short tweets</td> <td style='border-width:0'>174</td> </tr> <tr style='border-width:0'> <td style='border-width:0'>Tweets kept</td> <td style='border-width:0'>2400</td> </tr> </tbody> </table> Explore the data, which is tracked with W&B artifacts at every step of the pipeline. ## Training procedure The model is based on a pre-trained GPT-2 which is fine-tuned on @benedictevans's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. ## Intended uses & limitations ### How to use You can use this model directly with a pipeline for text generation: <pre><code><span style="color:#03A9F4">from</span> transformers <span style="color:#03A9F4">import</span> pipeline generator = pipeline(<span style="color:#FF9800">'text-generation'</span>, model=<span style="color:#FF9800">'huggingtweets/benedictevans'</span>) generator(<span style="color:#FF9800">"My dream is"</span>, num_return_sequences=<span style="color:#8BC34A">5</span>)</code></pre> ### Limitations and bias The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* </section> ![Follow](URL <section class='prose'> For more details, visit the project repository. </section> ![GitHub stars](URL
[ "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @benedictevans's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3236</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>662</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>174</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2400</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @benedictevans's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/benedictevans'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report.", "## Training data\n\nThe model was trained on @benedictevans's tweets.\n\n<table style='border-width:0'>\n<thead style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #CBD5E0'>\n<th style='border-width:0'>Data</th>\n<th style='border-width:0'>Quantity</th>\n</tr>\n</thead>\n<tbody style='border-width:0'>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Tweets downloaded</td>\n<td style='border-width:0'>3236</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Retweets</td>\n<td style='border-width:0'>662</td>\n</tr>\n<tr style='border-width:0 0 1px 0; border-color: #E2E8F0'>\n<td style='border-width:0'>Short tweets</td>\n<td style='border-width:0'>174</td>\n</tr>\n<tr style='border-width:0'>\n<td style='border-width:0'>Tweets kept</td>\n<td style='border-width:0'>2400</td>\n</tr>\n</tbody>\n</table>\n\nExplore the data, which is tracked with W&B artifacts at every step of the pipeline.", "## Training procedure\n\nThe model is based on a pre-trained GPT-2 which is fine-tuned on @benedictevans's tweets.\n\nHyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility.\n\nAt the end of training, the final model is logged and versioned.", "## Intended uses & limitations", "### How to use\n\nYou can use this model directly with a pipeline for text generation:\n\n<pre><code><span style=\"color:#03A9F4\">from</span> transformers <span style=\"color:#03A9F4\">import</span> pipeline\ngenerator = pipeline(<span style=\"color:#FF9800\">'text-generation'</span>,\n model=<span style=\"color:#FF9800\">'huggingtweets/benedictevans'</span>)\ngenerator(<span style=\"color:#FF9800\">\"My dream is\"</span>, num_return_sequences=<span style=\"color:#8BC34A\">5</span>)</code></pre>", "### Limitations and bias\n\nThe model suffers from the same limitations and bias as GPT-2.\n\nIn addition, the data present in the user's tweets further affects the text generated by the model.", "## About\n\n*Built by Boris Dayma*\n\n</section>\n\n![Follow](URL\n\n<section class='prose'>\nFor more details, visit the project repository.\n</section>\n\n![GitHub stars](URL" ]
[ 57, 34, 430, 76, 9, 168, 48, 58 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## How does it work?\n\nThe model uses the following pipeline.\n\n!pipeline\n\nTo understand how the model was developed, check the W&B report." ]
[ -0.04084593802690506, 0.035596683621406555, -0.0024457171093672514, 0.04662978649139404, 0.10991521924734116, 0.022836215794086456, 0.12812861800193787, 0.0424627922475338, -0.03746044635772705, -0.03597303107380867, 0.22758877277374268, 0.1009177565574646, 0.03089720755815506, 0.17962171137332916, 0.010350672528147697, -0.2703946828842163, 0.015237200073897839, 0.0647135004401207, -0.07720091193914413, 0.15752871334552765, 0.05562684312462807, -0.049801189452409744, 0.08214939385652542, 0.032038331031799316, -0.165513277053833, -0.004831716418266296, -0.02072383277118206, -0.04504403471946716, 0.09232694655656815, 0.06956911832094193, 0.07011176645755768, 0.034282486885786057, 0.017851393669843674, -0.0714372918009758, 0.06354191154241562, 0.014452377334237099, -0.02349161170423031, 0.13615116477012634, 0.028668763116002083, -0.0002947957837022841, 0.1798527091741562, 0.11815319955348969, 0.016722125932574272, 0.016196802258491516, -0.1166125237941742, -0.0606788769364357, 0.012365915812551975, 0.04470464214682579, 0.10005760192871094, 0.058856748044490814, 0.01943058706820011, 0.13215520977973938, -0.1164829432964325, 0.08512931317090988, 0.1828341782093048, -0.24191047251224518, -0.006823094096034765, 0.061043880879879, 0.08518847078084946, 0.02852868288755417, -0.0027500083670020103, 0.050437480211257935, 0.06484624743461609, 0.021347152069211006, 0.03636380657553673, -0.0638979896903038, 0.05009674280881882, 0.011186900548636913, -0.10215270519256592, -0.07660897821187973, 0.2294245809316635, -0.0602995827794075, 0.004886541981250048, -0.028186868876218796, -0.08490151911973953, -0.06026294082403183, -0.012795685790479183, -0.05738116800785065, -0.01700775697827339, 0.034475523978471756, 0.0214629378169775, -0.10922146588563919, -0.07046429812908173, -0.1146778091788292, -0.09578189998865128, 0.17879855632781982, -0.02900231070816517, 0.08940442651510239, -0.2603394091129303, 0.2254239320755005, 0.07019765675067902, -0.11909467726945877, 0.04757627099752426, -0.11814092099666595, 0.08530338853597641, 0.02207805961370468, 0.041434839367866516, 0.07761925458908081, 0.04161522537469864, 0.11940909177064896, 0.03257475048303604, -0.00314142694696784, 0.05456724017858505, 0.07224375009536743, 0.0941813513636589, 0.1230122447013855, -0.07287998497486115, -0.07873915135860443, 0.08034131675958633, -0.04709449037909508, -0.11787727475166321, -0.06861002743244171, -0.1452382206916809, -0.004202034790068865, -0.0335184708237648, 0.07922724634408951, 0.05637102574110031, 0.09526676684617996, -0.01433930266648531, -0.05777551606297493, 0.04157217592000961, -0.06985066086053848, 0.019995175302028656, -0.011708668433129787, -0.060763102024793625, 0.13121068477630615, 0.04688004031777382, -0.014275136403739452, -0.08673419058322906, 0.07951672375202179, -0.1204570084810257, -0.07978847622871399, -0.08239417523145676, -0.05098734796047211, -0.007686138618737459, -0.11281027644872665, 0.049203939735889435, -0.11458899080753326, -0.22705501317977905, -0.01313747651875019, 0.04550096020102501, -0.016462473198771477, -0.03707117214798927, -0.040630146861076355, -0.009473399259150028, 0.04880702868103981, -0.042894408106803894, 0.039052072912454605, -0.05325083062052727, 0.050018906593322754, -0.09618895500898361, 0.051321301609277725, -0.10470817238092422, 0.041251372545957565, -0.09671807289123535, 0.07595963031053543, 0.0017866486450657248, 0.0454767569899559, 0.010063555091619492, 0.08985432982444763, -0.03214199095964432, -0.044836416840553284, -0.07869677990674973, 0.026561295613646507, 0.02284199558198452, 0.20134314894676208, -0.10196486860513687, -0.0819794163107872, 0.12614667415618896, -0.07252102345228195, -0.1288122832775116, 0.0409054271876812, -0.02831157296895981, 0.17958028614521027, 0.07370224595069885, 0.16096454858779907, 0.12039237469434738, -0.03798284754157066, 0.1263856589794159, 0.14821170270442963, -0.1269366294145584, -0.004153167363256216, 0.039416827261447906, 0.014042570255696774, -0.20559121668338776, 0.04052022844552994, -0.01946125738322735, 0.06584736704826355, -0.10201486945152283, -0.00946728140115738, 0.0031527038663625717, -0.02320384420454502, 0.0023026331327855587, -0.0472555048763752, 0.061078161001205444, 0.038737643510103226, -0.022763127461075783, 0.029682613909244537, 0.06635552644729614, -0.026160234585404396, -0.00823320634663105, -0.04484035447239876, 0.10009831935167313, -0.07260703295469284, 0.06797321885824203, -0.13160665333271027, -0.003086181590333581, -0.012450510635972023, 0.0972909927368164, 0.03811546042561531, 0.10621625185012817, 0.05470104143023491, 0.03126294165849686, 0.07701993733644485, -0.02846951223909855, 0.0746634304523468, 0.01002051867544651, -0.0850096121430397, -0.1272277683019638, 0.015217535197734833, -0.10649837553501129, -0.004191380459815264, -0.08523700386285782, -0.0015737697249278426, -0.12802091240882874, 0.05826076865196228, -0.016465239226818085, 0.0539434514939785, -0.055182818323373795, -0.04337453469634056, -0.044662054628133774, -0.022007398307323456, 0.0482511967420578, -0.033139705657958984, -0.06687536835670471, 0.16832175850868225, -0.15395553410053253, 0.2592274248600006, 0.1313537061214447, -0.0889907032251358, 0.002020547864958644, -0.07812081277370453, -0.04194976016879082, -0.012868959456682205, 0.07527336478233337, -0.03700125962495804, 0.15495356917381287, -0.03386903926730156, 0.17382068932056427, -0.099679134786129, -0.0334031768143177, 0.02144831046462059, -0.10981054604053497, 0.057656705379486084, 0.08051177859306335, 0.04427298903465271, -0.159133642911911, 0.08837400376796722, 0.19756358861923218, 0.05472150072455406, 0.20321963727474213, -0.006869805511087179, -0.06112205237150192, -0.05358194187283516, -0.0808696523308754, -0.052568696439266205, 0.056259434670209885, -0.09903126955032349, -0.004000484477728605, 0.06423190236091614, 0.08783505111932755, 0.11237549781799316, -0.10904275625944138, -0.046180129051208496, 0.05125856027007103, -0.004819708876311779, -0.051060013473033905, 0.07006146013736725, -0.0659489631652832, 0.13217470049858093, 0.014598124660551548, -0.07049204409122467, 0.0036125897895544767, -0.004913401324301958, -0.11891388893127441, 0.20653130114078522, -0.08047540485858917, -0.27306002378463745, -0.16792123019695282, -0.16288253664970398, 0.07165426760911942, 0.038431257009506226, 0.033738572150468826, -0.08776884526014328, -0.020982403308153152, 0.004409478977322578, 0.11553267389535904, -0.09698133170604706, 0.013121976517140865, 0.008159824647009373, -0.018650712445378304, -0.07579360157251358, -0.09033482521772385, -0.0241270512342453, -0.02461584471166134, 0.020020704716444016, 0.03998296707868576, -0.11154978722333908, 0.06758414953947067, 0.2167699933052063, -0.015538511797785759, 0.06870997697114944, 0.00025148785789497197, 0.26176807284355164, -0.08426473289728165, 0.040830448269844055, 0.11926601082086563, -0.08760137856006622, 0.05199241638183594, 0.07132956385612488, 0.03210015222430229, -0.014074578881263733, 0.016441889107227325, -0.11233895272016525, -0.12864868342876434, -0.1923626959323883, -0.06961654871702194, -0.028241310268640518, 0.13464264571666718, 0.031150488182902336, 0.04321796074509621, 0.10346641391515732, 0.07471037656068802, 0.06701335310935974, 0.03259968012571335, -0.0005120337591506541, 0.0647427961230278, 0.024594781920313835, -0.05812343955039978, 0.054217349737882614, -0.04845457896590233, -0.0797470211982727, 0.08279551565647125, -0.011098933406174183, 0.0927528515458107, 0.06928195804357529, 0.02340286411345005, 0.018686039373278618, 0.04218229651451111, 0.15593960881233215, 0.22442668676376343, -0.012412761338055134, -0.041085485368967056, -0.05078154057264328, -0.040494389832019806, -0.01600850187242031, 0.015044075436890125, -0.05785144492983818, -0.033252447843551636, -0.0728597640991211, -0.015066487714648247, 0.011195010505616665, 0.015441779047250748, 0.07578693330287933, -0.22024130821228027, -0.038240667432546616, 0.042616840451955795, -0.013794191181659698, -0.10639895498752594, 0.05872863903641701, 0.016779562458395958, -0.17391349375247955, -0.07854076474905014, -0.016605399549007416, 0.1603294163942337, -0.030760308727622032, 0.0619782954454422, 0.005449770484119654, 0.02271227352321148, -0.013140208087861538, 0.11191333085298538, -0.27346712350845337, 0.1954270750284195, 0.001131516881287098, -0.04876048117876053, -0.016439033672213554, -0.04243995249271393, 0.0009058643481694162, 0.14556926488876343, 0.09718295931816101, 0.0028763783629983664, 0.0669604167342186, -0.07678256928920746, -0.11943262070417404, 0.05284353718161583, 0.08068333566188812, -0.07738065719604492, 0.029960619285702705, -0.029798466712236404, 0.027152907103300095, -0.007555682212114334, -0.030231619253754616, 0.002119861776009202, -0.11661309748888016, 0.02936525270342827, -0.08075195550918579, 0.06012337654829025, 0.02433968149125576, -0.02529163844883442, -0.012048180215060711, 0.1316436529159546, -0.013300766237080097, -0.08264251798391342, -0.08976204693317413, -0.02328740619122982, 0.09523095935583115, -0.05599937587976456, 0.03358715400099754, -0.08175740391016006, -0.04073614999651909, 0.005860272329300642, -0.16970814764499664, 0.06983034312725067, -0.10846570879220963, -0.09971687942743301, -0.050264790654182434, 0.15346404910087585, 0.013677009381353855, 0.025709833949804306, 0.03220117464661598, -0.04211581498384476, -0.18150363862514496, -0.15989434719085693, -0.007562890648841858, 0.0717545747756958, -0.04433317109942436, 0.03638565540313721, 0.007171243894845247, 0.10013602674007416, 0.004198792390525341, 0.07230839878320694, 0.2026015669107437, 0.16423118114471436, -0.08760133385658264, 0.17723721265792847, 0.16266676783561707, -0.12243213504552841, -0.2722402811050415, -0.09522651135921478, -0.05925937369465828, 0.03468820080161095, 0.02297091670334339, -0.13072867691516876, 0.06184706464409828, -0.011241482570767403, -0.004976592492312193, 0.13391432166099548, -0.2790721356868744, -0.07025358080863953, 0.13864430785179138, -0.012145180255174637, 0.2560276985168457, -0.042459286749362946, -0.08155408501625061, -0.060940731316804886, -0.2339130938053131, 0.1595010906457901, -0.12908293306827545, 0.030256805941462517, -0.06380902975797653, 0.1317017376422882, 0.04475972056388855, -0.051817599684000015, 0.13714583218097687, -0.0770399421453476, 0.03692200407385826, -0.1231972947716713, -0.01437266543507576, 0.05212629213929176, -0.014681367203593254, 0.10554680228233337, -0.053141020238399506, 0.10400939732789993, -0.12106935679912567, -0.052672889083623886, -0.054288461804389954, 0.017598338425159454, -0.023758167400956154, -0.05668776109814644, -0.039483629167079926, -0.05230721831321716, 0.00942184031009674, -0.024894973263144493, -0.008981208316981792, -0.02189256064593792, 0.08200293034315109, 0.10853444039821625, 0.1416669338941574, -0.04508063197135925, -0.02666328102350235, -0.029412275180220604, -0.043095141649246216, 0.07755832374095917, -0.1675589680671692, -0.020979177206754684, 0.15767353773117065, 0.008264025673270226, 0.08081416040658951, 0.07994852215051651, -0.043529048562049866, -0.04116993397474289, 0.09435915946960449, -0.23738352954387665, -0.032961416989564896, -0.07289689034223557, -0.032304681837558746, 0.05143286660313606, 0.06389017403125763, 0.11233682930469513, -0.055076416581869125, -0.015500548295676708, 0.038369257003068924, -0.013473432511091232, -0.10457789897918701, 0.12659704685211182, 0.07594829052686691, 0.04931824654340744, -0.13000807166099548, 0.03979043290019035, -0.02080575004220009, -0.024042857810854912, -0.009190280921757221, 0.09610513597726822, -0.13868926465511322, -0.061987441033124924, 0.01100219041109085, 0.1624082624912262, -0.08940329402685165, -0.054934311658144, -0.00678250240162015, -0.07782098650932312, 0.06215988099575043, 0.06269455701112747, 0.039047662168741226, 0.10006190836429596, -0.08492296934127808, -0.004345493856817484, -0.04427671059966087, 0.02742549031972885, 0.04004936292767525, -0.01839151792228222, -0.11644710600376129, 0.050648268312215805, 0.01261399406939745, 0.21786263585090637, -0.12195943295955658, -0.07748695462942123, -0.13975368440151215, 0.03579137846827507, -0.1441981941461563, -0.02782432734966278, -0.09455464035272598, -0.0542730838060379, -0.024786408990621567, -0.02354593575000763, -0.05044161155819893, -0.03595460206270218, -0.06568260490894318, 0.04963921010494232, -0.01889806240797043, -0.04201965406537056, -0.018809955567121506, 0.04780932888388634, 0.10624072700738907, -0.0022816911805421114, 0.11582330614328384, 0.10476028919219971, -0.06149300932884216, 0.06964143365621567, -0.08975338935852051, 0.049342647194862366, 0.010800108313560486, -0.03639211133122444, 0.07890737056732178, 0.033158838748931885, 0.011678727343678474, -0.02014644630253315, -0.05248590186238289, 0.015699470415711403, 0.019494805485010147, -0.09001129865646362, 0.04338252544403076, 0.03427375108003616, -0.07128193974494934, -0.06945458799600601, -0.02831537090241909, -0.04915383830666542, 0.10966338962316513, 0.09227382391691208, 0.01580313965678215, 0.11524862796068192, -0.09982031583786011, -0.0043287696316838264, 0.0288130734115839, -0.08074736595153809, -0.01706261746585369, -0.10044533759355545, -0.01304725930094719, -0.02274717018008232, 0.2554529011249542, 0.12089171260595322, -0.025309232994914055, -0.03230812028050423, 0.07114472985267639, 0.08105676621198654, -0.0211230106651783, 0.14824873208999634, 0.03444083034992218, -0.0007331980159506202, -0.1400776505470276, 0.10673409700393677, -0.060156434774398804, -0.010151425376534462, 0.09550673514604568, -0.08319920301437378, 0.048856139183044434, 0.07468824833631516, -0.01950058713555336, 0.05372466519474983, -0.11716536432504654, -0.2690386474132538, 0.023945249617099762, 0.027653370052576065, -0.0441947840154171, 0.07253700494766235, 0.145015150308609, 0.00042942073196172714, 0.05244648456573486, 0.061493102461099625, -0.05709811672568321, -0.17804701626300812, -0.19115881621837616, -0.0384756401181221, -0.11082857847213745, -0.023826930671930313, -0.10639674216508865, 0.04148538038134575, -0.02072913944721222, 0.05925795063376427, -0.09639845043420792, 0.12447383254766464, 0.06843417137861252, -0.11577396094799042, 0.05810433253645897, -0.008805959485471249, 0.048459235578775406, -0.07387776672840118, 0.08210063725709915, -0.10721065104007721, -0.026499031111598015, -0.016933415085077286, 0.03711435943841934, -0.05858420953154564, 0.0011270071845501661, -0.10357651859521866, -0.06808403134346008, -0.056935109198093414, 0.09072309732437134, -0.024477418512105942, 0.03998230770230293, -0.014557241462171078, -0.061277762055397034, -0.025446701794862747, 0.2273169606924057, -0.018587565049529076, -0.043939489871263504, -0.0661960318684578, 0.2851298749446869, -0.06544138491153717, 0.07253559678792953, -0.032977886497974396, -0.001274158013984561, -0.07127615064382553, 0.2931469976902008, 0.36314713954925537, -0.14264726638793945, 0.011796033009886742, -0.018389053642749786, 0.03556118905544281, 0.07535336911678314, 0.18024654686450958, 0.07291083037853241, 0.3107033371925354, -0.04080776497721672, -0.01225926261395216, -0.10546047985553741, -0.03835856914520264, 0.014304363168776035, 0.02947218343615532, 0.08378855139017105, -0.05586446449160576, -0.06808875501155853, 0.1039084792137146, -0.26703301072120667, -0.02056516334414482, -0.16380304098129272, -0.061613935977220535, -0.04166705906391144, 0.0007227687747217715, 0.07237391918897629, 0.028740311041474342, 0.05115301162004471, -0.039005450904369354, -0.047156207263469696, 0.057444483041763306, -0.02154913917183876, -0.12674635648727417, 0.0002557095722295344, 0.143532857298851, -0.07906237244606018, -0.0018181405030190945, 0.0032308290246874094, 0.060348983854055405, 0.044118594378232956, 0.04119637981057167, -0.10164451599121094, 0.02608482725918293, 0.01246592216193676, -0.03363148868083954, -0.028164468705654144, 0.008156497962772846, 0.07835527509450912, -0.21697945892810822, 0.0020338469184935093, -0.14078554511070251, 0.011757226660847664, -0.07641053944826126, -0.006896127946674824, -0.08222074061632156, 0.03242125362157822, 0.004625517874956131, 0.1118803396821022, 0.11125602573156357, -0.03202005848288536, -0.0006144302315078676, -0.06265610456466675, 0.06727221608161926, -0.06884542852640152, -0.02960195019841194, -0.025150567293167114, -0.09257599711418152, -0.09335606545209885, 0.09815482050180435, -0.022339481860399246, -0.1427105814218521, 0.007601875811815262, -0.09401176869869232, -0.04369132220745087, -0.021486658602952957, 0.09382037818431854, 0.11086808145046234, 0.09180203825235367, -0.007599277421832085, 0.047748953104019165, 0.03120456263422966, 0.07436691224575043, -0.12886843085289001, -0.10148585587739944 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1421907606105329672/ypXqcYtY_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Marc Benioff</div> <div style="text-align: center; font-size: 14px;">@benioff</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Marc Benioff. | Data | Marc Benioff | | --- | --- | | Tweets downloaded | 3242 | | Retweets | 2645 | | Short tweets | 67 | | Tweets kept | 530 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1jmodnqz/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @benioff's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/alpvpdqh) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/alpvpdqh/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/benioff') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benioff
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Marc Benioff @benioff I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Marc Benioff. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @benioff's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1241961365725540353/s1_Jg8dJ_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Steven Suptic 🤖 AI Bot </div> <div style="font-size: 15px">@stevensuptic bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@stevensuptic's tweets](https://twitter.com/stevensuptic). | Data | Quantity | | --- | --- | | Tweets downloaded | 3190 | | Retweets | 211 | | Short tweets | 507 | | Tweets kept | 2472 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3cj525fz/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @stevensuptic's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/55ewqb50) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/55ewqb50/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/stevensuptic') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benjinaesen
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Steven Suptic AI Bot @stevensuptic bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @stevensuptic's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @stevensuptic's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1108382380723597313/hDIUPnFe_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Ben Congdon 🤖 AI Bot </div> <div style="font-size: 15px">@benrcongdon bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@benrcongdon's tweets](https://twitter.com/benrcongdon). | Data | Quantity | | --- | --- | | Tweets downloaded | 3219 | | Retweets | 394 | | Short tweets | 515 | | Tweets kept | 2310 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3aazmqd6/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @benrcongdon's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/7zvkav4e) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/7zvkav4e/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/benrcongdon') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/benrcongdon/1616637140236/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benrcongdon
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Ben Congdon AI Bot @benrcongdon bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @benrcongdon's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @benrcongdon's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/573946913327878144/d67c1agn_400x400.jpeg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">bens-i kareem 🤖 AI Bot </div> <div style="font-size: 15px">@benskerim bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@benskerim's tweets](https://twitter.com/benskerim). | Data | Quantity | | --- | --- | | Tweets downloaded | 546 | | Retweets | 12 | | Short tweets | 152 | | Tweets kept | 382 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1vrptz87/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @benskerim's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2tluib3u) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2tluib3u/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/benskerim') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/benskerim/1616795883644/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/benskerim
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
bens-i kareem AI Bot @benskerim bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @benskerim's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @benskerim's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1340882135893364741/B3Xl8AlK_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Bentley Rubinstein 🤖 AI Bot </div> <div style="font-size: 15px">@bentley bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bentley's tweets](https://twitter.com/bentley). | Data | Quantity | | --- | --- | | Tweets downloaded | 119 | | Retweets | 43 | | Short tweets | 1 | | Tweets kept | 75 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/exsqu0gt/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bentley's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/212vaa17) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/212vaa17/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bentley') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bentley/1616467138651/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bentley
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Bentley Rubinstein AI Bot @bentley bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @bentley's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bentley's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1097820307388334080/9ddg5F6v_400x400.png&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/847818629840228354/VXyQHfn0_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1278259160644227073/MfCyF7CG_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Bernie Sanders & wint & CNN</div> <div style="text-align: center; font-size: 14px;">@berniesanders-cnn-dril</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Bernie Sanders & wint & CNN. | Data | Bernie Sanders | wint | CNN | | --- | --- | --- | --- | | Tweets downloaded | 3250 | 3229 | 3250 | | Retweets | 429 | 473 | 30 | | Short tweets | 10 | 300 | 6 | | Tweets kept | 2811 | 2456 | 3214 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1yapgpjj/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @berniesanders-cnn-dril's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1hmm651a) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1hmm651a/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/berniesanders-cnn-dril') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/berniesanders-cnn-dril/1646127802129/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/berniesanders-cnn-dril
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Bernie Sanders & wint & CNN @berniesanders-cnn-dril I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Bernie Sanders & wint & CNN. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @berniesanders-cnn-dril's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/794725967948181506/Zn4x_F6i_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/794619281271033856/Fs0QQaH7_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1097820307388334080/9ddg5F6v_400x400.png&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Coffee Burger & Bernie Sanders & Bernie Sanders</div> <div style="text-align: center; font-size: 14px;">@berniesanders-coffee__burger-sensanders</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Coffee Burger & Bernie Sanders & Bernie Sanders. | Data | Coffee Burger | Bernie Sanders | Bernie Sanders | | --- | --- | --- | --- | | Tweets downloaded | 2471 | 3249 | 3250 | | Retweets | 525 | 296 | 429 | | Short tweets | 337 | 5 | 10 | | Tweets kept | 1609 | 2948 | 2811 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/2k4t7tx8/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @berniesanders-coffee__burger-sensanders's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/31ey7s5h) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/31ey7s5h/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/berniesanders-coffee__burger-sensanders') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/berniesanders-coffee__burger-sensanders
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Coffee Burger & Bernie Sanders & Bernie Sanders @berniesanders-coffee\_\_burger-sensanders I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Coffee Burger & Bernie Sanders & Bernie Sanders. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @berniesanders-coffee\_\_burger-sensanders's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/794725967948181506/Zn4x_F6i_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1097820307388334080/9ddg5F6v_400x400.png&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Coffee Burger & Bernie Sanders</div> <div style="text-align: center; font-size: 14px;">@berniesanders-coffee__burger</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Coffee Burger & Bernie Sanders. | Data | Coffee Burger | Bernie Sanders | | --- | --- | --- | | Tweets downloaded | 2471 | 3250 | | Retweets | 525 | 429 | | Short tweets | 337 | 10 | | Tweets kept | 1609 | 2811 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/ltwd1tj1/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @berniesanders-coffee__burger's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/121buw7a) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/121buw7a/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/berniesanders-coffee__burger') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/berniesanders-coffee__burger
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Coffee Burger & Bernie Sanders @berniesanders-coffee\_\_burger I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Coffee Burger & Bernie Sanders. Data: Tweets downloaded, Coffee Burger: 2471, Bernie Sanders: 3250 Data: Retweets, Coffee Burger: 525, Bernie Sanders: 429 Data: Short tweets, Coffee Burger: 337, Bernie Sanders: 10 Data: Tweets kept, Coffee Burger: 1609, Bernie Sanders: 2811 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @berniesanders-coffee\_\_burger's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/847818629840228354/VXyQHfn0_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1097820307388334080/9ddg5F6v_400x400.png&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">wint & Bernie Sanders</div> <div style="text-align: center; font-size: 14px;">@berniesanders-dril</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from wint & Bernie Sanders. | Data | wint | Bernie Sanders | | --- | --- | --- | | Tweets downloaded | 3229 | 3250 | | Retweets | 473 | 429 | | Short tweets | 300 | 10 | | Tweets kept | 2456 | 2811 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/yw6378l1/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @berniesanders-dril's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3pydufi9) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3pydufi9/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/berniesanders-dril') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/berniesanders-dril
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG wint & Bernie Sanders @berniesanders-dril I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from wint & Bernie Sanders. Data: Tweets downloaded, wint: 3229, Bernie Sanders: 3250 Data: Retweets, wint: 473, Bernie Sanders: 429 Data: Short tweets, wint: 300, Bernie Sanders: 10 Data: Tweets kept, wint: 2456, Bernie Sanders: 2811 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @berniesanders-dril's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1097820307388334080/9ddg5F6v_400x400.png&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Bernie Sanders</div> <div style="text-align: center; font-size: 14px;">@berniesanders</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Bernie Sanders. | Data | Bernie Sanders | | --- | --- | | Tweets downloaded | 3250 | | Retweets | 387 | | Short tweets | 7 | | Tweets kept | 2856 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1peyp14p/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @berniesanders's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1w2aztyw) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1w2aztyw/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/berniesanders') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "http://www.huggingtweets.com/berniesanders/1661604495356/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/berniesanders
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Bernie Sanders @berniesanders I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Bernie Sanders. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @berniesanders's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/2113290180/images-1_400x400.jpeg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1284655541227323395/4E-Y6plH_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Best Music Lyric & Wisdom_by_GPT3</div> <div style="text-align: center; font-size: 14px;">@bestmusiclyric-bygpt3</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Best Music Lyric & Wisdom_by_GPT3. | Data | Best Music Lyric | Wisdom_by_GPT3 | | --- | --- | --- | | Tweets downloaded | 3248 | 293 | | Retweets | 1092 | 3 | | Short tweets | 834 | 86 | | Tweets kept | 1322 | 204 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/101pevjn/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bestmusiclyric-bygpt3's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3qkafun2) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3qkafun2/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bestmusiclyric-bygpt3') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bestmusiclyric-bygpt3/1621260459372/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bestmusiclyric-bygpt3
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Best Music Lyric & Wisdom\_by\_GPT3 @bestmusiclyric-bygpt3 I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Best Music Lyric & Wisdom\_by\_GPT3. Data: Tweets downloaded, Best Music Lyric: 3248, Wisdom\_by\_GPT3: 293 Data: Retweets, Best Music Lyric: 1092, Wisdom\_by\_GPT3: 3 Data: Short tweets, Best Music Lyric: 834, Wisdom\_by\_GPT3: 86 Data: Tweets kept, Best Music Lyric: 1322, Wisdom\_by\_GPT3: 204 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bestmusiclyric-bygpt3's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/2113290180/images-1_400x400.jpeg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/903769803768217600/EKtan_aM_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Best Music Lyric & mark normand</div> <div style="text-align: center; font-size: 14px;">@bestmusiclyric-marknorm</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Best Music Lyric & mark normand. | Data | Best Music Lyric | mark normand | | --- | --- | --- | | Tweets downloaded | 3247 | 3247 | | Retweets | 1112 | 138 | | Short tweets | 820 | 521 | | Tweets kept | 1315 | 2588 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/41htfpjc/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bestmusiclyric-marknorm's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2715uuiw) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2715uuiw/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bestmusiclyric-marknorm') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bestmusiclyric-marknorm/1621435674803/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bestmusiclyric-marknorm
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Best Music Lyric & mark normand @bestmusiclyric-marknorm I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Best Music Lyric & mark normand. Data: Tweets downloaded, Best Music Lyric: 3247, mark normand: 3247 Data: Retweets, Best Music Lyric: 1112, mark normand: 138 Data: Short tweets, Best Music Lyric: 820, mark normand: 521 Data: Tweets kept, Best Music Lyric: 1315, mark normand: 2588 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bestmusiclyric-marknorm's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/2113290180/images-1_400x400.jpeg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/818597091215716353/5ejD1Ojs_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Best Music Lyric & Poets.org</div> <div style="text-align: center; font-size: 14px;">@bestmusiclyric-poetsorg</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Best Music Lyric & Poets.org. | Data | Best Music Lyric | Poets.org | | --- | --- | --- | | Tweets downloaded | 3246 | 3249 | | Retweets | 1150 | 143 | | Short tweets | 807 | 13 | | Tweets kept | 1289 | 3093 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/11s270ah/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bestmusiclyric-poetsorg's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/p26o0mhb) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/p26o0mhb/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bestmusiclyric-poetsorg') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bestmusiclyric-poetsorg/1622303204122/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bestmusiclyric-poetsorg
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Best Music Lyric & URL @bestmusiclyric-poetsorg I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Best Music Lyric & URL. Data: Tweets downloaded, Best Music Lyric: 3246, URL: 3249 Data: Retweets, Best Music Lyric: 1150, URL: 143 Data: Short tweets, Best Music Lyric: 807, URL: 13 Data: Tweets kept, Best Music Lyric: 1289, URL: 3093 Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bestmusiclyric-poetsorg's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/2113290180/images-1_400x400.jpeg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Best Music Lyric 🤖 AI Bot </div> <div style="font-size: 15px">@bestmusiclyric bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bestmusiclyric's tweets](https://twitter.com/bestmusiclyric). | Data | Quantity | | --- | --- | | Tweets downloaded | 3244 | | Retweets | 1060 | | Short tweets | 853 | | Tweets kept | 1331 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/1ilv29ew/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bestmusiclyric's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/1wqx12s6) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/1wqx12s6/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bestmusiclyric') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bestmusiclyric/1620313468667/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bestmusiclyric
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "has_space", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
Best Music Lyric AI Bot @bestmusiclyric bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @bestmusiclyric's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bestmusiclyric's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 61 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n" ]
[ 0.023045063018798828, -0.02204611524939537, -0.0056217629462480545, 0.024998219683766365, 0.1312267929315567, 0.02466566488146782, 0.08025442808866501, 0.1621488779783249, -0.030768360942602158, 0.015156957320868969, 0.17928390204906464, 0.12833106517791748, -0.025381935760378838, 0.10042161494493484, -0.03942488506436348, -0.2719423770904541, 0.046165984123945236, 0.04379456117749214, -0.031213730573654175, 0.13666078448295593, 0.08867661654949188, -0.045726194977760315, 0.10844767093658447, -0.024391112849116325, -0.19374926388263702, 0.03891391307115555, 0.04400325194001198, -0.09513276815414429, 0.11447238177061081, 0.03918393328785896, 0.09885572642087936, 0.025941194966435432, -0.0787009596824646, -0.10229984670877457, 0.04564651846885681, 0.05498135834932327, -0.0677126869559288, 0.07756631076335907, 0.062077198177576065, -0.08898276835680008, 0.14696337282657623, 0.027552109211683273, -0.013472146354615688, 0.05225503072142601, -0.17611517012119293, -0.06823833286762238, -0.03582452982664108, 0.013555791229009628, 0.03382846713066101, 0.07916142791509628, -0.03099752962589264, 0.17965897917747498, -0.09157974272966385, 0.08549632132053375, 0.21223965287208557, -0.31330862641334534, -0.019718647003173828, 0.09505957365036011, 0.13277654349803925, 0.060757726430892944, -0.035037562251091, 0.08270368725061417, 0.060885775834321976, 0.014816065318882465, 0.053595565259456635, -0.05630973353981972, -0.11095286160707474, 0.05735233798623085, -0.08560370653867722, -0.053243570029735565, 0.2270948737859726, -0.03627891466021538, 0.06678362190723419, -0.06014703959226608, -0.0968315601348877, -0.05191762372851372, -0.014206932857632637, -0.0007905814563855529, -0.039079900830984116, 0.07661662995815277, -0.01842164434492588, -0.06635472923517227, -0.15472756326198578, 0.003009884152561426, -0.16963911056518555, 0.12417193502187729, -0.009206476621329784, 0.03786933794617653, -0.19481845200061798, 0.09602963179349899, -0.007850876078009605, -0.10225676745176315, 0.06259744614362717, -0.08888275921344757, 0.06638791412115097, 0.012942429631948471, -0.07640603184700012, -0.02106644958257675, 0.08056198060512543, 0.13388165831565857, -0.032740268856287, -0.00926523469388485, 0.013319301418960094, 0.0830560028553009, 0.06792842596769333, 0.04181957617402077, -0.041336700320243835, -0.05350780487060547, 0.029261786490678787, -0.10449601709842682, 0.0012442318256944418, -0.08886091411113739, -0.11949656903743744, -0.04972672089934349, 0.041117310523986816, 0.037330131977796555, 0.05952366814017296, 0.1147349551320076, -0.03513633459806442, 0.0018127808580175042, 0.06821189075708389, -0.04676239565014839, 0.02456514909863472, -0.021932292729616165, 0.026449091732501984, 0.09052641689777374, -0.009602625854313374, 0.029862230643630028, -0.07528015971183777, 0.04637369140982628, -0.11322551965713501, -0.012416922487318516, -0.023197615519165993, -0.08945703506469727, 0.03664110228419304, -0.11818758398294449, 0.006976603530347347, -0.17672543227672577, -0.0827573835849762, -0.0010230062762275338, -0.016674630343914032, -0.024935759603977203, -0.08746512979269028, -0.005465231370180845, -0.047830674797296524, 0.10002142190933228, -0.04745052009820938, 0.015952352434396744, -0.05889752879738808, 0.09694558382034302, -0.06273946166038513, 0.10444994270801544, -0.1365537941455841, 0.05791698396205902, -0.13272713124752045, -0.0095265107229352, -0.05663247033953667, 0.04933531954884529, -0.002125486033037305, 0.15809306502342224, -0.0019368804059922695, -0.0151921221986413, -0.11728960275650024, 0.0808638334274292, -0.014826769009232521, 0.2024649977684021, -0.10628247261047363, -0.12109203636646271, 0.19116753339767456, -0.05156968906521797, -0.13711145520210266, 0.1298442780971527, 0.011141160503029823, 0.06715473532676697, 0.05843217298388481, 0.23018912971019745, 0.015947241336107254, -0.02980491891503334, 0.026306381449103355, 0.09368371963500977, -0.1374131441116333, -0.006914369761943817, 0.017705213278532028, 0.006864378694444895, -0.08455020934343338, 0.03641775995492935, 0.11606235057115555, 0.08481859415769577, -0.061023809015750885, -0.02748442441225052, -0.0490991435945034, 0.004382774233818054, 0.06565552204847336, -0.0016684056026861072, 0.10515616089105606, -0.11406049132347107, -0.0735328420996666, -0.041395965963602066, 0.0035774358548223972, 0.023191820830106735, 0.04733319953083992, -0.026123911142349243, 0.12253347784280777, -0.010156016796827316, 0.04058242216706276, -0.14151473343372345, -0.08605361729860306, -0.03254232183098793, 0.15478350222110748, 0.03953593224287033, 0.11633965373039246, 0.05719606578350067, -0.06777982413768768, -0.01036068145185709, 0.002454544650390744, 0.14101839065551758, -0.015767211094498634, -0.08905056118965149, -0.07599052041769028, 0.07826608419418335, -0.07611092925071716, 0.020115256309509277, -0.057017967104911804, 0.037102315574884415, 0.06556723266839981, 0.11540397256612778, -0.0017086438601836562, 0.03440048173069954, -0.0123360063880682, -0.000012290506674617063, -0.08933671563863754, -0.013215054757893085, 0.08847546577453613, -0.008405817672610283, -0.05117173492908478, 0.2598556578159332, -0.19900687038898468, 0.24127228558063507, 0.231798455119133, -0.2356785386800766, -0.03353632614016533, -0.01462375745177269, -0.04819153994321823, 0.01492219790816307, 0.037874069064855576, -0.05811957269906998, 0.03794299066066742, -0.05634453147649765, 0.1481509953737259, -0.04520590230822563, -0.06490635871887207, 0.013088049367070198, -0.06494753062725067, -0.06585344672203064, 0.06484286487102509, 0.04194590449333191, -0.12645718455314636, 0.20141667127609253, 0.26711732149124146, 0.04010975360870361, 0.20662355422973633, 0.018375245854258537, -0.004053029697388411, 0.05102868750691414, -0.06542815268039703, -0.058470260351896286, -0.04150864854454994, -0.1588040441274643, -0.051093969494104385, 0.07530313730239868, 0.034314632415771484, 0.0898759514093399, -0.09657558798789978, -0.07673535495996475, 0.007029566913843155, 0.023532038554549217, -0.01306160632520914, 0.1372765153646469, 0.048913490027189255, 0.13490548729896545, -0.0032259346917271614, 0.011654623784124851, 0.06254645437002182, 0.02089114673435688, -0.08627534657716751, 0.13996493816375732, -0.14251753687858582, -0.36767199635505676, -0.11936365067958832, -0.09697052836418152, -0.011594709008932114, 0.029661882668733597, 0.12459489703178406, -0.12274473905563354, 0.006536944769322872, -0.04606909677386284, 0.09110275655984879, -0.09474314004182816, 0.046487726271152496, -0.09117124229669571, 0.0193075742572546, -0.053507283329963684, -0.08736079931259155, -0.04219098761677742, -0.021037515252828598, -0.08632376790046692, 0.16062548756599426, -0.0669471025466919, 0.07220233231782913, 0.17339450120925903, 0.005650070495903492, 0.029418056830763817, -0.04840487241744995, 0.1996528059244156, -0.10633594542741776, 0.03494453802704811, 0.15129250288009644, 0.019962413236498833, 0.08429192751646042, 0.10996142029762268, -0.010845727287232876, -0.06873574107885361, 0.04305357486009598, 0.009354572743177414, -0.11143358051776886, -0.1527336835861206, -0.10894794762134552, -0.10044675320386887, 0.13014699518680573, 0.05656032636761665, 0.06703919917345047, 0.16037318110466003, 0.061513472348451614, -0.019407641142606735, -0.009107081219553947, -0.019766777753829956, 0.07649408280849457, 0.1467757225036621, -0.026956025511026382, 0.14976274967193604, -0.05412312224507332, -0.11098915338516235, 0.12750805914402008, 0.02523878403007984, 0.02736341767013073, 0.030087651684880257, 0.004279325716197491, 0.00023339648032560945, 0.14612308144569397, 0.13334205746650696, 0.07964257895946503, -0.018630146980285645, -0.030937759205698967, -0.05317160487174988, -0.02153072878718376, -0.001635383814573288, 0.056616492569446564, 0.04805357754230499, -0.15345874428749084, -0.05428538843989372, -0.1453014612197876, 0.09931683540344238, 0.08077193051576614, 0.10425683856010437, -0.20023800432682037, -0.0020670441444963217, 0.08631721884012222, -0.031204866245388985, -0.10470310598611832, 0.07385390251874924, 0.0852615162730217, -0.10122890770435333, 0.050357360392808914, -0.00810049194842577, 0.11262702941894531, 0.03939642012119293, 0.10314954072237015, -0.056073401123285294, -0.05590416491031647, -0.017166707664728165, 0.08704748004674911, -0.29989248514175415, 0.16933052241802216, -0.026653995737433434, -0.09390314668416977, -0.06273353099822998, -0.027042074128985405, 0.02722407877445221, 0.08160185068845749, 0.08276555687189102, 0.03114314004778862, -0.03465403988957405, -0.07583022117614746, -0.021011626347899437, 0.0226152203977108, 0.11752773076295853, -0.05714089423418045, -0.009762411937117577, -0.04663566127419472, 0.01674286276102066, 0.010660070925951004, 0.09401563555002213, 0.010384885594248772, -0.16617059707641602, 0.06855206191539764, 0.042636506259441376, 0.033747557550668716, 0.029849642887711525, -0.031014347448945045, -0.13844801485538483, 0.17258141934871674, 0.03255597501993179, -0.050840288400650024, -0.12468960136175156, -0.052433181554079056, 0.053452592343091965, -0.04645095020532608, 0.022545570507645607, -0.06071555241942406, 0.008225022815167904, -0.07477552443742752, -0.20190678536891937, 0.1422843039035797, -0.07532759755849838, -0.06814900040626526, -0.0482926070690155, 0.1726374328136444, -0.08437072485685349, 0.021726032719016075, 0.00648227334022522, 0.05337712913751602, -0.15247385203838348, -0.10076600313186646, 0.07157283276319504, -0.04121299460530281, 0.0384424589574337, -0.002682593185454607, -0.029120588675141335, -0.0009653342422097921, -0.015056867152452469, 0.00028136532637290657, 0.27343273162841797, 0.23626449704170227, -0.08351079374551773, 0.17889991402626038, 0.06570972502231598, -0.0586962029337883, -0.32806596159935, -0.09495645016431808, -0.13385947048664093, -0.008901155553758144, 0.027000220492482185, -0.12693896889686584, 0.04112457484006882, 0.020501481369137764, -0.01586635410785675, 0.1080545112490654, -0.20121033489704132, -0.09839294105768204, 0.08066169917583466, -0.06320268660783768, 0.4066661596298218, -0.14013591408729553, -0.06666210293769836, -0.03479941561818123, -0.15767717361450195, 0.2101469784975052, -0.07493568956851959, 0.0867309719324112, -0.019003480672836304, 0.1150895357131958, 0.05205247923731804, -0.017986591905355453, 0.1076529324054718, -0.02941650152206421, 0.008291213773190975, -0.133688822388649, -0.0708819180727005, 0.07801494002342224, -0.013113176450133324, 0.0073072900995612144, -0.12169554829597473, 0.016744306311011314, -0.17788201570510864, 0.004354916512966156, -0.12219570577144623, 0.08279965072870255, 0.010220671072602272, -0.061546485871076584, -0.042094141244888306, -0.03481869027018547, -0.002012162236496806, -0.017395906150341034, 0.1828150898218155, -0.057601477950811386, 0.2122170478105545, 0.08566316217184067, 0.064366415143013, -0.13606293499469757, 0.05063160881400108, -0.0157748032361269, -0.07391754537820816, 0.08055251836776733, -0.19411487877368927, 0.050044044852256775, 0.07842409610748291, -0.03550063446164131, 0.049127377569675446, 0.09029321372509003, -0.004625566303730011, 0.007996995933353901, 0.17518295347690582, -0.2641419768333435, -0.00870419479906559, -0.06501568108797073, -0.06398701667785645, 0.08765042573213577, 0.03337143734097481, 0.16669827699661255, 0.002525120973587036, -0.050673842430114746, 0.014745346270501614, 0.021194254979491234, -0.05955834314227104, 0.031477998942136765, 0.02697918750345707, 0.012441786006093025, -0.11749376356601715, 0.04954813793301582, 0.030477186664938927, -0.15089958906173706, 0.032835379242897034, 0.20457783341407776, -0.10553225129842758, -0.1281069666147232, -0.049091774970293045, 0.04288773611187935, -0.08717567473649979, 0.02726605348289013, -0.009190280921757221, -0.09470494836568832, 0.0573163777589798, 0.13278797268867493, 0.062310438603162766, 0.10899180918931961, -0.0015607187524437904, -0.016328703612089157, -0.032409630715847015, -0.018961092457175255, 0.023581283167004585, 0.03034031018614769, -0.08721444010734558, 0.09516870975494385, -0.02652714028954506, 0.14469143748283386, -0.10320838540792465, -0.045907825231552124, -0.15597106516361237, -0.03938088193535805, -0.0625808909535408, -0.10799501836299896, -0.07772707939147949, -0.07070928066968918, 0.010625030845403671, -0.057112570852041245, -0.05386120453476906, -0.07970327138900757, -0.11000929772853851, -0.0029484170954674482, -0.03910250589251518, 0.04829269275069237, -0.07395005971193314, -0.002291048178449273, 0.1147882267832756, -0.020846085622906685, 0.15595434606075287, 0.10914915055036545, -0.0885118842124939, 0.0818740725517273, -0.11731013655662537, -0.09784740954637527, 0.08430901914834976, -0.012961963191628456, 0.03967556729912758, 0.10447244346141815, 0.0016526463441550732, 0.024455370381474495, 0.05921920761466026, 0.056686755269765854, 0.032198164612054825, -0.10841073840856552, 0.08236350864171982, -0.009807482361793518, -0.16146379709243774, -0.04428122565150261, -0.0764978900551796, 0.03825283423066139, 0.019274720922112465, 0.11204726994037628, -0.03897618129849434, 0.06797946989536285, -0.09499135613441467, 0.02731928415596485, -0.0030131423845887184, -0.18112139403820038, -0.02819901704788208, -0.05015208199620247, 0.03486237674951553, 0.0061057209968566895, 0.21226072311401367, 0.07469984889030457, -0.07797075062990189, 0.052126914262771606, 0.09802792966365814, 0.010076619684696198, 0.00902772881090641, 0.14953187108039856, 0.08701325207948685, -0.08160319179296494, -0.12231837213039398, 0.0692681148648262, 0.025796039029955864, -0.01994943618774414, 0.09121832251548767, 0.02836466394364834, 0.04820539429783821, 0.07800399512052536, -0.008208714425563812, 0.002804644638672471, -0.07332517206668854, -0.0980529934167862, -0.03354928642511368, 0.06204527989029884, -0.03742247819900513, 0.0858326330780983, 0.18433237075805664, -0.011270957067608833, 0.04023509472608566, -0.06021422520279884, -0.01479005441069603, -0.131768599152565, -0.1387748271226883, -0.06989447772502899, -0.16190116107463837, -0.01699049025774002, -0.08803019672632217, 0.06468790769577026, 0.06403729319572449, 0.05623135343194008, -0.0440327525138855, 0.0547567754983902, 0.0526677668094635, -0.09924907237291336, 0.07835037261247635, -0.02403833717107773, 0.045166827738285065, -0.019312484189867973, -0.0029918591026216745, -0.09736528247594833, 0.03896813467144966, -0.02290741354227066, 0.042304784059524536, -0.04907459020614624, 0.016129078343510628, -0.16801723837852478, -0.12587063014507294, -0.0551501139998436, 0.05602076277136803, -0.051323167979717255, 0.05448102205991745, 0.017261920496821404, 0.009536868892610073, 0.019009340554475784, 0.26213759183883667, -0.06562719494104385, -0.02456813119351864, -0.04132092744112015, 0.14051729440689087, -0.009935270063579082, 0.07769134640693665, -0.0549197718501091, -0.019326863810420036, -0.11462818831205368, 0.30631375312805176, 0.34365975856781006, -0.09576758742332458, 0.04471958056092262, -0.016151167452335358, 0.03212709724903107, 0.10395953804254532, 0.11348164826631546, 0.10550481081008911, 0.22288531064987183, -0.07528717070817947, -0.023958787322044373, -0.02763318084180355, -0.008543568663299084, -0.07610689103603363, 0.08119126409292221, 0.027744179591536522, -0.06762602180242538, -0.031160930171608925, 0.05028000473976135, -0.1926647573709488, 0.08593353629112244, -0.08396037667989731, -0.2089729756116867, -0.045300185680389404, 0.03321804851293564, 0.12435340881347656, -0.0017329715192317963, 0.12409063428640366, -0.001308076549321413, -0.06459707021713257, 0.004358434583991766, 0.01711258478462696, -0.2058516889810562, 0.010330619290471077, 0.07579771429300308, -0.1278086006641388, -0.008316628634929657, -0.036133430898189545, -0.007023192476481199, 0.08239703625440598, 0.030314216390252113, -0.04562307521700859, 0.013821984641253948, 0.002609855029731989, -0.023103008046746254, -0.03019971214234829, 0.05462737753987312, 0.03617123141884804, -0.13509878516197205, 0.08838478475809097, -0.1482846587896347, 0.0359911173582077, -0.06373808532953262, -0.0009798946557566524, -0.006681718863546848, 0.009986492805182934, -0.038703128695487976, 0.08425160497426987, 0.0816703662276268, -0.022384973242878914, -0.012395314872264862, -0.058895740658044815, -0.054292429238557816, -0.022564271464943886, -0.07696029543876648, -0.10243724286556244, -0.11822693049907684, -0.09812650829553604, 0.09495946764945984, -0.02070685848593712, -0.18472422659397125, 0.009433140978217125, -0.09159061312675476, 0.05415434390306473, -0.14695541560649872, 0.1135123074054718, 0.10353317111730576, 0.0020196076948195696, 0.004518974106758833, -0.041260309517383575, 0.08009263873100281, 0.1161307767033577, -0.08494146913290024, -0.07230672985315323 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1442634650703237120/mXIcYtIs_400x400.jpg&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1441096557944737802/y56EUiiU_400x400.png&#39;)"> </div> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1431003324157812739/QYyroq6k_400x400.jpg&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI CYBORG 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Elon Musk & Beth Kindig & I/O Fund Official</div> <div style="text-align: center; font-size: 14px;">@beth_kindig-elonmusk-iofundofficial</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Elon Musk & Beth Kindig & I/O Fund Official. | Data | Elon Musk | Beth Kindig | I/O Fund Official | | --- | --- | --- | --- | | Tweets downloaded | 2400 | 3247 | 1935 | | Retweets | 127 | 484 | 143 | | Short tweets | 642 | 273 | 8 | | Tweets kept | 1631 | 2490 | 1784 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3pyiqrq2/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @beth_kindig-elonmusk-iofundofficial's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/3anxlpvl) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/3anxlpvl/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/beth_kindig-elonmusk-iofundofficial') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/beth_kindig-elonmusk-iofundofficial
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI CYBORG Elon Musk & Beth Kindig & I/O Fund Official @beth\_kindig-elonmusk-iofundofficial I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Elon Musk & Beth Kindig & I/O Fund Official. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @beth\_kindig-elonmusk-iofundofficial's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1005205545144336384/tRX1NN3Z_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">Brian likes Communism 🤖 AI Bot </div> <div style="font-size: 15px">@bfkelleher bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bfkelleher's tweets](https://twitter.com/bfkelleher). | Data | Quantity | | --- | --- | | Tweets downloaded | 3097 | | Retweets | 1424 | | Short tweets | 248 | | Tweets kept | 1425 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3gd8lb66/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bfkelleher's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/2xmh1eim) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/2xmh1eim/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bfkelleher') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bfkelleher/1614109725850/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bfkelleher
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Brian likes Communism AI Bot @bfkelleher bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @bfkelleher's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bfkelleher's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]
null
null
transformers
<div class="inline-flex flex-col" style="line-height: 1.5;"> <div class="flex"> <div style="display:inherit; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;https://pbs.twimg.com/profile_images/1399217176343547904/rFLGp2F7_400x400.jpg&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> <div style="display:none; margin-left: 4px; margin-right: 4px; width: 92px; height:92px; border-radius: 50%; background-size: cover; background-image: url(&#39;&#39;)"> </div> </div> <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div> <div style="text-align: center; font-size: 16px; font-weight: 800">Harsha Bhogle</div> <div style="text-align: center; font-size: 14px;">@bhogleharsha</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-Model-to-Generate-Tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on tweets from Harsha Bhogle. | Data | Harsha Bhogle | | --- | --- | | Tweets downloaded | 3250 | | Retweets | 1399 | | Short tweets | 92 | | Tweets kept | 1759 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/i02fyzkh/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bhogleharsha's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/39kwdhq0) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/39kwdhq0/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bhogleharsha') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://github.com/borisdayma/huggingtweets/blob/master/img/logo.png?raw=true", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bhogleharsha
[ "transformers", "pytorch", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
AI BOT Harsha Bhogle @bhogleharsha I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on tweets from Harsha Bhogle. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bhogleharsha's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 54 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0001764487533364445, -0.01891571842133999, -0.0068881697952747345, 0.01242890115827322, 0.16224369406700134, 0.04406825825572014, 0.08452208340167999, 0.14250440895557404, -0.026455026119947433, -0.016114573925733566, 0.17334569990634918, 0.17106501758098602, -0.014037161134183407, 0.08718273043632507, -0.05552244931459427, -0.2646014094352722, 0.044212065637111664, 0.058431971818208694, -0.020032864063978195, 0.14111687242984772, 0.0714879110455513, -0.01828647032380104, 0.10845158249139786, -0.02953636273741722, -0.18877948820590973, 0.03499612212181091, 0.05605728179216385, -0.09992336481809616, 0.11936124414205551, 0.04713086411356926, 0.08659981191158295, 0.015409729443490505, -0.07482189685106277, -0.12249794602394104, 0.03866785019636154, 0.041891295462846756, -0.0629689022898674, 0.05911329388618469, 0.08818957209587097, -0.11120674759149551, 0.1456013321876526, 0.07799911499023438, -0.01863223686814308, 0.07941857725381851, -0.17164963483810425, -0.019008882343769073, -0.036983806639909744, 0.005464354529976845, 0.057414017617702484, 0.07476010918617249, -0.01932354085147381, 0.1732589453458786, -0.0767555758357048, 0.09587065875530243, 0.16117197275161743, -0.2913956344127655, -0.005072304047644138, 0.0498935841023922, 0.06722559779882431, 0.03902119770646095, -0.015738610178232193, 0.08706029504537582, 0.06277379393577576, 0.02536560781300068, -0.0014978112885728478, -0.06339331716299057, -0.0928240567445755, 0.04014677554368973, -0.0745372325181961, -0.06578013300895691, 0.20811960101127625, -0.039430033415555954, 0.050536420196294785, -0.03807967156171799, -0.10491003096103668, -0.02197154052555561, -0.015753688290715218, 0.00720712635666132, -0.06031509116292, 0.08949250727891922, -0.014659170061349869, -0.07363511621952057, -0.15025688707828522, -0.016003666445612907, -0.18369624018669128, 0.1574522852897644, 0.003967532888054848, 0.04864242300391197, -0.2093716263771057, 0.11408322304487228, 0.020872395485639572, -0.08021171391010284, 0.047296058386564255, -0.09546594321727753, 0.07179957628250122, 0.002116252202540636, -0.05267130210995674, -0.02264278009533882, 0.08682441711425781, 0.15258505940437317, -0.026140356436371803, 0.0017382961232215166, -0.027155736461281776, 0.07059825956821442, 0.05281682312488556, 0.040018972009420395, -0.017763059586286545, -0.04289618879556656, 0.045782025903463364, -0.15945611894130707, -0.007582054473459721, -0.06918198615312576, -0.1068587526679039, -0.05112413689494133, 0.022331949323415756, 0.06442617624998093, 0.031878020614385605, 0.11345649510622025, -0.04609488695859909, -0.014168480411171913, 0.06350603699684143, -0.042460083961486816, -0.0172110628336668, -0.016047241166234016, 0.015319211408495903, 0.14095398783683777, -0.018181998282670975, 0.030711950734257698, -0.11251416057348251, 0.0645761713385582, -0.09946728497743607, -0.01914660818874836, -0.0071241967380046844, -0.04078202322125435, 0.029947424307465553, -0.13505136966705322, 0.010975461453199387, -0.1711256355047226, -0.14874492585659027, 0.009897243231534958, -0.02876151353120804, -0.018498392775654793, -0.059200938791036606, -0.03973992541432381, -0.01719699800014496, 0.06084148958325386, -0.04528016224503517, 0.0009662628290243447, -0.05951235070824623, 0.11524756252765656, -0.05245914310216904, 0.06888316571712494, -0.13250301778316498, 0.05976588651537895, -0.15375985205173492, -0.00870948750525713, -0.04503806680440903, 0.08287355303764343, 0.017666861414909363, 0.16653785109519958, -0.006960130762308836, -0.012972959317266941, -0.09847598522901535, 0.06441020965576172, -0.023426895961165428, 0.24133971333503723, -0.06349262595176697, -0.143473818898201, 0.22233937680721283, -0.06944137066602707, -0.1420930027961731, 0.12657590210437775, 0.020838137716054916, 0.07354387640953064, 0.10204131156206131, 0.19502143561840057, 0.014321080408990383, 0.006638950202614069, 0.054760824888944626, 0.0820525735616684, -0.17468082904815674, -0.03090454451739788, -0.00916894432157278, -0.01691337674856186, -0.1358218789100647, 0.043003637343645096, 0.11889315396547318, 0.10293904691934586, -0.07099705934524536, -0.013554582372307777, -0.03317642584443092, -0.004439891315996647, 0.06940841674804688, -0.007896981202065945, 0.09806080162525177, -0.09929461032152176, -0.040301576256752014, -0.058301206678152084, -0.006890235934406519, 0.0024084753822535276, 0.04120192304253578, -0.040066637098789215, 0.10096675902605057, -0.0006905568297952414, 0.05225489288568497, -0.14046427607536316, -0.07798092812299728, -0.020940499380230904, 0.1575685739517212, 0.04120592027902603, 0.04776391014456749, 0.057414863258600235, -0.0481991246342659, -0.0158701092004776, -0.009968787431716919, 0.16312187910079956, -0.0394844189286232, -0.06968989968299866, -0.056453973054885864, 0.10656707733869553, -0.058365050703287125, 0.03222460299730301, -0.04202231392264366, 0.022449776530265808, 0.060929615050554276, 0.12114907056093216, -0.0026538248639553785, 0.029062092304229736, -0.01023187953978777, -0.0038288652431219816, -0.07459788024425507, -0.020947640761733055, 0.10031349956989288, -0.004600553773343563, -0.08327498286962509, 0.23685328662395477, -0.17454755306243896, 0.19663569331169128, 0.2115958034992218, -0.2628093659877777, -0.024321777746081352, -0.06840608268976212, -0.05017746612429619, 0.003011465771123767, 0.05876095965504646, -0.04692309722304344, 0.09809201210737228, -0.02521132305264473, 0.16497591137886047, -0.04652651026844978, -0.07362692058086395, 0.016456644982099533, -0.05898859724402428, -0.0463954322040081, 0.0659271627664566, 0.08106416463851929, -0.15480613708496094, 0.18694530427455902, 0.20838376879692078, 0.07612863928079605, 0.19334357976913452, 0.004058185499161482, -0.014812501147389412, 0.08005014806985855, -0.03805047646164894, -0.04202093929052353, -0.07553261518478394, -0.16944189369678497, -0.01902174763381481, 0.07485251128673553, 0.03750864416360855, 0.11274250596761703, -0.10172852873802185, -0.07372885197401047, -0.016179129481315613, -0.005032413639128208, 0.005167648661881685, 0.1174640879034996, 0.045775800943374634, 0.14043675363063812, -0.019972821697592735, 0.03493902459740639, 0.08747350424528122, 0.02448674477636814, -0.10759711265563965, 0.16407065093517303, -0.14081640541553497, -0.38538745045661926, -0.16212545335292816, -0.13394121825695038, -0.029274288564920425, 0.04825805127620697, 0.11038947850465775, -0.13598975539207458, 0.0011978530092164874, -0.003706524148583412, 0.12342415004968643, -0.0806080624461174, 0.03755999356508255, -0.07838296890258789, 0.026997538283467293, -0.06349453330039978, -0.07917723804712296, -0.036463622003793716, -0.03232228383421898, -0.10000553727149963, 0.1757805496454239, -0.11054177582263947, 0.057571277022361755, 0.1741490364074707, 0.022440658882260323, 0.034390855580568314, -0.0513761006295681, 0.17275545001029968, -0.11779367178678513, 0.02093288116157055, 0.16278521716594696, -0.01799617148935795, 0.08678310364484787, 0.08059167861938477, -0.015366556122899055, -0.10777527838945389, 0.05196633189916611, 0.0019955262541770935, -0.1096891239285469, -0.20052047073841095, -0.12150565534830093, -0.0784008651971817, 0.14483654499053955, 0.05303339660167694, 0.05915789678692818, 0.17167222499847412, 0.08591149002313614, -0.04288473725318909, -0.004711467772722244, -0.012867298908531666, 0.07781979441642761, 0.1684085726737976, -0.017248503863811493, 0.11789125204086304, -0.05446818470954895, -0.11601924896240234, 0.13826869428157806, 0.02504623495042324, 0.050291191786527634, 0.04182872176170349, 0.008374262601137161, -0.009610554203391075, 0.09969738125801086, 0.12988939881324768, 0.118865467607975, -0.008107239380478859, -0.0232877004891634, -0.03601100295782089, -0.00860752072185278, -0.03570752218365669, 0.034572016447782516, 0.011757065542042255, -0.16013272106647491, -0.05726486071944237, -0.12173470109701157, 0.0964120477437973, 0.09787409752607346, 0.08039643615484238, -0.2033146470785141, -0.004589703865349293, 0.07378163933753967, -0.03603411093354225, -0.11624246090650558, 0.086527980864048, 0.033109065145254135, -0.1271866261959076, 0.0817195475101471, -0.03352120518684387, 0.115711510181427, -0.017423994839191437, 0.09427224844694138, -0.04346824064850807, -0.0329415462911129, -0.012381686829030514, 0.10430185496807098, -0.30799204111099243, 0.17485815286636353, -0.019660785794258118, -0.07034741342067719, -0.07672256976366043, -0.025566134601831436, 0.017929747700691223, 0.07530300319194794, 0.09619415551424026, 0.024311896413564682, 0.04642496258020401, -0.09243860840797424, -0.03940937668085098, 0.034113768488168716, 0.13641610741615295, -0.0638844221830368, -0.015862328931689262, -0.04075292870402336, 0.01116214320063591, -0.019626103341579437, -0.027355113998055458, 0.018256209790706635, -0.1504947543144226, 0.05358212813735008, 0.017237937077879906, 0.0753018707036972, 0.03889141231775284, -0.007973386906087399, -0.10224062204360962, 0.18268363177776337, -0.03004412353038788, -0.08521177619695663, -0.127006396651268, -0.04812724515795708, 0.04587927460670471, -0.051474425941705704, 0.034381575882434845, -0.06691597402095795, -0.011345877312123775, -0.06886660307645798, -0.21483656764030457, 0.12495172768831253, -0.0775398537516594, -0.07874035835266113, -0.03474915772676468, 0.20981398224830627, -0.05076101794838905, -0.00018431349599268287, 0.01172169204801321, 0.014822970144450665, -0.1086968258023262, -0.10466268658638, 0.06874550879001617, -0.034708425402641296, 0.02743770368397236, 0.02760813757777214, -0.03700246661901474, 0.02073092758655548, -0.06074898689985275, -0.01314478274434805, 0.2849438786506653, 0.22848764061927795, -0.035804633051157, 0.1875685602426529, 0.10711772739887238, -0.07248730212450027, -0.30828598141670227, -0.0999293103814125, -0.13133330643177032, -0.033457282930612564, -0.02052777260541916, -0.17081500589847565, 0.07089676707983017, 0.04558560997247696, 0.00998434517532587, 0.14900504052639008, -0.21140912175178528, -0.08518896996974945, 0.1092359870672226, -0.03466132655739784, 0.42140644788742065, -0.1164151057600975, -0.09652433544397354, -0.05340435355901718, -0.15239587426185608, 0.20086675882339478, -0.01557826716452837, 0.08761543780565262, -0.031178412958979607, 0.14360429346561432, 0.04995222017168999, -0.01680157333612442, 0.08329997956752777, 0.0014065488940104842, 0.004026432521641254, -0.12653036415576935, -0.022627411410212517, 0.0504734143614769, 0.021120509132742882, 0.0054380460642278194, -0.07736434042453766, 0.028706049546599388, -0.14863882958889008, -0.024921666830778122, -0.10909338295459747, 0.08278775960206985, 0.03857516124844551, -0.07422378659248352, -0.010892878286540508, -0.05615931376814842, -0.023589344695210457, -0.012528443709015846, 0.13453009724617004, -0.050522636622190475, 0.1720355898141861, 0.036824267357587814, 0.11522943526506424, -0.13816054165363312, 0.06146138161420822, -0.07406572997570038, -0.07532623410224915, 0.06773588061332703, -0.13577620685100555, 0.05240122601389885, 0.10075365751981735, -0.03372569754719734, 0.04758675396442413, 0.08927652984857559, 0.000919255951885134, 0.009403540752828121, 0.15953567624092102, -0.2761637568473816, 0.01791755110025406, -0.07046908140182495, -0.07692829519510269, 0.112159863114357, 0.07484955340623856, 0.181244894862175, 0.02791808731853962, -0.0472177192568779, 0.012221097014844418, 0.019269011914730072, -0.05120278522372246, 0.0548672117292881, 0.006578541360795498, -0.011816216632723808, -0.14148055016994476, 0.08790901303291321, -0.0015104453777894378, -0.1429632604122162, 0.02210722118616104, 0.19665491580963135, -0.13288317620754242, -0.10024755448102951, -0.05081937462091446, 0.057288672775030136, -0.13861924409866333, 0.008680099621415138, -0.01990172080695629, -0.09562872350215912, 0.0756591260433197, 0.1573420912027359, 0.05079081282019615, 0.12963363528251648, -0.02916029281914234, -0.008411908522248268, -0.04279141128063202, -0.051675889641046524, 0.02788730151951313, 0.019720058888196945, -0.07752392441034317, 0.08735781908035278, -0.024015765637159348, 0.14329436421394348, -0.10051412135362625, -0.06987810134887695, -0.1344141960144043, -0.005019306670874357, -0.09607285261154175, -0.0959741547703743, -0.08079583197832108, -0.061935000121593475, 0.004180733114480972, -0.039079975336790085, -0.0417400486767292, -0.08063078671693802, -0.10278100520372391, 0.016344387084245682, -0.02737213484942913, 0.028547246009111404, -0.07021904736757278, 0.008450011722743511, 0.12212047725915909, -0.029432062059640884, 0.17478644847869873, 0.15045183897018433, -0.10443241149187088, 0.10585668683052063, -0.17528948187828064, -0.10057486593723297, 0.10386897623538971, -0.01424362976104021, 0.02509693056344986, 0.12928596138954163, 0.018586233258247375, 0.04242280498147011, 0.03278495371341705, 0.06450606882572174, 0.04438134282827377, -0.11879310011863708, 0.08265755325555801, -0.002198860514909029, -0.15595629811286926, -0.061124756932258606, -0.09235648810863495, 0.02707000821828842, 0.02035105973482132, 0.11072126775979996, -0.04177020862698555, 0.08901136368513107, -0.06589431315660477, 0.023006385192275047, 0.025188656523823738, -0.17641755938529968, -0.03692740947008133, -0.05073147267103195, 0.03235582634806633, 0.028527792543172836, 0.22216679155826569, 0.015049049630761147, -0.030550595372915268, 0.0387740358710289, 0.12439186871051788, 0.015807392075657845, -0.0005460345419123769, 0.1709187924861908, 0.10334094613790512, -0.07364179939031601, -0.14011329412460327, 0.0676020160317421, 0.012262817472219467, -0.05599943920969963, 0.11458845436573029, -0.01812131516635418, -0.006866521667689085, 0.0670672059059143, -0.019556820392608643, 0.03449620306491852, -0.06607513129711151, -0.12762659788131714, -0.02478908933699131, 0.04269528388977051, 0.0029638311825692654, 0.12794137001037598, 0.15664103627204895, -0.005168106406927109, 0.026651626452803612, -0.015845872461795807, -0.023246217519044876, -0.13421551883220673, -0.15646639466285706, -0.06222613900899887, -0.14998294413089752, 0.022462334483861923, -0.08663632720708847, 0.047867026180028915, 0.05302877724170685, 0.07266844809055328, -0.06541765481233597, 0.06867647916078568, 0.049353908747434616, -0.11939282715320587, 0.08969518542289734, -0.024748487398028374, 0.04455447196960449, -0.0018985075876116753, -0.03077637031674385, -0.10359742492437363, 0.042003192007541656, -0.01649407483637333, 0.045346152037382126, -0.04421991854906082, 0.022579338401556015, -0.1750071942806244, -0.10995419323444366, -0.04818427562713623, 0.06838665902614594, -0.06683412194252014, 0.04056481271982193, 0.01606888137757778, 0.011690732091665268, 0.030521972104907036, 0.22083373367786407, -0.041100796312093735, -0.04384056478738785, -0.041319385170936584, 0.1635150909423828, -0.014866671524941921, 0.08733032643795013, -0.027159888297319412, -0.008534550666809082, -0.09159335494041443, 0.3551705777645111, 0.29626867175102234, -0.08988689631223679, 0.018322573974728584, -0.023403000086545944, 0.04054094851016998, 0.13749836385250092, 0.13930056989192963, 0.09693139791488647, 0.23870247602462769, -0.069560706615448, -0.05235563591122627, -0.01810343936085701, -0.013992605730891228, -0.06698887050151825, 0.09740167111158371, 0.02308662235736847, -0.05992421507835388, -0.0399320125579834, 0.09157036244869232, -0.2363831102848053, 0.1119081899523735, -0.1049022525548935, -0.15424089133739471, -0.03413557633757591, 0.011207441799342632, 0.07758733630180359, 0.01318280678242445, 0.11218693852424622, 0.013466300442814827, -0.0844072625041008, 0.014990749768912792, 0.034041877835989, -0.25270384550094604, 0.004782035481184721, 0.053680628538131714, -0.12391778826713562, -0.004765757359564304, -0.025234686210751534, 0.013625700026750565, 0.05700398609042168, 0.04094824194908142, -0.03069460764527321, 0.016096249222755432, -0.006652043666690588, -0.020144162699580193, -0.008288858458399773, 0.05953062325716019, 0.04713095352053642, -0.1559005230665207, 0.06380777060985565, -0.13577982783317566, 0.04075455665588379, -0.023885579779744148, -0.012463473714888096, -0.0012933483812958002, 0.01776987873017788, -0.0522073432803154, 0.06581555306911469, 0.07255802303552628, -0.0144086554646492, 0.008549283258616924, -0.08346639573574066, -0.034440234303474426, -0.023841137066483498, -0.10951124131679535, -0.08404874801635742, -0.13158579170703888, -0.12121031433343887, 0.10822955518960953, -0.02819518744945526, -0.18475614488124847, 0.03248962014913559, -0.12301548570394516, 0.05860345810651779, -0.17328192293643951, 0.11332228779792786, 0.07014153152704239, 0.018578365445137024, 0.01264720968902111, -0.007008096668869257, 0.08073210716247559, 0.11507359892129898, -0.0813264548778534, -0.08199906349182129 ]
null
null
transformers
<div> <div style="width: 132px; height:132px; border-radius: 50%; background-size: cover; background-image: url('https://pbs.twimg.com/profile_images/1080013557717192707/Cm9m3alh_400x400.jpg')"> </div> <div style="margin-top: 8px; font-size: 19px; font-weight: 800">┼aco d▲ve 🤖 AI Bot </div> <div style="font-size: 15px">@bibliobabble bot</div> </div> I was made with [huggingtweets](https://github.com/borisdayma/huggingtweets). Create your own bot based on your favorite user with [the demo](https://colab.research.google.com/github/borisdayma/huggingtweets/blob/master/huggingtweets-demo.ipynb)! ## How does it work? The model uses the following pipeline. ![pipeline](https://github.com/borisdayma/huggingtweets/blob/master/img/pipeline.png?raw=true) To understand how the model was developed, check the [W&B report](https://app.wandb.ai/wandb/huggingtweets/reports/HuggingTweets-Train-a-model-to-generate-tweets--VmlldzoxMTY5MjI). ## Training data The model was trained on [@bibliobabble's tweets](https://twitter.com/bibliobabble). | Data | Quantity | | --- | --- | | Tweets downloaded | 3206 | | Retweets | 77 | | Short tweets | 695 | | Tweets kept | 2434 | [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/3sdlkzqx/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline. ## Training procedure The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @bibliobabble's tweets. Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/e6mj14rl) for full transparency and reproducibility. At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/e6mj14rl/artifacts) is logged and versioned. ## How to use You can use this model directly with a pipeline for text generation: ```python from transformers import pipeline generator = pipeline('text-generation', model='huggingtweets/bibliobabble') generator("My dream is", num_return_sequences=5) ``` ## Limitations and bias The model suffers from [the same limitations and bias as GPT-2](https://huggingface.co/gpt2#limitations-and-bias). In addition, the data present in the user's tweets further affects the text generated by the model. ## About *Built by Boris Dayma* [![Follow](https://img.shields.io/twitter/follow/borisdayma?style=social)](https://twitter.com/intent/follow?screen_name=borisdayma) For more details, visit the project repository. [![GitHub stars](https://img.shields.io/github/stars/borisdayma/huggingtweets?style=social)](https://github.com/borisdayma/huggingtweets)
{"language": "en", "tags": ["huggingtweets"], "thumbnail": "https://www.huggingtweets.com/bibliobabble/1614116708796/predictions.png", "widget": [{"text": "My dream is"}]}
text-generation
huggingtweets/bibliobabble
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "huggingtweets", "en", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
┼aco d▲ve AI Bot @bibliobabble bot I was made with huggingtweets. Create your own bot based on your favorite user with the demo! How does it work? ----------------- The model uses the following pipeline. !pipeline To understand how the model was developed, check the W&B report. Training data ------------- The model was trained on @bibliobabble's tweets. Explore the data, which is tracked with W&B artifacts at every step of the pipeline. Training procedure ------------------ The model is based on a pre-trained GPT-2 which is fine-tuned on @bibliobabble's tweets. Hyperparameters and metrics are recorded in the W&B training run for full transparency and reproducibility. At the end of training, the final model is logged and versioned. How to use ---------- You can use this model directly with a pipeline for text generation: Limitations and bias -------------------- The model suffers from the same limitations and bias as GPT-2. In addition, the data present in the user's tweets further affects the text generated by the model. About ----- *Built by Boris Dayma* ![Follow](URL For more details, visit the project repository. ![GitHub stars](URL
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 57 ]
[ "passage: TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #huggingtweets #en #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 0.004547144751995802, -0.006708405911922455, -0.007013476919382811, 0.01947171241044998, 0.15818242728710175, 0.03448796644806862, 0.08709780126810074, 0.15389476716518402, -0.019877297803759575, -0.022431448101997375, 0.18047170341014862, 0.173692986369133, -0.012988686561584473, 0.09047263860702515, -0.05271327868103981, -0.2622397541999817, 0.03682216629385948, 0.05513067543506622, -0.007422737777233124, 0.14252057671546936, 0.07580838352441788, -0.023790201172232628, 0.11380083113908768, -0.02966974675655365, -0.202972412109375, 0.03197307139635086, 0.0615268237888813, -0.09518525749444962, 0.11083168536424637, 0.04628797993063927, 0.08698221296072006, 0.022143812850117683, -0.07331052422523499, -0.120787613093853, 0.04532235115766525, 0.045263588428497314, -0.06358368694782257, 0.06480421870946884, 0.08820623904466629, -0.1065920814871788, 0.1416475921869278, 0.07373794168233871, -0.01588049717247486, 0.07824484258890152, -0.17789237201213837, -0.03725104406476021, -0.036331940442323685, 0.007741300854831934, 0.07058489322662354, 0.0750737413764, -0.019116664305329323, 0.1746976524591446, -0.06598041951656342, 0.09777773916721344, 0.17528840899467468, -0.2887236773967743, -0.018040433526039124, 0.0492081381380558, 0.0887371376156807, 0.04900359362363815, -0.024227341637015343, 0.08339477330446243, 0.06365471333265305, 0.01686069741845131, 0.014271941967308521, -0.06960906833410263, -0.09346919506788254, 0.03645368665456772, -0.06932124495506287, -0.05699722096323967, 0.22001419961452484, -0.0334535576403141, 0.04674676060676575, -0.03953840583562851, -0.09316058456897736, -0.028927378356456757, -0.027232296764850616, -0.00907184462994337, -0.05413005128502846, 0.08754174411296844, -0.015151693485677242, -0.06331931799650192, -0.1435878872871399, -0.012912428006529808, -0.15805892646312714, 0.13816505670547485, 0.004333257209509611, 0.04586424678564072, -0.22094038128852844, 0.1012546569108963, 0.022817784920334816, -0.08995530754327774, 0.04930093511939049, -0.09425957500934601, 0.0717538446187973, 0.0007676240638829768, -0.04885277524590492, -0.02944779396057129, 0.08848895877599716, 0.14690880477428436, -0.02718975953757763, 0.005980455316603184, -0.01338018849492073, 0.0733228251338005, 0.059399042278528214, 0.028748195618391037, -0.006081049330532551, -0.052236080169677734, 0.05618719011545181, -0.1417204737663269, -0.010511515662074089, -0.07227712869644165, -0.10605388879776001, -0.04232237488031387, 0.03443120792508125, 0.060671232640743256, 0.042230576276779175, 0.11220116913318634, -0.04771716892719269, -0.01857093721628189, 0.05281376466155052, -0.03979083523154259, -0.008994937874376774, -0.01990325190126896, 0.018122754991054535, 0.13074275851249695, -0.019943278282880783, 0.03407962992787361, -0.10256942361593246, 0.05431444197893143, -0.10281401127576828, -0.01971535198390484, -0.014149561524391174, -0.04367954283952713, 0.031883664429187775, -0.12165860831737518, 0.016123656183481216, -0.16833168268203735, -0.14714312553405762, 0.002859292319044471, -0.016588665544986725, -0.017911825329065323, -0.07954888790845871, -0.04400517791509628, -0.02466505579650402, 0.06924423575401306, -0.04276731237769127, -0.00935916043817997, -0.05846982076764107, 0.11090090870857239, -0.05349889397621155, 0.07203050702810287, -0.1194647029042244, 0.0557217076420784, -0.14930842816829681, -0.013004516251385212, -0.04842504858970642, 0.07119924575090408, 0.015398351475596428, 0.1813964694738388, -0.006925920024514198, -0.003623353084549308, -0.09382472932338715, 0.06455672532320023, -0.02733452245593071, 0.24096953868865967, -0.0756828561425209, -0.14226967096328735, 0.21630549430847168, -0.06334739923477173, -0.14993034303188324, 0.1314547061920166, 0.01843975856900215, 0.08251222223043442, 0.10434340685606003, 0.19023460149765015, 0.01808990351855755, -0.007808534894138575, 0.054424818605184555, 0.07603957504034042, -0.1683882623910904, -0.033340878784656525, 0.0012923459289595485, -0.00014291972911451012, -0.1366809904575348, 0.04632483050227165, 0.1230006217956543, 0.09730340540409088, -0.07249721139669418, -0.018487868830561638, -0.030607668682932854, 0.0016078021144494414, 0.04144361615180969, -0.0005212334799580276, 0.09951234608888626, -0.1033509373664856, -0.04366454482078552, -0.06751791387796402, -0.002970147645100951, 0.011176802217960358, 0.03924661502242088, -0.04455869272351265, 0.09700342267751694, -0.007412149105221033, 0.0545678474009037, -0.13708296418190002, -0.07981666922569275, -0.016090448945760727, 0.1597585678100586, 0.040224816650152206, 0.04663374274969101, 0.0566885769367218, -0.05624469742178917, -0.015493324026465416, -0.010199432261288166, 0.16243304312229156, -0.04404180869460106, -0.07694169133901596, -0.07860849797725677, 0.10474636405706406, -0.06389671564102173, 0.026263169944286346, -0.051667314022779465, 0.024654213339090347, 0.04686986654996872, 0.1110762283205986, 0.004046999383717775, 0.026442723348736763, -0.012835992500185966, -0.007690808270126581, -0.07657550275325775, -0.01617686077952385, 0.1077079176902771, -0.0017721779877319932, -0.06809886544942856, 0.2437063455581665, -0.16884316504001617, 0.21163912117481232, 0.20976658165454865, -0.2492678016424179, -0.02882898785173893, -0.04848965257406235, -0.04766342043876648, -0.0012878701090812683, 0.06041788309812546, -0.034700244665145874, 0.09027024358510971, -0.03288675472140312, 0.16564396023750305, -0.051203593611717224, -0.07646744698286057, 0.019007064402103424, -0.05823178589344025, -0.05114857107400894, 0.07018019258975983, 0.08213616907596588, -0.1630844622850418, 0.18756183981895447, 0.21879082918167114, 0.06839460134506226, 0.2044064849615097, 0.00858453568071127, -0.010656360536813736, 0.07200875878334045, -0.04608747735619545, -0.03843220695853233, -0.06601633131504059, -0.15238076448440552, -0.03009703755378723, 0.06625645607709885, 0.030863380059599876, 0.09900964051485062, -0.09019728004932404, -0.08104760944843292, -0.017665131017565727, 0.004776675254106522, 0.00156646769028157, 0.11991100758314133, 0.03676433861255646, 0.13820022344589233, -0.01955524832010269, 0.022415857762098312, 0.08040772378444672, 0.016582515090703964, -0.10843544453382492, 0.16101348400115967, -0.13329310715198517, -0.3788211941719055, -0.14546175301074982, -0.13134250044822693, -0.020925991237163544, 0.03777816519141197, 0.1120775043964386, -0.1329103261232376, 0.005511005409061909, -0.007893978618085384, 0.10391844809055328, -0.08707519620656967, 0.039245378226041794, -0.07586963474750519, 0.0314689576625824, -0.060405436903238297, -0.07552991807460785, -0.03722400963306427, -0.028465405106544495, -0.09132689982652664, 0.16675986349582672, -0.11130212247371674, 0.06035055220127106, 0.16001324355602264, 0.021197395399212837, 0.03523072600364685, -0.05174810439348221, 0.18330632150173187, -0.112345851957798, 0.020098978653550148, 0.15624848008155823, -0.013005592860281467, 0.08254575729370117, 0.08188403397798538, -0.013132697902619839, -0.10316278785467148, 0.05240294709801674, 0.001463406952098012, -0.10209372639656067, -0.1950312703847885, -0.10119245946407318, -0.08230090886354446, 0.15922248363494873, 0.06361804902553558, 0.058937788009643555, 0.17968137562274933, 0.07578518986701965, -0.038606274873018265, -0.00038743947516195476, -0.00239798822440207, 0.08808282762765884, 0.13635766506195068, -0.01442645862698555, 0.1225903332233429, -0.04975935071706772, -0.10913994163274765, 0.12899059057235718, 0.01750512234866619, 0.03937286511063576, 0.051435839384794235, 0.021011192351579666, -0.011281835846602917, 0.11866551637649536, 0.13484057784080505, 0.10447502881288528, -0.015693627297878265, -0.0293489471077919, -0.04774824157357216, -0.01359935849905014, -0.033305928111076355, 0.03640862926840782, 0.008061517030000687, -0.14140670001506805, -0.06158366799354553, -0.11537835001945496, 0.08758961409330368, 0.10668005049228668, 0.07567808032035828, -0.21108253300189972, -0.003950516227632761, 0.07933880388736725, -0.03630997985601425, -0.11126025766134262, 0.08416172116994858, 0.03095286712050438, -0.1277567446231842, 0.07218055427074432, -0.03519461303949356, 0.12458370625972748, -0.0032897875644266605, 0.09583556652069092, -0.03598680719733238, -0.027483470737934113, -0.013308011926710606, 0.09818253666162491, -0.3191508650779724, 0.1621316522359848, -0.017933005467057228, -0.0618131123483181, -0.06667962670326233, -0.02528184838593006, 0.015994107350707054, 0.07729468494653702, 0.10861869156360626, 0.021759910508990288, 0.01640525460243225, -0.07345785945653915, -0.042352862656116486, 0.038021303713321686, 0.12403716146945953, -0.06827268749475479, -0.012903391383588314, -0.04523605480790138, 0.00796645786613226, -0.017124788835644722, -0.008793274872004986, 0.006911922711879015, -0.14962191879749298, 0.05182485654950142, 0.014736213721334934, 0.07058768719434738, 0.0436982735991478, -0.014969068579375744, -0.09180716425180435, 0.18274778127670288, -0.015714606270194054, -0.07271543145179749, -0.12616917490959167, -0.05262751132249832, 0.030376195907592773, -0.05518756061792374, 0.021047864109277725, -0.06501689553260803, -0.0035362408962100744, -0.06755607575178146, -0.22007296979427338, 0.1278373897075653, -0.08437205106019974, -0.07192739844322205, -0.04912353679537773, 0.2010866105556488, -0.051223888993263245, 0.003238252131268382, 0.010222852230072021, 0.021994104608893394, -0.11474784463644028, -0.09469719231128693, 0.07112357765436172, -0.03247172012925148, 0.03123478777706623, 0.0022505864035338163, -0.04091062396764755, 0.016593176871538162, -0.06314414739608765, -0.011381587944924831, 0.27866554260253906, 0.23951324820518494, -0.040407944470644, 0.1904350072145462, 0.11012271791696548, -0.08163551241159439, -0.3069863021373749, -0.10166139155626297, -0.12140648066997528, -0.02996143139898777, -0.017288926988840103, -0.16865339875221252, 0.06477722525596619, 0.038930367678403854, 0.009261871688067913, 0.13778774440288544, -0.20730599761009216, -0.08823523670434952, 0.09138026833534241, -0.02557477355003357, 0.43079736828804016, -0.1257614940404892, -0.08959750831127167, -0.051866497844457626, -0.16516901552677155, 0.2173919379711151, -0.021592965349555016, 0.07857322692871094, -0.029561417177319527, 0.11770006269216537, 0.04697660356760025, -0.010707763023674488, 0.08040876686573029, -0.00884756539016962, 0.008373050950467587, -0.12410011142492294, -0.02768467366695404, 0.04874192550778389, 0.012378438375890255, 0.0013600040692836046, -0.09389680624008179, 0.020313434302806854, -0.15990203619003296, -0.018549781292676926, -0.11233476549386978, 0.07682323455810547, 0.025788001716136932, -0.06466120481491089, -0.003637736663222313, -0.04986237734556198, -0.015892893075942993, -0.01400828268378973, 0.1717434972524643, -0.04862768203020096, 0.19366511702537537, 0.03501616790890694, 0.11570870876312256, -0.1362973153591156, 0.06143493950366974, -0.06429426372051239, -0.07528600096702576, 0.07427702099084854, -0.1537967324256897, 0.05111055448651314, 0.09430045634508133, -0.030276626348495483, 0.05380253866314888, 0.08795086294412613, -0.003969982732087374, 0.004800081253051758, 0.15867236256599426, -0.2786487936973572, 0.01320126373320818, -0.07396841049194336, -0.06665283441543579, 0.10506758838891983, 0.06261139363050461, 0.17162823677062988, 0.011681869626045227, -0.056615445762872696, 0.01595049723982811, 0.02499506063759327, -0.04915530979633331, 0.04529924690723419, 0.008104361593723297, -0.010991688817739487, -0.13640300929546356, 0.08699746429920197, 0.0042801909148693085, -0.1531187742948532, 0.024680746719241142, 0.2155698835849762, -0.1260155886411667, -0.10237220674753189, -0.03444112092256546, 0.08444061875343323, -0.11519137024879456, 0.01753072999417782, -0.030764780938625336, -0.09109894186258316, 0.07448896765708923, 0.15248911082744598, 0.049206193536520004, 0.11775100976228714, -0.015379221178591251, -0.011753370985388756, -0.05147303268313408, -0.0317845419049263, 0.025745956227183342, 0.017857374623417854, -0.08257177472114563, 0.06648801267147064, -0.022109810262918472, 0.14559012651443481, -0.09791336953639984, -0.06602771580219269, -0.1468091756105423, -0.009785634465515614, -0.0695481076836586, -0.09207163751125336, -0.08133620768785477, -0.062133077532052994, 0.0010387726360931993, -0.03962359577417374, -0.04795864596962929, -0.0791037380695343, -0.10289866477251053, 0.009435068815946579, -0.02305566892027855, 0.03256045654416084, -0.06115729361772537, 0.007872066460549831, 0.12092912197113037, -0.028174830600619316, 0.16686207056045532, 0.1458095908164978, -0.09536580741405487, 0.10568815469741821, -0.16346460580825806, -0.08964221179485321, 0.0939340740442276, -0.01729099079966545, 0.027899714186787605, 0.11666940152645111, 0.014932696707546711, 0.04195788502693176, 0.035977672785520554, 0.06045130267739296, 0.03587699308991432, -0.11899011582136154, 0.07665140181779861, 0.009481414221227169, -0.1612047255039215, -0.06303887814283371, -0.08555969595909119, 0.030386725440621376, 0.021575886756181717, 0.12225193530321121, -0.045776769518852234, 0.0887017622590065, -0.07972796261310577, 0.027257539331912994, 0.02293219044804573, -0.181223064661026, -0.047844018787145615, -0.053065262734889984, 0.032686229795217514, 0.018960151821374893, 0.1893557906150818, 0.027213018387556076, -0.03697650134563446, 0.04549255222082138, 0.1042066365480423, 0.005313898902386427, 0.004829791374504566, 0.16259528696537018, 0.09423433989286423, -0.07654286175966263, -0.12226779758930206, 0.07556461542844772, 0.019673259928822517, -0.044067107141017914, 0.10607215762138367, -0.002448870101943612, 0.020163848996162415, 0.06910120695829391, -0.014892932027578354, 0.034322552382946014, -0.044286008924245834, -0.10698256641626358, -0.023580113425850868, 0.046367425471544266, 0.00669879000633955, 0.12847968935966492, 0.177873894572258, -0.002574790036305785, 0.025011489167809486, -0.0363602340221405, -0.024931130930781364, -0.13864666223526, -0.1558164656162262, -0.06855984032154083, -0.14875617623329163, 0.012976853176951408, -0.0915176048874855, 0.04695429280400276, 0.028682325035333633, 0.06887643784284592, -0.07052405923604965, 0.04384735971689224, 0.06974220275878906, -0.12065785378217697, 0.09397104382514954, -0.028081456199288368, 0.03704333305358887, -0.006730496883392334, -0.012833851389586926, -0.10013298690319061, 0.035936567932367325, -0.01747855544090271, 0.045271266251802444, -0.04546798765659332, 0.030429324135184288, -0.1703072488307953, -0.124412901699543, -0.04034453630447388, 0.06420420855283737, -0.06510858237743378, 0.03512151539325714, 0.019115818664431572, 0.013339218683540821, 0.03305599465966225, 0.23020225763320923, -0.03704051673412323, -0.02329315058887005, -0.042310282588005066, 0.16692522168159485, -0.014016710221767426, 0.08088304847478867, -0.03037172369658947, 0.0002500463742762804, -0.08417443931102753, 0.3385351300239563, 0.3027777075767517, -0.09020252525806427, 0.019915465265512466, -0.030905582010746002, 0.03936264291405678, 0.11892254650592804, 0.13376617431640625, 0.09784641861915588, 0.2282467782497406, -0.07217609137296677, -0.03032243251800537, -0.020507147535681725, -0.011079044081270695, -0.06650827825069427, 0.0879674032330513, 0.02507801540195942, -0.05553486570715904, -0.031693898141384125, 0.0812700018286705, -0.2327648252248764, 0.10665327310562134, -0.11289316415786743, -0.1636168211698532, -0.039189815521240234, 0.0042042857967317104, 0.08908319473266602, 0.015396242961287498, 0.11228121817111969, 0.009163780137896538, -0.07585213333368301, 0.017798418179154396, 0.028085503727197647, -0.24201616644859314, -0.008133855648338795, 0.060310713946819305, -0.12939085066318512, -0.004324504639953375, -0.027167800813913345, 0.007199867628514767, 0.059822265058755875, 0.029368450865149498, -0.04319324716925621, -0.001257759635336697, -0.010450302623212337, -0.008644461631774902, -0.011618612334132195, 0.07065588980913162, 0.046958792954683304, -0.13329142332077026, 0.06869500875473022, -0.11774353682994843, 0.033477768301963806, -0.05866728723049164, -0.015255378559231758, 0.000037100471672602, 0.03460683673620224, -0.04829782620072365, 0.07058211416006088, 0.07688362896442413, -0.015606098808348179, 0.000610517687164247, -0.0802936851978302, -0.036274004727602005, -0.019796574488282204, -0.09252054989337921, -0.08371094614267349, -0.13031646609306335, -0.11573562026023865, 0.1029667928814888, -0.02224794402718544, -0.19213621318340302, 0.03111329674720764, -0.12165344506502151, 0.045619383454322815, -0.1751558482646942, 0.11076030135154724, 0.08046020567417145, 0.01831907220184803, 0.011516088619828224, -0.02576824277639389, 0.08821021765470505, 0.11728470027446747, -0.07783648371696472, -0.08528783172369003 ]