Johnesss commited on
Commit
0e43283
·
verified ·
1 Parent(s): 04a3f0c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -16,13 +16,14 @@ This model used for text classification with toxic and non-toxic labels.
16
  ## Intended uses & limitations
17
 
18
  If you want to reuse model, try copy this
 
19
  from huggingface_hub import from_pretrained_keras
20
  reloaded_model = from_pretrained_keras('Johnesss/Toxic-Comment-Classification')
21
 
22
  y_testing=reloaded_model.predict(x_testing,verbose=1,batch_size=32)
23
  test_df['Toxic']=['Not Toxic' if x<0.5 else 'Toxic' for x in y_testing]
24
  test_df[['comment_text','Toxic']].head(20)
25
-
26
  ## Training and evaluation data
27
 
28
  More information needed
 
16
  ## Intended uses & limitations
17
 
18
  If you want to reuse model, try copy this
19
+ ```
20
  from huggingface_hub import from_pretrained_keras
21
  reloaded_model = from_pretrained_keras('Johnesss/Toxic-Comment-Classification')
22
 
23
  y_testing=reloaded_model.predict(x_testing,verbose=1,batch_size=32)
24
  test_df['Toxic']=['Not Toxic' if x<0.5 else 'Toxic' for x in y_testing]
25
  test_df[['comment_text','Toxic']].head(20)
26
+ ```
27
  ## Training and evaluation data
28
 
29
  More information needed