wannaphong commited on
Commit
71b8d91
·
1 Parent(s): 8ddee2d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -6
app.py CHANGED
@@ -4,7 +4,7 @@ import gradio as gr
4
  import torch
5
  torch.set_num_threads(1)
6
  from transformers import AutoModelForCausalLM, AutoTokenizer
7
- from typing import Optional, Union, List, Dict, Any
8
  import random
9
  import time
10
  import datetime
@@ -31,6 +31,10 @@ from typing import List
31
  import numpy as np
32
  import tensorflow_hub as hub
33
  import tensorflow_text
 
 
 
 
34
 
35
 
36
  class Encoder(ABC):
@@ -52,11 +56,7 @@ class MUSEEncoder(Encoder):
52
  return embeds
53
 
54
 
55
- from dataclasses import dataclass
56
- from typing import Dict, List, Tuple
57
 
58
- import numpy as np
59
- import tensorflow as tf
60
 
61
 
62
  @dataclass
@@ -132,7 +132,6 @@ model = AutoModelForCausalLM.from_pretrained(
132
  offload_folder="./",
133
  low_cpu_mem_usage=True,
134
  )
135
- model=torch.compile(model)
136
  tokenizer = AutoTokenizer.from_pretrained("facebook/xglm-7.5B")
137
 
138
 
@@ -294,6 +293,21 @@ with gr.Blocks(height=900) as demo:
294
  verbose=True,
295
  memory=ConversationBufferWindowMemory(k=2),
296
  )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
297
  with gr.Tab("Text Generation"):
298
  with gr.Row():
299
  with gr.Column():
 
4
  import torch
5
  torch.set_num_threads(1)
6
  from transformers import AutoModelForCausalLM, AutoTokenizer
7
+ from typing import Optional, Union, List, Dict, Any, Tuple
8
  import random
9
  import time
10
  import datetime
 
31
  import numpy as np
32
  import tensorflow_hub as hub
33
  import tensorflow_text
34
+ from dataclasses import dataclass
35
+
36
+ import numpy as np
37
+ import tensorflow as tf
38
 
39
 
40
  class Encoder(ABC):
 
56
  return embeds
57
 
58
 
 
 
59
 
 
 
60
 
61
 
62
  @dataclass
 
132
  offload_folder="./",
133
  low_cpu_mem_usage=True,
134
  )
 
135
  tokenizer = AutoTokenizer.from_pretrained("facebook/xglm-7.5B")
136
 
137
 
 
293
  verbose=True,
294
  memory=ConversationBufferWindowMemory(k=2),
295
  )
296
+ gr.Markdown(
297
+ """
298
+ # 🐘 WangChanGLM v0.2 demo
299
+
300
+ [Blog](https://medium.com/@iwishcognitivedissonance/wangchanglm-the-thai-turned-multilingual-instruction-following-model-7aa9a0f51f5f) | [Codes](https://github.com/pythainlp/wangchanglm) | [Demo](https://colab.research.google.com/github/pythainlp/WangChanGLM/blob/main/demo/WangChanGLM_v0_1_demo.ipynb)
301
+
302
+ ![](https://i.imgur.com/RL5Mqs3.png)
303
+
304
+
305
+ This demo use CPU only, so It may be slow or very slow. If you want the speed, try [Google colab](https://colab.research.google.com/github/pythainlp/WangChanGLM/blob/main/demo/WangChanGLM_v0_1_demo.ipynb).
306
+
307
+
308
+ **We do not guarantee a reply message.**
309
+ """
310
+ )
311
  with gr.Tab("Text Generation"):
312
  with gr.Row():
313
  with gr.Column():