Kkordik commited on
Commit
023aaad
·
verified ·
1 Parent(s): be7c477

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -10,7 +10,7 @@ subprocess.run('pip install mistral_inference mamba-ssm --no-build-isolation', e
10
  subprocess.run('pip install causal-conv1d --no-build-isolation', env={'CAUSAL_CONV1D_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
11
 
12
  # Import after installation
13
- from mistral_inference.transformer import Transformer
14
  from mistral_inference.generate import generate
15
  from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
16
  from mistral_common.protocol.instruct.messages import UserMessage, AssistantMessage
@@ -28,7 +28,7 @@ MODEL_PATH = str(mistral_models_path)
28
 
29
  # Load model and tokenizer
30
  tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
31
- model = Transformer.from_folder(MODEL_PATH)
32
 
33
 
34
  @spaces.GPU()
 
10
  subprocess.run('pip install causal-conv1d --no-build-isolation', env={'CAUSAL_CONV1D_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
11
 
12
  # Import after installation
13
+ from mistral_inference.mamba import Mamba
14
  from mistral_inference.generate import generate
15
  from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
16
  from mistral_common.protocol.instruct.messages import UserMessage, AssistantMessage
 
28
 
29
  # Load model and tokenizer
30
  tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
31
+ model = Mamba.from_folder(MODEL_PATH)
32
 
33
 
34
  @spaces.GPU()