metadata
library_name: transformers
tags: []
Model Card for Model ID
This repository contains the transfromers
compatible mamba-2.8b
. The checkpoints are untouched, but the full config.json
and tokenizer are pushed to this repo.
Usage
from transformers import MambaConfig, MambaForCausalLM, AutoTokenizer
import torch
tokenizer = AutoTokenizer.from_pretrained("state-spaces/mamba-2.8b-hf")
tokenizer.pad_token = tokenizer.eos_token
model = MambaForCausalLM.from_pretrained("state-spaces/mamba-2.8b-hf")
input_ids = tokenizer("Hey how are you doing?", return_tensors= "pt")["input_ids"]
out = model.generate(input_ids, max_new_tokens=10)
print(tokenizer.batch_decode(out))