mamba-2.8b-hf / README.md
ArthurZ's picture
ArthurZ HF staff
Update README.md
03e6b8b verified
|
raw
history blame
762 Bytes
metadata
library_name: transformers
tags: []

Model Card for Model ID

This repository contains the transfromers compatible mamba-2.8b. The checkpoints are untouched, but the full config.json and tokenizer are pushed to this repo.

Usage

from transformers import MambaConfig, MambaForCausalLM, AutoTokenizer
import torch

tokenizer = AutoTokenizer.from_pretrained("state-spaces/mamba-2.8b-hf")
tokenizer.pad_token = tokenizer.eos_token

model = MambaForCausalLM.from_pretrained("state-spaces/mamba-2.8b-hf")
input_ids = tokenizer("Hey how are you doing?", return_tensors= "pt")["input_ids"]

out = model.generate(input_ids, max_new_tokens=10)
print(tokenizer.batch_decode(out))