faceless-void commited on
Commit
7b34995
·
verified ·
1 Parent(s): a876c8b

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ProtBERT-Unmasking
2
+
3
+ This model is a fine-tuned version of ProtBERT specifically optimized for unmasking protein sequences. It can predict masked amino acids in protein sequences based on the surrounding context.
4
+
5
+ ## Model Description
6
+
7
+ - **Base Model**: ProtBERT
8
+ - **Task**: Protein Sequence Unmasking
9
+ - **Training**: Fine-tuned on masked protein sequences
10
+ - **Use Case**: Predicting missing or masked amino acids in protein sequences
11
+
12
+ ## Usage
13
+
14
+ ```python
15
+ from transformers import AutoModelForMaskedLM, AutoTokenizer
16
+
17
+ # Load model and tokenizer
18
+ model = AutoModelForMaskedLM.from_pretrained("faceless-void/protbert-sequence-unmasking")
19
+ tokenizer = AutoTokenizer.from_pretrained("faceless-void/protbert-sequence-unmasking")
20
+
21
+ # Example usage
22
+ sequence = "MALN[MASK]KFGP[MASK]LVRK"
23
+ inputs = tokenizer(sequence, return_tensors="pt")
24
+ outputs = model(**inputs)
25
+ predictions = outputs.logits
26
+ ```
27
+
28
+ ## Limitations and Biases
29
+
30
+ This model is specifically designed for protein sequence unmasking and may not perform optimally for other protein-related tasks.
31
+
32
+ ## Training Details
33
+
34
+ The model was fine-tuned from the original ProtBERT model with specific focus on masked sequence prediction.