mohammadmahdinouri commited on
Commit
41c6323
·
1 Parent(s): 2f63d6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -1,3 +1,25 @@
1
  ---
2
- license: mit
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: README
3
+ emoji: 🏃
4
+ colorFrom: gray
5
+ colorTo: purple
6
+ sdk: static
7
+ pinned: false
8
  ---
9
+
10
+ # Model Description
11
+ TinyClinicalBERT is a distilled version of the [BioClinicalBERT](https://huggingface.co/emilyalsentzer/Bio_ClinicalBERT) which is distilled for 3 epochs using a total batch size of 192 on the MIMIC-III notes dataset.
12
+
13
+ # Distillation Procedure
14
+ This model uses a unique distillation method called ‘transformer-layer distillation’ which is applied on each layer of the student to align the attention maps and the hidden states of the student with those of the teacher.
15
+
16
+ # Architecture and Initialisation
17
+ This model uses 4 hidden layers with a hidden dimension size and an embedding size of 768 resulting in a total of 15M parameters. Due to the model's small hidden dimension size, it uses random initialisation.
18
+
19
+ # Citation
20
+
21
+ If you use this model, please consider citing the following paper:
22
+
23
+ ```bibtex
24
+
25
+ ```