Fill-Mask
Transformers
PyTorch
English
bert
exbert
Ozan Oktay commited on
Commit
1d47dbc
·
1 Parent(s): e5a7bfb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -33,7 +33,7 @@ First, we pretrain **CXR-BERT-general** from a randomly initialized BERT model v
33
 
34
  ```
35
  @misc{https://doi.org/10.48550/arxiv.2204.09817,
36
- title = {Making the Most of Text Semantics to Improve Biomedical Vision--Language Processing},
37
  author = {Boecking, Benedikt and Usuyama, Naoto and Bannur, Shruthi and Castro, Daniel C. and Schwaighofer, Anton and Hyland, Stephanie and Wetscherek, Maria and Naumann, Tristan and Nori, Aditya and Alvarez-Valle, Javier and Poon, Hoifung and Oktay, Ozan},
38
  publisher = {arXiv},
39
  year = {2022},
@@ -91,7 +91,7 @@ CXR-BERT also contributes to better vision-language representation learning thro
91
  | **BioViL** | **CXR-BERT** | **1.027** |
92
  | **BioViL-L** | **CXR-BERT** | **1.142** |
93
 
94
- Additional details about performance can be found in the corresponding paper, [Making the Most of Text Semantics to Improve Biomedical Vision--Language Processing](https://arxiv.org/abs/2204.09817).
95
 
96
  ## Limitations
97
 
@@ -99,4 +99,4 @@ This model was developed using English corpora, and thus can be considered Engli
99
 
100
  ## More Information
101
 
102
- Refer to the corresponding paper, [Making the Most of Text Semantics to Improve Biomedical Vision--Language Processing](https://arxiv.org/abs/2204.09817) for additional details and performance information.
 
33
 
34
  ```
35
  @misc{https://doi.org/10.48550/arxiv.2204.09817,
36
+ title = {Making the Most of Text Semantics to Improve Biomedical Vision-Language Processing},
37
  author = {Boecking, Benedikt and Usuyama, Naoto and Bannur, Shruthi and Castro, Daniel C. and Schwaighofer, Anton and Hyland, Stephanie and Wetscherek, Maria and Naumann, Tristan and Nori, Aditya and Alvarez-Valle, Javier and Poon, Hoifung and Oktay, Ozan},
38
  publisher = {arXiv},
39
  year = {2022},
 
91
  | **BioViL** | **CXR-BERT** | **1.027** |
92
  | **BioViL-L** | **CXR-BERT** | **1.142** |
93
 
94
+ Additional details about performance can be found in the corresponding paper, [Making the Most of Text Semantics to Improve Biomedical Vision-Language Processing](https://arxiv.org/abs/2204.09817).
95
 
96
  ## Limitations
97
 
 
99
 
100
  ## More Information
101
 
102
+ Refer to the corresponding paper, [Making the Most of Text Semantics to Improve Biomedical Vision-Language Processing](https://arxiv.org/abs/2204.09817) for additional details and performance information.