Update README.md
Browse files
README.md
CHANGED
@@ -20,4 +20,31 @@ library_name: transformers
|
|
20 |
For further details about Libra-v0.5, including its architecture, training process, and use cases, please refer to the following resources:
|
21 |
- Article: Comprehensive paper describing Libra’s design and experiments [Gla-AI4BioMed at RRG24: Visual Instruction-tuned Adaptation for Radiology Report Generation](https://aclanthology.org/2024.bionlp-1.54/)
|
22 |
- Code Repository: Open-source implementation and pre-trained models (GitHub: [Gla-AI4BioMed at RRG24](https://github.com/X-iZhang/RRG-BioNLP-ACL2024))
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
For further details about Libra-v0.5, including its architecture, training process, and use cases, please refer to the following resources:
|
21 |
- Article: Comprehensive paper describing Libra’s design and experiments [Gla-AI4BioMed at RRG24: Visual Instruction-tuned Adaptation for Radiology Report Generation](https://aclanthology.org/2024.bionlp-1.54/)
|
22 |
- Code Repository: Open-source implementation and pre-trained models (GitHub: [Gla-AI4BioMed at RRG24](https://github.com/X-iZhang/RRG-BioNLP-ACL2024))
|
23 |
+
|
24 |
+
## How to Cite ✒️
|
25 |
+
|
26 |
+
If you use this model in academic or research contexts, please cite:
|
27 |
+
|
28 |
+
```bibtex
|
29 |
+
@inproceedings{zhang-etal-2024-gla,
|
30 |
+
title = "Gla-{AI}4{B}io{M}ed at {RRG}24: Visual Instruction-tuned Adaptation for Radiology Report Generation",
|
31 |
+
author = "Zhang, Xi and
|
32 |
+
Meng, Zaiqiao and
|
33 |
+
Lever, Jake and
|
34 |
+
Ho, Edmond S.L.",
|
35 |
+
editor = "Demner-Fushman, Dina and
|
36 |
+
Ananiadou, Sophia and
|
37 |
+
Miwa, Makoto and
|
38 |
+
Roberts, Kirk and
|
39 |
+
Tsujii, Junichi",
|
40 |
+
booktitle = "Proceedings of the 23rd Workshop on Biomedical Natural Language Processing",
|
41 |
+
month = aug,
|
42 |
+
year = "2024",
|
43 |
+
address = "Bangkok, Thailand",
|
44 |
+
publisher = "Association for Computational Linguistics",
|
45 |
+
url = "https://aclanthology.org/2024.bionlp-1.54/",
|
46 |
+
doi = "10.18653/v1/2024.bionlp-1.54",
|
47 |
+
pages = "624--634",
|
48 |
+
abstract = "This paper introduces a radiology-focused visual language model designed to generate radiology reports from chest X-rays. Building on previous findings that large language models can acquire multimodal capabilities when aligned with pretrained vision encoders, we demonstrate similar potential with chest X-ray images. The model combines an image encoder (CLIP) with a fine-tuned large language model (LLM) based on the Vicuna-7B architecture. The training process involves a two-stage approach: initial alignment of chest X-ray features with the LLM, followed by fine-tuning for radiology report generation. The study highlights the importance of generating both FINDINGS and IMPRESSIONS sections in radiology reports and evaluates the model`s performance using various metrics, achieving notable accuracy in generating high-quality medical reports. The research also addresses the need for domain-specific fine-tuning to capture the intricate details necessary for accurate medical interpretations and reports."
|
49 |
+
}
|
50 |
+
```
|