X-iZhang commited on
Commit
e9a88b3
·
verified ·
1 Parent(s): 1e2b51b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -1
README.md CHANGED
@@ -20,4 +20,31 @@ library_name: transformers
20
  For further details about Libra-v0.5, including its architecture, training process, and use cases, please refer to the following resources:
21
  - Article: Comprehensive paper describing Libra’s design and experiments [Gla-AI4BioMed at RRG24: Visual Instruction-tuned Adaptation for Radiology Report Generation](https://aclanthology.org/2024.bionlp-1.54/)
22
  - Code Repository: Open-source implementation and pre-trained models (GitHub: [Gla-AI4BioMed at RRG24](https://github.com/X-iZhang/RRG-BioNLP-ACL2024))
23
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  For further details about Libra-v0.5, including its architecture, training process, and use cases, please refer to the following resources:
21
  - Article: Comprehensive paper describing Libra’s design and experiments [Gla-AI4BioMed at RRG24: Visual Instruction-tuned Adaptation for Radiology Report Generation](https://aclanthology.org/2024.bionlp-1.54/)
22
  - Code Repository: Open-source implementation and pre-trained models (GitHub: [Gla-AI4BioMed at RRG24](https://github.com/X-iZhang/RRG-BioNLP-ACL2024))
23
+
24
+ ## How to Cite ✒️
25
+
26
+ If you use this model in academic or research contexts, please cite:
27
+
28
+ ```bibtex
29
+ @inproceedings{zhang-etal-2024-gla,
30
+ title = "Gla-{AI}4{B}io{M}ed at {RRG}24: Visual Instruction-tuned Adaptation for Radiology Report Generation",
31
+ author = "Zhang, Xi and
32
+ Meng, Zaiqiao and
33
+ Lever, Jake and
34
+ Ho, Edmond S.L.",
35
+ editor = "Demner-Fushman, Dina and
36
+ Ananiadou, Sophia and
37
+ Miwa, Makoto and
38
+ Roberts, Kirk and
39
+ Tsujii, Junichi",
40
+ booktitle = "Proceedings of the 23rd Workshop on Biomedical Natural Language Processing",
41
+ month = aug,
42
+ year = "2024",
43
+ address = "Bangkok, Thailand",
44
+ publisher = "Association for Computational Linguistics",
45
+ url = "https://aclanthology.org/2024.bionlp-1.54/",
46
+ doi = "10.18653/v1/2024.bionlp-1.54",
47
+ pages = "624--634",
48
+ abstract = "This paper introduces a radiology-focused visual language model designed to generate radiology reports from chest X-rays. Building on previous findings that large language models can acquire multimodal capabilities when aligned with pretrained vision encoders, we demonstrate similar potential with chest X-ray images. The model combines an image encoder (CLIP) with a fine-tuned large language model (LLM) based on the Vicuna-7B architecture. The training process involves a two-stage approach: initial alignment of chest X-ray features with the LLM, followed by fine-tuning for radiology report generation. The study highlights the importance of generating both FINDINGS and IMPRESSIONS sections in radiology reports and evaluates the model`s performance using various metrics, achieving notable accuracy in generating high-quality medical reports. The research also addresses the need for domain-specific fine-tuning to capture the intricate details necessary for accurate medical interpretations and reports."
49
+ }
50
+ ```