GunaKoppula commited on
Commit
ff9cb56
Β·
verified Β·
1 Parent(s): e5701a3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +36 -43
app.py CHANGED
@@ -70,53 +70,46 @@ theme = gr.themes.Default(primary_hue="blue").set(
70
 
71
 
72
  with gr.Blocks(theme='upsatwal/mlsc_tiet') as demo:
73
- # gr.Markdown("## MulitModal Phi2 Model Pretraining and Finetuning from Scratch", justify="center")
74
-
75
- # with gr.Row() as title_row:
76
- # with gr.Column():
77
- # # Create an empty column on the left for spacing
78
- # pass
79
-
80
- # with gr.Column():
81
- # # Add Markdown with centered text
82
- # gr.Markdown("## MulitModal Phi2 Model Pretraining and Finetuning from Scratch")
83
- # gr.Markdown("This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/phi-2) model. Please find the source code and training details [here](https://github.com/RaviNaik/ERA-CAPSTONE/MultiModalPhi2).")
84
-
85
-
86
- # # with gr.Column():
87
- # # # Create an empty column on the right for spacing
88
- # # pass
89
 
90
- with gr.Column():
91
- # Add an empty line above the text for vertical centering
92
- gr.Text("")
93
 
94
- # Add the title text in the center
95
- gr.Text("MulitModal Phi2 Model Pretraining and Finetuning from Scratch", style={"text-align": "center"})
96
-
97
- # Add an empty line below the text for vertical centering
98
- gr.Text("")
 
 
 
 
 
 
99
 
100
- gr.Markdown(
101
- """This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/phi-2) model. Please find the source code and training details [here](https://github.com/RaviNaik/ERA-CAPSTONE/MultiModalPhi2).
102
 
103
- ### Details:
104
- 1. LLM Backbone: [Phi2](https://huggingface.co/microsoft/phi-2)
105
- 2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
106
- 3. Audio Model: [Whisper Tiny](https://huggingface.co/openai/whisper-tiny)
107
- 4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
108
- 5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
109
- 6. Finetuned Model: [GunaKoppula/Llava-Phi2](https://huggingface.co/GunaKoppula/Llava-Phi2)
110
- """
111
- )
112
-
113
- # with gr.Row(scale=8):
114
- # with gr.Box():
115
- # with gr.Row():
116
- # chatbot = gr.Chatbot(
117
- # avatar_images=("πŸ§‘", "πŸ€–"),
118
- # height=550,
119
- # )
 
 
 
 
 
 
 
120
 
121
  with gr.Row():
122
  chatbot = gr.Chatbot(
 
70
 
71
 
72
  with gr.Blocks(theme='upsatwal/mlsc_tiet') as demo:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
 
74
+ with gr.Row():
 
 
75
 
76
+ gr.Markdown("## MulitModal Phi2 Model Pretraining and Finetuning from Scratch")
77
+
78
+ # with gr.Row() as title_row:
79
+ # with gr.Column():
80
+ # # Create an empty column on the left for spacing
81
+ # pass
82
+
83
+ # with gr.Column():
84
+ # # Add Markdown with centered text
85
+ # gr.Markdown("## MulitModal Phi2 Model Pretraining and Finetuning from Scratch")
86
+ # gr.Markdown("This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/phi-2) model. Please find the source code and training details [here](https://github.com/RaviNaik/ERA-CAPSTONE/MultiModalPhi2).")
87
 
 
 
88
 
89
+ # # with gr.Column():
90
+ # # # Create an empty column on the right for spacing
91
+ # # pass
92
+
93
+ gr.Markdown(
94
+ """This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/phi-2) model. Please find the source code and training details [here](https://github.com/RaviNaik/ERA-CAPSTONE/MultiModalPhi2).
95
+
96
+ ### Details:
97
+ 1. LLM Backbone: [Phi2](https://huggingface.co/microsoft/phi-2)
98
+ 2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
99
+ 3. Audio Model: [Whisper Tiny](https://huggingface.co/openai/whisper-tiny)
100
+ 4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
101
+ 5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
102
+ 6. Finetuned Model: [GunaKoppula/Llava-Phi2](https://huggingface.co/GunaKoppula/Llava-Phi2)
103
+ """
104
+ )
105
+
106
+ # with gr.Row(scale=8):
107
+ # with gr.Box():
108
+ # with gr.Row():
109
+ # chatbot = gr.Chatbot(
110
+ # avatar_images=("πŸ§‘", "πŸ€–"),
111
+ # height=550,
112
+ # )
113
 
114
  with gr.Row():
115
  chatbot = gr.Chatbot(