Caleb Spradlin
commited on
Commit
·
76005a3
1
Parent(s):
ab687e7
added text and image
Browse files- app.py +5 -0
- data/.DS_Store +0 -0
- data/figures/reconstruction.png +0 -0
- text.py +22 -0
app.py
CHANGED
|
@@ -3,6 +3,7 @@ import numpy as np
|
|
| 3 |
import os
|
| 4 |
import pathlib
|
| 5 |
from inference import infer, InferenceModel
|
|
|
|
| 6 |
|
| 7 |
# -----------------------------------------------------------------------------
|
| 8 |
# class SatvisionDemoApp
|
|
@@ -77,6 +78,10 @@ class SatvisionDemoApp:
|
|
| 77 |
with col3:
|
| 78 |
st.image(output, use_column_width=True, caption="Reconstruction")
|
| 79 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 80 |
# -------------------------------------------------------------------------
|
| 81 |
# load_selected_image
|
| 82 |
# -------------------------------------------------------------------------
|
|
|
|
| 3 |
import os
|
| 4 |
import pathlib
|
| 5 |
from inference import infer, InferenceModel
|
| 6 |
+
from text import intro
|
| 7 |
|
| 8 |
# -----------------------------------------------------------------------------
|
| 9 |
# class SatvisionDemoApp
|
|
|
|
| 78 |
with col3:
|
| 79 |
st.image(output, use_column_width=True, caption="Reconstruction")
|
| 80 |
|
| 81 |
+
st.markdown(intro)
|
| 82 |
+
|
| 83 |
+
st.image('data/figures/reconstruction.png')
|
| 84 |
+
|
| 85 |
# -------------------------------------------------------------------------
|
| 86 |
# load_selected_image
|
| 87 |
# -------------------------------------------------------------------------
|
data/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
data/figures/reconstruction.png
ADDED
|
text.py
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
|
| 3 |
+
intro = '''Remote sensing images from NASA's fleet of Earth-observing satellites are pivotal for applications as broad as land cover mapping,
|
| 4 |
+
disaster monitoring, urban planning, and environmental analysis. The potential of AI-based geospatial foundation models for performing
|
| 5 |
+
visual analysis tasks on these remote sensing images has garnered significant attention. To realize that potential, the crucial first
|
| 6 |
+
step is to develop foundation models – computer models that acquire competence in a broad range of tasks, which can then be specialized
|
| 7 |
+
with further training for specific applications. In this case, the foundation model is based on a large-scale vision transformer model
|
| 8 |
+
trained with satellite imagery.
|
| 9 |
+
|
| 10 |
+
Vision transformers employ AI/deep learning techniques to fine-tune the model to answer specific science questions. Through training
|
| 11 |
+
on extensive remote sensing datasets, vision transformers can learn general relationships between the spectral data given as inputs,
|
| 12 |
+
as well as capture high-level visual patterns, semantics, and spatial relationships that can be leveraged for a wide range of analysis tasks.
|
| 13 |
+
Trained vision transformers can handle large-scale, high-resolution data; learn global reorientations; extract robust features; and support
|
| 14 |
+
multi-modal data fusion – all with improved performance.
|
| 15 |
+
|
| 16 |
+
The Data Science Group at NASA Goddard Space Flight Center's Computational and Information Sciences and Technology Office (CISTO)
|
| 17 |
+
has implemented an end-to-end workflow to generate a pre-trained vision transformer which could evolve into a foundation model.
|
| 18 |
+
A training dataset of over 2 million 128x128 pixel “chips” has been created from NASA’s Moderate Resolution Imaging Spectroradiometer (MODIS)
|
| 19 |
+
surface reflectance products (MOD09). These data were used to train a SwinV2 vision transformer that we call SatVision.
|
| 20 |
+
'''
|
| 21 |
+
|
| 22 |
+
|