Skip to content

rayford295/Sat2Street-DisasterGen

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

106 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Sat2Street-DisasterGen

Synthesizing Post-Disaster Street Views from Satellite Imagery via Generative Vision Models

arXiv IGARSS Python PyTorch License

Accepted at IEEE IGARSS 2026


Overview

Ground-level street-view imagery is critical for post-disaster damage assessment, yet is often unavailable immediately after events due to access constraints. This project benchmarks whether realistic and semantically consistent street views can be synthesized from satellite imagery, and evaluates how different generative paradigms perform under this cross-view setting.


Evaluated Methods

Method Paradigm Key Characteristic
Pix2Pix Conditional GAN Preserves spatial layout; prone to texture blur
SD 1.5 + ControlNet Diffusion High realism; may underestimate fine-grained damage
ControlNet + VLM (Gemini) Diffusion + Semantic Damage-aware; possible semantic–geometry tension
Disaster-MoE Mixture of Experts Severity-adaptive; decouples structure from texture

Results

Qualitative Comparison

Qualitative Comparison

Semantic Consistency (ResNet-18 Classifier)

Confusion Matrices

  • SD 1.5 + ControlNet achieves the highest semantic consistency (F1 ≈ 0.71)
  • Pix2Pix exhibits strong mode collapse toward mild damage predictions
  • Gemini-guided and MoE improve visual realism with a slight trade-off in class separability

Citation

@article{yang2026satellite,
  title   = {Satellite-to-Street: Synthesizing Post-Disaster Views from Satellite
             Imagery via Generative Vision Models},
  author  = {Yang, Yifan and Zou, Lei and Jepson, Wendy},
  journal = {arXiv preprint arXiv:2603.20697},
  year    = {2026}
}

Acknowledgement

Supported by the Texas A&M University Environment and Sustainability Initiative (ESI) through the Environment and Sustainability Graduate Fellow Award.


Contact

Yifan Yang — Department of Geography, Texas A&M University yyang295@tamu.edu · rayford295.github.io

About

A multimodal generation framework that synthesizes realistic street-view imagery from satellite and remote-sensing inputs, designed for post-disaster assessment, urban resilience analysis, and cross-view understanding.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages