Skip to content

dsrestrepo/MIT_Multimodal_Satellite_Images_Fusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Multimodal Satellite Images Fusion for Dengue Prediction and Socioeconomic Analysis

Overview

This repository contains the implementation of a data fusion framework that combines satellite images and tabular data for dengue prediction and socioeconomic analysis. The framework leverages variational autoencoders (VAE) for generating embeddings and employs an LSTM-based data fusion model to integrate time series of embeddings and metadata. The experiments in this repository demonstrate the effectiveness of the proposed approach in predicting dengue outbreaks and analyzing socioeconomic indicators like poverty and access to water.

Table of Contents

Introduction

Data fusion of satellite images and tabular data has shown promising results in various domains. In this project, we propose a multimodal data fusion framework that combines the visual information from satellite images with contextual tabular data to predict dengue outbreaks and analyze socioeconomic indicators. The framework consists of two main components: (1) Variational Autoencoders (VAE) for generating embeddings from satellite images and (2) an LSTM-based data fusion model to integrate the embeddings with temporal metadata.

Installation

To use the code in this repository, follow these steps:

  1. Clone the repository:
git clone https://github.com/dsrestrepo/MIT_Multimodal_Satellite_Images_Fusion.git
cd MIT_Multimodal_Satellite_Images_Fusion
  1. Set up the Python environment. It is recommended to use a virtual environment:
python -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate
  1. Install the required dependencies:
pip install -r requirements.txt

Usage

Once you have installed the required dependencies, you can use the code for generating embeddings, training the fusion model, and conducting experiments on dengue prediction and socioeconomic analysis.

Experiments

This repository includes the following experiments:

  1. Embedding Generation: Implementation of variational autoencoders (VAE) for generating embeddings from satellite images.
  2. Data Fusion Model: Implementation of the LSTM-based data fusion model that integrates the time series of embeddings with temporal metadata.
  3. Dengue Prediction: Using the fused embeddings to predict dengue outbreaks based on the multimodal data.
  4. Socioeconomic Analysis: Utilizing the fused embeddings to analyze socioeconomic indicators like poverty and access to water.

Results

The results of the experiments are provided in the results directory. We include evaluation metrics and visualizations that demonstrate the performance of the data fusion framework for dengue prediction and socioeconomic analysis.

Contributing

We welcome contributions to this repository! If you find any issues or have improvements to suggest, please open an issue or a pull request.

License

This project is licensed under the MIT License.

Contact

For any questions or inquiries, please feel free to reach out to David Restrepo at [email protected].

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published