[go: up one dir, main page]

Skip to content

Image Reconstruction from Brain fMRI imaging using Diffusion Models

License

Notifications You must be signed in to change notification settings

ahmed-alllam/BrainLens

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Contributors Forks Stargazers Issues LinkedIn


fmri

BrainLens: Image Reconstruction from Brain fMRI imaging using Diffusion Models

Report Bug · Request Feature

About The Project

This project aims to reconstruct visual images from brain fMRI data using diffusion models. The project is inspired by the paper Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors.

The basic idea of the project is to train a model to reconstruct images from fMRI data. The model is split into two main pipelines, one to capute higher-level semantic features of the image from fMRI data, and the other one to capture the lower-level details of the image. The first pipeline is trained using contrastive learning to make the latent representation of the fMRI similar to the CLIP embedding of the image, and thus the model is able to capture the semantic features of the image. The second pipeline is trained using diffusion autoencoder models to capture the lower-level details of the image. The two pipelines are then combined to reconstruct the image from the fMRI data using Versatile Diffusion model.

Note This project is still under development and is not yet ready for deployment, but feel free to fork it and use it for your own purposes!

Data Source

The dataset used in this project is from the Natural Scenes Dataset (NSD), which is a large-scale fMRI dataset conducted at ultra-high-field (7T) strength at the Center of Magnetic Resonance Research (CMRR) at the University of Minnesota. The dataset consists of whole-brain, high-resolution fMRI measurements of 8 healthy adult subjects while they viewed thousands of color natural scenes over the course of 30–40 scan sessions. While viewing these images, subjects were engaged in a continuous recognition task in which they reported whether they had seen each given image at any point in the experiment. 1

Tech Stack

The project utilizes the following technologies and libraries:

  • PyTorch
  • NumPy
  • Hugging Face
  • Diffusers
  • MLflow
  • DVC

References

Footnotes

  1. Natural Scenes Dataset (NSD)