Forecasting Whole-Brain Neuronal Activity from Volumetric Video
Abstract
Large-scale neuronal activity recordings with fluorescent calcium indicators are increasingly common, yielding high-resolution 2D or 3D videos. Traditional analysis pipelines reduce this data to 1D traces by segmenting regions of interest, leading to inevitable information loss. Inspired by the success of deep learning on minimally processed data in other domains, we investigate the potential of forecasting neuronal activity directly from volumetric videos. To capture long-range dependencies in high-resolution volumetric whole-brain recordings, we design a model with large receptive fields, which allow it to integrate information from distant regions within the brain. Our model outperforms trace-based forecasting approaches on ZAPBench, a benchmark on whole-brain activity prediction in zebrafish, demonstrating the advantages of preserving the spatial structure of neuronal activity. We perform extensive model selection, analyze effects of input resolution and spatio-temporal trade-offs, explore the impact of pre-training, provide comprehensive controls, and propose complementary performance metrics.