{"id":204977,"date":"2025-01-31T00:07:15","date_gmt":"2025-01-31T06:07:15","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2025\/01\/a-neural-network-for-long-term-super-resolution-imaging-of-live-cells-with-reliable-confidence-quantification"},"modified":"2025-01-31T00:07:15","modified_gmt":"2025-01-31T06:07:15","slug":"a-neural-network-for-long-term-super-resolution-imaging-of-live-cells-with-reliable-confidence-quantification","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2025\/01\/a-neural-network-for-long-term-super-resolution-imaging-of-live-cells-with-reliable-confidence-quantification","title":{"rendered":"A neural network for long-term super-resolution imaging of live cells with reliable confidence quantification"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/a-neural-network-for-long-term-super-resolution-imaging-of-live-cells-with-reliable-confidence-quantification.jpg\"><\/a><\/p>\n<p>TISR or video SR (VSR) neural network models are designed to leverage temporal neighbor frames to assist the SR of the current frame and are, therefore, expected to achieve better performance than SISR models<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 19\" title=\"Liu, H. et al. Video super-resolution based on deep learning: a comprehensive survey. Artif. Intell. Rev. 55, 5981&ndash;6035 (2022).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR19\" id=\"ref-link-section-d57257314e729\">19<\/a><\/sup> (Supplementary Note <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#MOESM1\">1<\/a>). Although TISR models have been widely explored in natural image SR to improve video definition, whether such models can be applied to super-resolve biological images (that is, enhancing both sampling rate and optical resolution) has been poorly investigated. Here, we used the total internal reflection fluorescence (TIRF) SIM, grazing incidence (GI) SIM and nonlinear SIM<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 20\" title=\"Li, D. et al. Extended-resolution structured illumination imaging of endocytic and cytoskeletal dynamics. Science 349, aab3500 (2015).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR20\" id=\"ref-link-section-d57257314e736\">20<\/a><\/sup> modes of our home-built Multi-SIM system to acquire an extensive TISR dataset of five different biological structures: clathrin-coated pits (CCPs), lysosomes, outer mitochondrial membranes (Mitos), microtubules (MTs) and F-actin filaments (Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig6\">1<\/a>). For each type of specimen, we generally acquired over 50 sets of raw SIM images with 20 consecutive time points at 2\u20134 levels of excitation light intensity (<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"section anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Sec9\">Methods<\/a>). Each set of raw SIM images was averaged out to a diffraction-limited wide-field (WF) image sequence and was used as the network input, while the raw SIM images acquired at the highest excitation level were reconstructed into SR-SIM images as the ground truth (GT) used in network training. In particular, the image acquisition configuration was modified into a special running order where each illumination pattern is applied 2\u20134 times at escalating excitation light intensity before changing to the next phase or orientation, so as to minimize the motion-induced difference between WF inputs and SR-SIM targets.<\/p>\n<p>To effectively use the temporal continuity of time-lapse data, SOTA TISR neural networks consist of mainly two important components<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 21\" title=\"Chan, K. C., Zhou, S., Xu, X. & Loy, C. C. BasicVSR++: improving video super-resolution with enhanced propagation and alignment. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (ed. O\u2019Conner, L.) 5972&ndash;5981 (IEEE, 2022).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR21\" id=\"ref-link-section-d57257314e749\">21<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 22\" title=\"Chan, K. C., Wang, X., Yu, K., Dong, C. & Loy, C. C. BasicVSR: the search for essential components in video super-resolution and beyond. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (ed. O\u2019Conner, L.) 4947&ndash;4956 (IEEE, 2021).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR22\" id=\"ref-link-section-d57257314e752\">22<\/a><\/sup>: temporal information propagation and neighbor feature alignment. We selected two popular types of propagation approaches, sliding window (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1a<\/a>) and recurrent network (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1b<\/a>), and three representative neighbor feature alignment mechanisms, explicit warping using OF<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 15\" title=\"Wang, L. et al. Deep video super-resolution using HR optical flow estimation. IEEE Trans. Image Process. 29, 4323&ndash;4336 (2020).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR15\" id=\"ref-link-section-d57257314e762\">15<\/a><\/sup> (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1c<\/a>) and implicit alignment by nonlocal attention<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 23\" title=\"Yu, J., Liu, J., Bo, L. & Mei, T. Memory-augmented non-local attention for video super-resolution. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (ed. O\u2019Conner, L.) 17834&ndash;17843 (IEEE, 2022).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR23\" id=\"ref-link-section-d57257314e770\">23<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 24\" title=\"Wang, X., Girshick, R., Gupta, A. & He, K. Non-local neural networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (ed. O\u2019Conner, L.) 7794&ndash;7803 (IEEE, 2018).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR24\" id=\"ref-link-section-d57257314e773\">24<\/a><\/sup> (NA; Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1d<\/a>) or deformable convolution<sup><a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 21\" title=\"Chan, K. C., Zhou, S., Xu, X. & Loy, C. C. BasicVSR++: improving video super-resolution with enhanced propagation and alignment. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (ed. O\u2019Conner, L.) 5972&ndash;5981 (IEEE, 2022).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR21\" id=\"ref-link-section-d57257314e780\">21<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 25\" title=\"Wang, X., Chan, K. C., Yu, K., Dong, C. & Loy, C. C. EDVR: video restoration with enhanced deformable convolutional networks. In Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (ed. O\u2019Conner, L.) 0&ndash;0 (IEEE, 2019).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR25\" id=\"ref-link-section-d57257314e783\">25<\/a>,<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 26\" title=\"Dai, J. et al. Deformable convolutional networks. In Proc. IEEE International Conference on Computer Vision (ed. O\u2019Conner, L.) 764&ndash;773 (2017).\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#ref-CR26\" id=\"ref-link-section-d57257314e786\">26<\/a><\/sup> (DC; Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1e<\/a>), resulting in six combinations in total. For fair comparison, we custom-designed a general TISR network architecture composed of a feature extraction module, a propagation and alignment module and a reconstruction module (Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig7\">2<\/a>) and kept the architecture of the feature extraction module and reconstruction module unchanged while only modifying the propagation and alignment module during evaluation (<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"section anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Sec9\">Methods<\/a>). We then examined the six models on five different data types: linear SIM data of MTs, lysosomes and Mito, three of the most common biological structures in live-cell experiments, nonlinear SIM data of F-actin, which is of the highest structural complexity and upscaling factor in BioTISR, and simulated data of tubular structure with infallible GT references (Supplementary Note <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#MOESM1\">2<\/a>). As is shown in Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1f<\/a>, Extended Data Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig8\">3<\/a> and Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#MOESM1\">2<\/a>, all models denoised and sharpened the input noisy WF image evidently, among which the model constructed with a recurrent scheme and DC alignment resolved the finest details compared to the GT-SIM image (indicated by white arrows in Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1f<\/a>). Furthermore, we calculated time-lapse correlation matrices (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1g<\/a>) and image fidelity metrics (Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Fig1\">1h\u2013j<\/a>) (that is, peak SNR (PSNR) and structural similarity (SSIM)) for the output SR images to quantitatively evaluate the temporal consistency and reconstruction fidelity, respectively. According to the evaluation, we found that recurrent network-based propagation (RNP) outperformed sliding window-based propagation (SWP) in both temporal consistency and image fidelity with fewer trainable parameters (<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"section anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#Sec9\">Methods<\/a>) and propagation mechanisms had little effect on the temporal consistency of the reconstructed SR time-lapse data, while the DC-based alignment generally surpassed the other two mechanisms with a similar number of parameters for all types of datasets (Supplementary Fig. <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"supplementary material anchor\" href=\"https:\/\/www.nature.com\/articles\/s41587-025-02553-8#MOESM1\">3<\/a>).<\/p>\n","protected":false},"excerpt":{"rendered":"<p>TISR or video SR (VSR) neural network models are designed to leverage temporal neighbor frames to assist the SR of the current frame and are, therefore, expected to achieve better performance than SISR models19 (Supplementary Note 1). Although TISR models have been widely explored in natural image SR to improve video definition, whether such models [\u2026]<\/p>\n","protected":false},"author":661,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[11,6],"tags":[],"class_list":["post-204977","post","type-post","status-publish","format-standard","hentry","category-biotech-medical","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/204977","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/661"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=204977"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/204977\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=204977"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=204977"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=204977"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}