TY - JOUR
T1 - Three-dimensional virtual refocusing of fluorescence microscopy images using deep learning
AU - Wu, Yichen
AU - Rivenson, Yair
AU - Wang, Hongda
AU - Luo, Yilin
AU - Ben-David, Eyal
AU - Bentolila, Laurent A.
AU - Pritz, Christian
AU - Ozcan, Aydogan
N1 - Publisher Copyright:
© 2019, The Author(s), under exclusive licence to Springer Nature America, Inc.
PY - 2019/12/1
Y1 - 2019/12/1
N2 - We demonstrate that a deep neural network can be trained to virtually refocus a two-dimensional fluorescence image onto user-defined three-dimensional (3D) surfaces within the sample. Using this method, termed Deep-Z, we imaged the neuronal activity of a Caenorhabditis elegans worm in 3D using a time sequence of fluorescence images acquired at a single focal plane, digitally increasing the depth-of-field by 20-fold without any axial scanning, additional hardware or a trade-off of imaging resolution and speed. Furthermore, we demonstrate that this approach can correct for sample drift, tilt and other aberrations, all digitally performed after the acquisition of a single fluorescence image. This framework also cross-connects different imaging modalities to each other, enabling 3D refocusing of a single wide-field fluorescence image to match confocal microscopy images acquired at different sample planes. Deep-Z has the potential to improve volumetric imaging speed while reducing challenges relating to sample drift, aberration and defocusing that are associated with standard 3D fluorescence microscopy.
AB - We demonstrate that a deep neural network can be trained to virtually refocus a two-dimensional fluorescence image onto user-defined three-dimensional (3D) surfaces within the sample. Using this method, termed Deep-Z, we imaged the neuronal activity of a Caenorhabditis elegans worm in 3D using a time sequence of fluorescence images acquired at a single focal plane, digitally increasing the depth-of-field by 20-fold without any axial scanning, additional hardware or a trade-off of imaging resolution and speed. Furthermore, we demonstrate that this approach can correct for sample drift, tilt and other aberrations, all digitally performed after the acquisition of a single fluorescence image. This framework also cross-connects different imaging modalities to each other, enabling 3D refocusing of a single wide-field fluorescence image to match confocal microscopy images acquired at different sample planes. Deep-Z has the potential to improve volumetric imaging speed while reducing challenges relating to sample drift, aberration and defocusing that are associated with standard 3D fluorescence microscopy.
UR - http://www.scopus.com/inward/record.url?scp=85074766196&partnerID=8YFLogxK
U2 - 10.1038/s41592-019-0622-5
DO - 10.1038/s41592-019-0622-5
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 31686039
AN - SCOPUS:85074766196
SN - 1548-7091
VL - 16
SP - 1323
EP - 1331
JO - Nature Methods
JF - Nature Methods
IS - 12
ER -