The goal of this work is to develop machine learning tools that can infer information from fluorescence localisation imaging with photobleaching (FLImP) images collected at Central Laser Facilities (CLF). FLImP allows to map out the spacing of molecules in complexes with a resolution better than 5 nm. FLImP has been used to characterise the molecular architecture of complexes of epidermal growth factor receptor (EGFR), the target of a number of drugs in clinical use. Signals from EGFR are responsible for the control of cell growth and EGFR mutations are implicated in many cancers. At present, successful FLImP imaging is a user-intensive process, requiring manual intervention for the selection of regions of interest (ROI) for image segmentation, autofocusing of images, and track selection. More efficient tools need to be developed to automate FLImP and achieve translation to clinical use.
Efficient autofocusing
At present, the optimal focus is determined by using a deconvolution technique on top of the Oxford Nanoimager autofocusing. In this method, the defocused image is treated as a convolution of the in-focus image and the defocusing point spread function (Gaussian Kernel approximating PSF determined from experiments). We plan to develop a machine learning-based offline autofocusing method that enables the prediction of the focusing distance from a couple of defocused images without any prior knowledge of the defocus distance, its direction, or the PSF.
We have developed three convolutional neural network (CNN) based models, one deep CNN, and two pre-trained models with MobileNetV3 and InceptionV3, respectively, to predict the focusing distance of the microscope images. The images obtained from CLF contain both fiducial markers and cell images. The deep-CNN model outperformed the pre-trained models and we found a 91% correlation between the predicted focusing distance and the true focusing distance. The Bland-Altman analysis between the prediction and ground truth showed 95% limits of agreement between +0.66 $\mu$m and -0.65 $\mu$m which is good but not satisfactory enough. We are exploring a deep Q-network-based RL model to achieve autofocusing with the desired accuracy.
Automatic ROI detection
At present, a classical image segmentation approach is used to determine ROI. First, ROIs are drawn on the cells of interest (Hoechst channel) and the EGFR receptors (FLImP labelling channel) by applying the Otsu threshold and triangle threshold, respectively. Then the ROI is selected for recording a FLImP video if both fractions satisfy pre-determined thresholds.
We are exploring various machine learning models to automatize the ROI detection process. A two-step model has been developed. First, this model classifies data frames with and without cells and then segments ROIs from the frames with cells. We compared several architectures made of different backbones and additional neural network modules for segmentation. Three segmentation models with 10 backbones have been designed to classify and segment ROIs from the image frames with cells. JPU-FCN (Joint Pyramid Upsampling-Fully Convoluted Network) and DeeplabV3 are the best-performing among the segmentation models, and MobileNetV3 is the best among the backbones. These models utilize information from two channels and demonstrate promising performances. Now we are validating our model against different sets of manual ROIs.
Image below: ROI or mask prediction using our model and its comparison with the manual mask.
Automatic track selection
Each FLImP series typically returned between 1,000 and 10,000 track objects of which only a small fraction was suitable for FLImP analysis. At present, the identification of FLImP suitable tracks is a laborious process, requiring trained operators to manually go through track lists from each FLImP series to identify tracks that may be suitable for downstream FLImP fitting processes. The goal of this work is to identify hidden features between good and bad tracks and classify them automatically.
At first, an autoencoder is used for unsupervised feature extraction. The track data are transformed into Fourier space to obtain a more continuous representation of the data. After that, Kalman Filter-based method is used for denoising and level-classification in each of the individual tracks to reconstruct and polish conventionally unused tracks, which can boost the overall processing pipeline.
Image below: Classifcation result from autoencoder
Image below: Track feature extraction
Image below: Kalman filter based method for denoising and level-classification
Note: This project is in collaboration with scientists from CLF.