Figures
Abstract
Mitosis is the process by which eukaryotic cells divide to produce two similar daughter cells with identical genetic material. Research into the process of mitosis is therefore of critical importance both for the basic understanding of cell biology and for the clinical approach to manifold pathologies resulting from its malfunctioning, including cancer. In this paper, we propose an approach to study mitotic progression automatically using deep learning. We used neural networks to predict different mitosis stages. We extracted video sequences of cells undergoing division and trained a Recurrent Neural Network (RNN) to extract image features. The use of RNN enabled better extraction of features. The RNN-based approach gave better performance compared to classifier based feature extraction methods which do not use time information. Evaluation of precision, recall, and F-score indicates the superiority of the proposed model compared to the baseline. To study the loss in performance due to confusion between adjacent classes, we plotted the confusion matrix as well. In addition, we visualized the feature space to understand why RNNs are better at classifying the mitosis stages than other classifier models, which indicated the formation of strong clusters for the different classes, clearly confirming the advantage of the proposed RNN-based approach.
Citation: Jose A, Roy R, Moreno-Andrés D, Stegmaier J (2024) Automatic detection of cell-cycle stages using recurrent neural networks. PLoS ONE 19(3): e0297356. https://doi.org/10.1371/journal.pone.0297356
Editor: Xiao Luo, University of California Los Angeles, UNITED STATES
Received: August 25, 2023; Accepted: January 2, 2024; Published: March 11, 2024
Copyright: © 2024 Jose et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data underlying the results presented in this study are in location: https://osf.io/8qdgm/files/osfstorage For information about data please refer to: (https://osf.io/b6gy5/wiki/home/).
Funding: This work is supported by internal funds.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Chromatin segregation errors during mitosis are a source of chromosomal instability and a hallmark of diverse pathologies [1, 2], including cancer [3, 4]. Therefore studying in detail the highly dynamic process of mitosis, as well as the fate of chromosomes during cell division is of great importance for both basic and clinical research. The study of mitotic cytology and its pathological phenotypes date back more than a century with the pioneering work of Walther Flemming [5]. However, the quantitative study of morphological parameters in fixed samples [6–8] or time-lapse images of living cells have recently emerged as a powerful tool to elucidate underlying molecular mechanisms [9–11]. In the pursuit of less biased and time-consuming quantitative approaches, an immense effort is being made to generalise algorithms capable of automatically segmenting, tracking, extracting and quantifying different feature aspects of the mitotic events in time-lapse microscopy image sequences. A sequence indicates a series of frames. Such tools for unbiased and comprehensive cytological analysis reflecting natural and pathological changes in different phases of the cell-cycle are urgently needed in both basic and translational research [12]. The technological advancement in the area of machine learning with deep neural networks can potentially support biological and clinical researchers. For example, these automated methods could give a faster and more precise analysis of cell behaviors under different drug treatments [13, 14].
Related work
The state-of-the-art methods identifying cell-cycle stages in time-lapse microscopy records extract image features and then expert biologists train machine learning algorithms to generate cytological classifiers. For example, CellCognition [10] is a technique that is applied for annotating time-resolved mitosis stages from live cell image sequences. The high similarity between images in some stages of mitosis and smooth transition will introduce high classification noise at the state transitions. CellCognition demonstrates that the incorporation of time information into the annotation can compress the classification noise and reduce the confusion between images of similar morphology. In this approach, an object detection method is implemented initially to identify the location of each cell. They used local adaptive thresholding [15] with watershed split-and-merge error correction [16] to detect individual cells with high accuracy. Then for each object, they calculated features describing texture and shape. From these features, they combined a supervised machine learning technique for classification with a Hidden Markov Model (HMM) to reduce misclassification by incorporating time information. Even though this method gives good performance, some of the mitosis stages are wrongly classified. This is due to the high degree of similarity between some classes as well as the fact that some classes have fewer training samples than others. They assumed that each state of the cell at a given time point depends on the previous state and considered an HMM [17] for the error correction. All the parameters of the HMM like prior probabilities, transition probabilities, and prediction probabilities are derived automatically from the output of a Support Vector Machine classifier. Then they derived the overall maximum likelihood path for the sequence by the Viterbi algorithm [18]. HMM increased the overall accuracy of the model by eliminating the misclassifications at stage transitions.
In 2012, [19] proposed an unsupervised method for identifying the stages of mitosis from the images captured using time-lapse microscopy. This paper introduces a clustering algorithm based on a temporally-constrained combinatorial clustering (TC3) method as a module in the CellCognition [10] software. This approach uses the features extracted from the time-lapse microscopy images of human tissue culture cells (HeLa ‘Kyoto’ cells). This method classifies the cells into interphase and five stages of mitosis, which are prophase, prometaphase, metaphase, anaphase, and telophase. The CellCognition software calculates the synchronized time series of cell features based on the shape and texture of the tracked cells over time. The authors then convert these features to a lower data dimension using principal component analysis [20] (PCA). They use this converted data as the starting point of the TC3 algorithm. For each cell trajectory, the TC3 algorithm will cluster temporally liked features to a user-defined number of classes. The authors first used a binary clustering algorithm to divide the PCA features into 3 clusters based on mitotic subgraph properties. Further, they used the TC3 algorithm to cluster into subclasses. The authors also state that the performance of the model can be increased by using TC3 results to initialize a Gaussian Mixture Model (GMM). Also, further extended the performance by extending this GMM model results to HMM. This model predicted the cell behaviors closer to the user annotations. For each subcluster, the TC3 algorithm will do an exhaustive search with all possible combinations of labels. Within each set, it calculates the distance measures of features assigned to the same class of labels and selects the set of labels with the least distance measure as the final output. TC3 algorithm on the set of features used for this dataset gives results similar to the user annotations.
LiveCellMiner [11] is an open-source tool developed to analyse live-cell time-lapse records obtained in different microscopy platforms in a quantitative and unbiased manner. This software tool can track single cell fates and extracts, analyses, and visualises biological relevant image features from 2D+t microscopy images. The study of human cells passing through mitosis under different experimental conditions has settled the proof of principal application of this tool. Some of the functionalities of this software include fully-automatic segmentation and tracking of the cells. This tool can be also used to extract the quantitative features of the cells being tracked. This tool is available as an extension package in the MATLAB toolbox SciXMiner [21]. Object detection functionalities of this software are utilized to locate different cell nuclei in an image containing many cells and then crop a square region around each cell. One of the datasets used in this paper is extracted by this tool. Segmentation labels for each cell are extracted by using a modified version of Otsu’s method [22]. The cell trajectory synchronization application of LiveCellMiner is used to identify the duration of different mitotic stages. Using this tool mitosis as well as wrongly detected trajectories can be automatically identified. The default settings for mitotic analysis on LiveCellMiner divides cell-cycle into three classes: interphase, early mitosis (prophase to the end of metaphase), and late mitosis (anaphase, telophase and until G1 when the nucleus recovers to interphase). LiveCellMiner uses three different methods for synchronizing cell trajectories. The first method is a TC3 clustering method [19]. In this method, classical image features like area, circularity, and intensity of each cell are calculated and clusters with minimum within-class variance are extracted for detecting the interphase to prophase transition. The metaphase-to-anaphase transition comes from the tracking and uses a user-defined distance of sister chromatin masses to identify anaphase onset. The second approach is based on the first approach and additionally uses trainable LSTM [23] networks which evaluate each trajectory as a whole, identifying erroneous trajectories. The third method uses Convolutional Neural Networks (CNNs) features extracted by using GoogleNet and uses LSTMs to predict the state sequence for all time points. These predicted time points are post-processed with an HMM model which allows only valid state transitions. Then the most likely sequence of stages is identified using the Viterbi algorithm [18].
A Recurrent Neural Network (RNN) [24] is a type of deep learning technique used with temporal sequence data. Artificial neural networks such as CNNs are meant for single data points which are independent of each other. However, in sequential data or time series data, one data point is dependent on the previous data points. Since RNNs are temporally connected, they can store the information from prior inputs to influence the current input and output. Thus, the output of an RNN is dependent on not only the current time point but also on the previous time points. This type of network is commonly used in speech recognition and natural language processing. RNNs can also be used with convolutional layers to extend their application to video data. When the sequence data is very large, the gradient information during the training of an RNN is not able to propagate back to the earlier time points. This problem is called vanishing gradients. Cho et al. [25] proposed a type of RNN called Gated Recurrent Unit (GRU) to solve the vanishing gradient problem. GRU solves this problem by using an update gate and a reset gate, which decides what information has to be passed to the output.
In 2016, Ondruska et al. [26] proposed an RNN-based approach as ‘Deep Tracking’, for the end-to-end detection of objects from sensor data. In this paper, the authors used RNN networks to extract features from the input data. For the end-to-end object tracking directly from raw sensor data for robots, in 2016, Ondruska et al. [26] proposed the deep tracking approach. The idea was to use raw sensor data without any feature engineering as the input and produce an output that also included the detection of the occluded objects. This paper also followed an unsupervised training approach to achieve this result. The authors stated that this is the first approach that uses unsupervised end-to-end tracking of objects using sensor data. Classical approaches at the time of this work solved object detection in separate stages, with an object detection stage and a tracking stage. The proposed deep tracking approach is trained end-to-end, thereby overcoming the hand-engineering required in such a separately executed model. This is achieved by exploiting the sequential model in the form of RNNs to learn complex dynamics from raw data to object tracks. In this model, the hidden states of the RNNs can capture certain appearances and motion patterns of objects. Inspired by the Bayesian filtering approach and considering a generative model, the authors stated that there exists a Markov process, which completely captures the state of the world. From the hidden state representation, the location of each object at each time frame is trained to predict using binary cross-entropy loss. In a later work, Ondruska et al. [27] proposed that each of the object classes can be learned from this hidden state distribution by using RNNs. The network can perform a semantic classification from the rich information learned in the hidden states. They proposed that this network can be trained end-to-end with a small amount of labeled data. This classification network is introduced once the deep tracking network is learned. This method outperformed many semantic classification methods available at that time. To be able to track moving objects throughout the time sequence, the network must remember the location and other properties of each object. To achieve this tracking, the authors used GRUs as the processing steps at each layer of the RNN. Convolutional GRUs are utilized to maintain the resolution. From the learned hidden layer, two convolutional networks are employed to predict the semantic segmentation as well as the class of the object. The network is trained with softmax loss for the classification and a binary cross-entropy loss for the segmentation.
This paper
In this paper, we propose an approach that identifies different stages of a cell during mitosis using deep learning techniques from tracked cell images. The tracked cell images follow a single cell in different time frames and avoid background clutter from the surrounding cells. The aim is to find the phases of mitosis of the cell in different time frames. In other words, the aim is to find the temporal segmentation of a video sequence of cell data. This means that the class labels are assigned to each frame of the video sequence to classify the mitotic phases. Inspired by the model proposed by [27], we modify the network architecture and incorporate time-related propagation of features for better classification of mitosis stages. The main contributions of this paper are summarized here:
- We propose a novel network architecture, Time Encoded ResNet18 Model, which uses the GRUs to capture the inherent time dependency between different frames during mitosis, resulting in better classification.
- We compare the performance of the proposed RNN-based approaches to the original deep learning-based classification networks such as ResNet18 [28].
- Experiments were conducted on two different datasets: a) LiveCellMiner dataset [11], which contains images in three different phases of mitosis. b) Zhong et al.’s dataset which contains images in six different phases of mitosis.
- By visualizing the feature space using PCA [20], the effect of the incorporation of time information is studied.
- We also plotted the confusion matrices to identify the classes that cause confusion and identified potential areas for further improvement in classification accuracy.
- For quantitative evaluation, we have measured the precision, recall, and F-score.
- The reconstruction of the center cell using tracking network is visualized for the images in different mitotic phases.
Proposed approach
Given a cell mitosis video sequence, it has become of utmost importance in the context of cell-biology research to find the temporal segmentation of cell-cycle stages. While there are several successful methods to track [29, 30] the cells from a video sequence, identification of time-dependent cell-cycle stages or cytological phenotypes using deep learning methods is still a challenging task. Most existing approaches [10, 11, 19] for cell stage classification start by extracting features separately for each of the frames and then try to identify the stage sequence by a similarity-based grouping of the feature vectors. However, the temporal dependency of features that were extracted from successive frames is not considered here. To incorporate this time dependency, using normal CNNs would not be sufficient. In this context, we propose an RNN-based approach to address the problem of time-related features. Since in RNN, the networks are connected in time, it can effectively transfer information across temporal data.
Base model
Each cell follows a set of unique steps during cell division. Thus, the different phases of cell mitosis always occur in the same order and the current state can be considered as being dependent on its previous states. Therefore, a cell in any frame of a cell mitosis video sequence is dependent on the same cell from earlier frames. To this end, inspired by the networks proposed by [26, 27], we propose our first architecture by considering time-related features for the classification of the cell mitosis stages. The network design of the model is given in Fig 1. Our proposed model consists of four main layers:
- A backbone network for feature extraction.
- A time encoding network to incorporate time information into the features.
- A tracking network to track the cell.
- A classification network for the mitosis stage prediction.
Illustration of the architecture of our proposed base model applied to the nth frame of a video sequence [31].
The initial results and experiments using this base model are summarized in our previous work [31]. The experiments show that adding the time information into the features, helps to better identify cell-cycle stages. Each of these modules is explained in detail in [31].
We further modified this baseline architecture, with a view that deeper networks can create deep feature representations with complex functions and are more efficient. The modified model, time encoded ResNet18, which we elaborate in this paper also has the aforementioned modules. The details of this model are explained in next subsection.
Time encoded ResNet18
This model introduces deeper layers between each GRU layer by combining the architecture of ResNet18 [28] with RNN layers. Combining these two modules introduces a deeper architecture with convolutional layers between each RNN layer. This in turn enhances the feature extraction capability of the model. Combining these two modules helps to propagate the features extracted at different layers and increases the classification performance. RNN combines the features of the current frame with the information from the previous frames in the video. In a process like a cell mitosis, this encoding of time information helps in predicting the class to which the cell belongs, because one stage is dependent on the previous stages and always occurs in the same order. After the time encoded features are extracted, tracking and classification networks are used for reconstructing the center-cell and predicting the class of the cell respectively.
Time encoded backbone network.
The main difference from the base model [31], is that there is no separate backbone and time encoding network. They are combined together into a single module, as shown in Fig 2. The time encoded backbone network extracts the features with time information for each frame in the video sequence. This is achieved by combining a state-of-the-art model like ResNet18 with the properties of an RNN. ResNet18 contains eighteen deep layers with eight residual blocks. The proposed network uses an architecture of ResNet18 with convolutional GRUs between the residual blocks to transfer time information between frames. Similar to the base model, this model has three convolutional GRUs. The GRU layers in this model have been positioned such that the dimensions of each GRU layer is different. This helps to propagate information at multiple scales. In the ResNet18 architecture, a max-pooling layer reduces the dimensionality after every two residual blocks. Thus, each convolutional GRU is placed between two residual blocks of the ResNet18 architecture. This addition of convolutional GRU layers does not change the dimensionality of the features of the ResNet18, but help in propagating time information by combining features from the previous time frames to the present frame. The channel sizes of the convolution GRU layers are 128, 256, and 512. The original grayscale image is rescaled and converted to an RGB image for matching the input requirements of the ResNet18 architecture. Starting with the initial image (RGB image), the time encoded backbone generates a lower resolution activation map . Typical values used for channels and image resolution are C = 512, and H,W = respectively. The ResNet18 layers are loaded with pre-trained weights of network trained on ImageNet dataset and the complete network is fine-tuned with the cell sequence dataset.
Illustration of the architecture of our proposed time encoded ResNet18 model applied to the nth frame of a video sequence where GRU blocks are connected between residual blocks of ResNet18.
Tracking network.
Since the time encoded backbone network has a deeper architecture compared to that of the base model, the output features are at a lower spatial dimension. So compared to the base model, which uses three transpose convolutional layers, this tracking network uses four transpose convolutional layers to reconstruct the center-cell to the original image dimension from the output of the time encoded backbone network. 3×3 kernels are used in all transpose convolutional layers. The typical values of the channel dimensions used in each transpose convolutional layer are 256, 128, 64, and 3. The final layer has an output with three channels matching the RGB image size with each channel representing the grayscale content. Thus a resolution equal to the RGB input image is reconstructed and the output dimensions are .
Classification network.
A shallow classification network is enough to predict the class from the output features of the time-encoding backbone network. A convolutional layer, a max-pool layer, and two fully-connected layers make up this network. The convolutional layer has 1024 filters and uses 3×3 kernels. The feature dimensions are then reduced by a 2×2 max-pooling layer. The features are flattened after the max-pooling layer. Then a series of two fully-connected networks is added that predicts the class to which each frame belongs. The classification network’s last layer will provide an output whose size is equal to the number of classes.
ResNet18 classifier
The two models that were discussed previously use RNN to propagate information between subsequent frames. To compare the performance of the models without using RNN, we propose to use a state-of-the-art deep learning classification model. This model consider each frame in the video as independent of other frames in the same sequence. This approach uses transfer learning on a ResNet18 [28] model. A pre-trained model architecture with the last layer replaced with the number of classes is used. Then the complete network is trained with image datasets of the mitosis sequence. ResNet18 has eighteen deep layers with eight residual block connections. In the final layer, the number of outputs is updated with the number of classes in our dataset. Thus the network predicts the stage of mitosis for each input image. The complete network is then trained with the cell mitosis dataset. In this way, the network can learn more information related to microscopy images, whereas the pre-trained ResNet18 model is trained with non-microscopy images. The network consists of seventeen convolutional layers and one fully-connected layer. All the convolutional layers have a kernel size of 3×3. Eight residual blocks are implemented with two convolutional layers. A max-pooling operation reduces the dimension after every two residual blocks. Finally, the fully-connected layer predicts the class to which each image belongs.
Training details
This section explains training with the supervised approaches proposed in this paper. The models are trained end-to-end. The proposed models that use GRUs to deal with time-based features have a tracking network and a classification network. The tracking network reconstructs only the center-cell out of many cells in the frame so that the feature space will contain information about this cell. The classification network predicts the stage of mitosis of the cell. The fully-connected network in the final layer of the classification network has an output dimension equal to the number of classes to predict. The losses for these two networks are combined simultaneously by using a weighting factor. The deep learning classification models, which do not have a GRU, have only the classification network that predicts the stage of the mitosis of a cell, and the loss is optimized during training. The following section explains how these models are trained in different scenarios.
Training base model and time encoded ResNet18
Since time encoding networks are used to help the flow of information across time frames, each image in a sequence is dependent on previous images. Therefore in these models, the sequence of images belonging to the same cell is given as the input.
Tracking network loss.
The input images used for training base model are grayscale images with a resolution of 96×96. Time encoded ResNet18 uses 3-channel RGB images as input and has a resolution of 224×224. Hence, the original images are converted to 3 channels and resized for training these models. The output of the tracking network in both models has a dimensionality equal to that of the input training images. The tracking network reconstructs only the center-cell from the input image. This is achieved by calculating the loss between the predicted output and a masked image. The masked image is the input image masked with the segmentation of the center-cell. The last layer of the tracking network uses a sigmoid activation layer. Then the loss between the predicted and the expected input is calculated using a binary cross-entropy loss [32] as given below: (1) where y(w, h) and are the input and predicted values of the network at pixel location ((w, h)). W0 and H0 are the input image dimensions. The segmentation mask of new images can also be extracted using a simple intensity threshold operation on the predicted output of the tracking network.
Classification network loss.
In the classification network, the final layer has a number of outputs equal to the number of mitosis stages. This layer has a sigmoid activation function. The network is trained with a one-hot vector of the ground-truth values. This helps the network to train to predict the probability of the image belonging to each class. The training loss function used is binary cross-entropy loss [32] as shown below: (2) where yc is the one-hot embedding of the ground truth values and is the output predicted values of the classification network for the cth class. Nclass is the total number of classes in the dataset. The class with the highest probability is chosen as the predicted class during inference. The total loss is calculated as a weighted sum of the tracking and classification losses. A parameter λwt is used to weigh the losses. (3)
Training ResNet18 classifier
The deep learning architectures which used ResNet18 classifier has only the classification network loss. In these models, each image is considered to be independent of the previous image in the same sequence. The final layer of these networks has a number of outputs equal to the number of classes to predict. A one-hot embedding vector of the ground truth is used to train this network. This network is trained with the cross-entropy loss [32] to predict the probability of the image belonging to each class as shown below: (4) where yc is the one-hot embedding of the ground truth values and is the classification network’s output prediction values for the cth class. Nclass is the total number of classes in the dataset. The class with the highest probability is chosen as the predicted class during inference. The total loss of models with deep learning architectures is the same as the classification loss as shown below: (5)
Datasets
For experimental evaluation, we have used two main datasets. The first dataset is provided by Moreno-Andrés et al. [11] and the second dataset is provided by [19]. Both datasets contain microscopic image sequences of the mitotic process from human HeLa cells expressing H2B-mCherry as fluorescent chromatin marker. These images are acquired using time-lapse microscopy. In the following subsections, these two datasets are explained.
LiveCellMiner dataset
As explained in, the LiveCellMiner [11] tool allows the analysis of mitotic phases in 2D+t microscopy images. These images contain human tissue culture cells as they undergo mitosis and are acquired using widefield and confocal microscopy. These acquired images have many cells in each time frame. The software tracks the position of each cell in the image and then extracts the tracked single cell image. The software then checks and eliminates cells that do not undergo cell division. A region is cropped around each tracked cell from each time frame if the cell undergoes cell division. Each of these cropped images has a resolution of 96 × 96 pixels, and the target cell is in the center of the cropped image. For each cell division sequence, 90 frames are available with this resolution. This dataset is divided into three classes. This software identifies interphase to early prophase transition and metaphase/early anaphase to late anaphase transition as reference patterns for the alignment of interphase or postmitotic frames, and then automatically detects interphase to prophase and metaphase to anaphase transitions. Then it divides the data into interphase, mitosis, and post-mitosis classes. A sequence of images belonging to different classes is illustrated in Fig 3. Experts corrected the predicted labels of the LiveCellMiner tool and released them as ground truth annotations. Along with this tool, four different image datasets are published. All four datasets are acquired on the human HeLa cells expressing H2B-mCherry transfected with indicated siRNA oligonucleotides in eight-well µ-slide chambers. These datasets contain images that were taken three minutes apart. The total number of training sequences in each of these datasets and a detailed description of these datasets are given below. The first dataset is the LSM710 dataset. In this dataset, the cells are imaged using an LSM710 confocal microscope (Zeiss) and ZEN software (Zeiss). 1458 sequences are available for training with the LSM710 dataset. The second dataset is the LSD1 dataset [33]. It was acquired using an LSM5 live confocal microscope (Zeiss) and ZEN software. With this dataset 1042 sequences are available for training. RecQL4 dataset [34] is the third dataset and was acquired using LSM5 live confocal microscope and ZEN software. 1214 sequences with this dataset are available for training. The last dataset is the NikonXLight dataset [35]. This dataset is imaged with a widefield module of a Ti2 Eclipse (Nikon) equipped with a LED light engine SpectraX and GFP/mCherry filter sets and using elements software (Nikon). This dataset contains 1229 sequences for training. Fig 4 illustrates a few images of these four datasets at different stages of cell splitting.
Illustration of the images of a cell that undergoes cell division and also assigned to the three different cell-cycle classes: interphase, mitosis (prophase, prometaphase, metaphase), and post-mitosis (early and late anaphase and telophase) classes [11]. Two sequences are available after the cell-splitting—each follows one daughter cell after mitosis.
Images belonging to four datasets of LiveCellMiner at different cell-cycle stages [11].
Zhong et al.’s dataset
The dataset provided by Zhong et al. [19] also uses time-lapse microscopy images of human tissue culture cells (HeLa cells). This dataset contains cells as they undergo cell division. Each frame in the dataset is labeled as interphase or one of the five mitosis stages. Cell images belonging to these six stages, with their state diagram is shown in Fig 5. The labeling by different biologists can be inconsistent despite the well-defined chromatin morphology. The biologist’s annotations were subjected to a dissimilarity study that showed minor inconsistencies between annotations made by the same person on different days, but significant differences between annotations made by different users. The gold standard for the labels is chosen by a majority vote among these user annotations. The dataset consists of seven image sequences with a total of 326 cell division events uniformly sampled with an interval of 4.6 minutes. Since mitosis stages occur in a shorter time compared to the interphase stage, the distribution of cells in different classes is highly uneven. In this dataset, each image has a resolution of 96 × 96 pixels, and each cell sequence has a length of 40 frames.
Illustration of the different stages of cell-cycle [19] as a state diagram. Cell images belonging to different classes are labeled with different colors.
Evaluation criteria
This section explains the different evaluation techniques, used to compare the models. We compared our results with some of the state-of-the-art methods as well as between our different models. Here we measured classification performance such as classification accuracy, confusion matrix, and some of the measures which are derived from the confusion matrix. For the evaluation of multi-class classification, three possible cases for predictions are available for each class. These are true positives (TP), type I error or false positives (FP), and type II error or false negatives (FN) [36]. TPs of a class are the correctly classified data, whereas FPs and FNs are the incorrectly classified data. The data that belongs to this class but is predicted as one of the other classes is FN. FPs are the data that are predicted into this class that belongs to another class.
Accuracy
Accuracy is defined as the sum of all TPs from all the classes divided by the total number of data points. Here in this paper, we calculate frame-to-frame accuracy from each sequence and average it over the total length of data. The accuracy that is discussed in this paper is the average accuracy generated by evaluating the ground-truth labels and the predictions from the test dataset. A higher value of accuracy means better performance. The value is in the range of 0 to 1.
Confusion matrix
The confusion matrix can visualize the three cases of prediction, that is TPs, FPs, and FNs. The TPs for each class lie on the diagonal of the confusion matrix [36]. The FPs and FNs for a class will be on the corresponding columns and rows respectively, such that the true class is on the y-axis and the predicted class is on the x-axis. In this paper, we present the normalized confusion matrix.
Precision and recall
The precision represents the proportion of correct classified images within a class to the total number of images classified in this class. I“t is the ratio of TP to the sum of TP and FP as shown in Eq 6. A higher value of precision means better performance. Recall again indicates the ratio of TP to the sum of TP and FN. (6) (7)
F-score
The F-score denotes the harmonic mean between precision and recall. The values of the F-score lie in the range of 0 to 1. A higher F-score means that the classification has a better performance. The F-score is defined as in Eq 8. (8)
Experiments
In line with our previous work [31] we conduct the experiments for each of the five datasets mentioned in Section. We compared the performance of the proposed RNN-based methods with ResNet18 classifier. The classification models treat each image independently. Hence, the number of images in the batch equals the batch size. Models with GRU layers always consider each image in a sequence dependent on the previous images from the same sequence. Each sequence is used together as input during training to produce this dependency. So, in this case, the batch size equals the number of sequences. The experiments are carried out with these different models to evaluate their performance. In experiments, we observed that a batch size of two or four sequences performs better than a batch size of one. In this paper, the typical batch chosen has 4 sequences for the models with GRU layers. This is because, due to the graphic processing unit’s memory limitations, some deeper models cannot employ batch sizes greater than 4. For the models using ResNet18 classifier, the batch size used is 64 images. For the training of the base model where a pre-trained model is not available, a learning rate of 0.01 proved to provide better training than lower or higher values. The typical learning rate used for models with a pre-trained model is 0.001. Inspired by [27], the learning rate is scheduled to reduce to 10 percent of the current value after 2500 training iterations. The weighting parameter value controls the weighting between multiple losses. We observed that the weighting parameter value of 0.01 has better accuracy with the base model and a value of 0.1 with the time encoded ResNet18 model by using the basic grid search [37] approach. Besides the aforementioned hyperparameters, data augmentation is introduced in all experiments to increase the robustness and avoid overfitting. Rotations by multiples of 90 degrees and flips in vertical and horizontal directions are the data augmentations used. We also split the datasets into separate training and testing datasets. The typical value of the train-test ratio used is 0.85, such that 85% of the total data is used to train the models, and the remaining 15% is used as test dataset. 10% out of the test images were used as the validation set. The experiments were done for 20 repetitions and the standard deviation is measured. Table 1 shows the typical values of the hyperparameters used in the experiments, which were determined by grid search-based parameter tuning. The cells will grow back to the interphase stage after they undergo mitosis. The expert annotation of Zhong et al.’s dataset considers this, while the LiveCellMiner dataset does not consider this aspect. In LiveCellMiner datasets, the cells moving back to the interphase stage are annotated as the part of the post-mitosis class. Due to the resemblance between these cells in the post-mitosis class and those in the interphase class, this will pose issues during training. Studies from the LiveCellMiner [11] showed that the recovery back to the interphase stage happens around 15 to 20 frames after mitosis. To overcome the above issue, during training, we labeled all the frames after around 15 to 20 frames from the post-mitosis stage as the interphase class. During inference, the predicted interphase recovered class after the mitosis stage is assigned to the post-mitosis stage and evaluated. This also solves the problem of similarity between images for deep learning classification networks like ResNet18. Thus, the recovery back to interphase does not affect the prediction accuracies of these models.
The hyperparameter values chosen for experiments after initial grid-search based tuning.
Experimental results
This section presents the results of the proposed approaches on different datasets. The code is available in the git link, https://github.com/Rijo756/cell-cycle-stages-identification. Accuracy, confusion matrix, precision, recall, and F-score are used to quantify the classification performance. The results are estimated on two models using time encoding with GRU layers and also on the ResNet18 classifier. The models are trained for 12 epochs with the LiveCellMiner dataset and 40 epochs with Zhong et al.’s dataset (both around 10, 000 batch iterations). We also plotted the label matrix which helps in comparing the class predictions of proposed models to the user annotations. Another experiment to study how the learned feature space looks for different models is also plotted for different datasets by taking the first 3 principle components, to visualize how the feature space looks in a lower dimensional space.
Results on LiveCellMiner dataset
In this section, the results are presented in four parts. Each part belongs to results from one of the four datasets available from LiveCellMiner.
LSM710 dataset.
The results of the proposed models trained with the LSM710 dataset from LiveCellMiner are presented in this subsection. Fig 6 visualizes the label matrix of the ground truth annotation and the labels generated by the proposed models on 50 test sequences. The y-axis denotes these sequences, and the x-axis is the length of each sequence. Each of the four proposed models produced excellent results on this dataset. The performance of each of the predicted labels can be computed using the accuracy of the predictions.
Label matrices of user annotation and the predictions by proposed models for 50 sequences selected from the test data of the LSM710 dataset. The y-axis represents different cell trajectories, and the x-axis represents the length of each trajectory. Green, magenta, and red represent the interphase, mitosis, and post-mitosis classes respectively.
Table 2 demonstrates the frame-to-frame accuracy of each of the proposed models. With the LSM710 dataset, the classification using the ResNet18 model got the least accuracy, and the time encoded ResNet18 model has the maximum performance in terms of accuracy. Both the proposed models have higher accuracy than the ResNet18 architecture. This indicates that introducing features from the previous time step using GRU layers seems to help increase the performance.
The frame-to-frame accuracy value of various proposed models using the LSM710 dataset. Boldface indicates the best performance.
The LSM710 datasets are labeled as interphase, mitosis, and post-mitosis stages. The confusion matrix represents the correct predictions as well as the wrong predictions. Fig 7 visualizes the normalized confusion matrices for the proposed models. It is evident from the confusion matrices that all the models were able to classify the images into correct stages in most of the cases as the values of the diagonal elements are higher. The precision, recall, and F-score of the three classes of various models are presented in Table 3. The F-score, precision, and recall are highest for the interphase and post-mitosis stages for the time encoded ResNet18 model. For the mitosis stage, recall and F-score for LiveCellMiner have slightly better performance but the time encoded ResNet18 also has comparable performance here as well.
Normalized confusion matrices of the prediction with the LSM710 test dataset.
Average precision, recall, and F-score for each stage of mitosis for the LSM710 dataset. Boldface indicates the best performance.
Fig 8 visualizes the reconstructed output image from the tracking network. The base model and the time encoded ResNet18 model has a tracking network to reconstruct the center-cell from the input image. The ResNet18 model is a classification network and only outputs the stage each input image belongs to. Both models using a tracking network were able to reconstruct the center-cell. The reconstructed cell from the base model seems to be better compared to the time encoded ResNet18 model. This could be because the embedding space of the base model is at a higher spatial dimension compared to the other model. This tracking network can give the segmentation of the center-cell, with a simple intensity thresholding operation as shown in Fig 8. The embedding space or the feature space gives characteristics related to some properties of the input data. From this feature space in a lower dimension, it can be understood approximately how the classes are divided. Fig 9 illustrates the first three principal components of the feature space of our proposed models. It can be seen that time encoded ResNet18 modules have a proper separation between the features belonging to each class. The higher performance scores of this model could arise because the features belonging to each class are well separated compared to other models. The features of the baseline model also look well separated. During the postmitotic stage, the daughter cells will grow back to the interphase stage. Since the ResNet18 operates on individual images instead of sequences, less separation between interphase and post-mitosis features is visible. This does not occur in models with GRU layers because the feature space includes the time information which is the main advantage of the proposed model. Hence as seen in Fig 9, time encoded ResNet18 model got better separation of feature embeddings of the classes.
Illustration of the output of the tracking network of the proposed models on the LSM710 dataset. The tracking network reconstructs the center-cell given an input image. Additionally shown the ground-truth mask and the segmentation using intensity thresholding of each output.
Illustration of the first three principal components of the embeddings of the proposed models. a) Base Model b) Time Encoded ResNet18 c) ResNet18.
LSD1 dataset.
The results of the proposed models trained with the LSD1 dataset from LiveCellMiner are presented in this subsection. Fig 10 visualizes the label matrix of the ground truth annotation and the labels generated by the proposed models on 50 test sequences from the LSD1 dataset. Each sequence has a length of 90 frames.
a) Label matrices of user annotation and the predictions by proposed models for 50 sequences selected from the test data of the LSD1 dataset. The y-axis represents different cell trajectories, and the x-axis represents the length of each trajectory. Green, magenta, and red represent the interphase, mitosis, and post-mitosis classes respectively.
The predicted label matrix of the proposed models compared to the user annotation labels looks similar. This means that each of the four proposed models produced excellent results on this dataset. Further detailed evaluation of the performance is achieved by computing the frame-to-frame accuracy of the predictions on the test dataset. In this dataset, the classification using the ResNet18 model got the least accuracy, and the time encoded ResNet18 model has the maximum performance in terms of accuracy. Thus it seems that introducing features from the previous time step using GRU layers helps in increasing the performance. The precision, recall, and F-score of the three classes of various models are presented in Table 4. Table 5 demonstrates the frame-to-frame accuracy of the proposed models. Except for the precision values in the mitosis and post-mitosis classes, the proposed time encoded ResNet18 model has the highest scores in all other cases. Thus it seems that introducing GRU layers between residual blocks ResNet18 architecture helps in increasing the performance of the model.
Average precision, recall, and F-score for each stage of mitosis for the LSD1 dataset. Boldface indicates the best performance.
The frame-to-frame accuracy value of various proposed models using the LSD1 dataset. Boldface indicates the best performance.
The confusion matrix represents the correct predictions as well as the wrong predictions. Fig 11 visualizes the normalized confusion matrices for the proposed models. The diagonal values of the confusion matrix represent the true positive predictions for each class. With the LSD1 dataset, the diagonal elements of the confusion matrix have higher values than the off-diagonal values. This implies that the proposed models were able to classify the images into the correct classes. The PCA plot in Fig 12, indicates better clustering of feature embedding, for the time encoded ResNet18 model compared to the classifier model.
Normalized confusion matrices of the prediction with the LSD1 test dataset.
a) Base Model b) Time Encoded ResNet18 c) ResNet18.
RecQL4 dataset.
The results of the proposed models trained with the RecQL4 dataset from LiveCellMiner are demonstrated in this subsection. Fig 13 visualizes the label matrix of the ground-truth annotation and the labels generated by the proposed models on test sequences. The label matrix generated by the ResNet18 model contains some misclassifications in the post-mitosis class and also in the mitosis class for some sequences which is eliminated by the proposed RNN-based model. The same behaviour is observed for the other datasets as well which is shown in the respective subsections. The predicted label matrix of the proposed models compared to the user annotation show a very good correspondence. This indicates that in this dataset, both the proposed models produced good results. By calculating the frame-to-frame accuracy of the predictions on the test dataset, a further assessment of the performance is made possible.
Label matrices of user annotation and the predictions by proposed models for 50 sequences selected from the test data of the NikonXLight dataset. The y-axis represents different cell trajectories, and the x-axis represents the length of each trajectory. Green, magenta and red represent the interphase, mitosis, and post-mitosis classes respectively.
Table 6 demonstrates the frame-to-frame accuracy of each of the proposed models. With the RecQL4 dataset, the base model, and the time encoded ResNet18 model have similar accuracies with a slightly higher value for the time encoded ResNet18 model. The ResNet18 model has the least frame-to-frame accuracy clearly showing the advantage of the proposed model. The precision, recall, and F-score of the three classes of various models for the RecQL4 dataset are presented in Table 7. These values help in the analysis of each model’s performance in predicting each class.
The frame-to-frame accuracy values of various proposed models using the RecQL4 dataset. Boldface indicates the best performance.
Average precision, recall, and F-score for each stage of mitosis when trained using the RecQL4 dataset on various proposed models. Boldface indicates the best performance.
It is evident from Table 7 that the time encoded ResNet18 model gave the highest performance scores for most metrics among all models. For some cases, the performance is slightly better for the LiveCellMiner approach. However, the overall performance scores are highest for the time encoded ResNet18 model for most of the cases. The RecQL4 dataset is also annotated into three stages of cell-cycle. The confusion matrix portrays the correct predictions as well as the incorrect predictions. Fig 14 visualizes the normalized confusion matrices for the proposed models. The diagonal values of the confusion matrix represent the true positive predictions for each of the three classes. With the RecQL4 dataset, the diagonal elements of the confusion matrix have higher values than the off-diagonal values. This implies that the proposed models mostly were able to classify the images into correct classes. The PCA plot (Fig 15) has similar behaviour as the other two datasets, with embeddings corresponding to time encoded ResNet18 model showing time continuity as well as clear clustering in the embedding space.
Normalized confusion matrices of the prediction with the RecQL4 test dataset with proposed models.
Illustration of the first three principal components of the embeddings of the proposed models. a) Base Model b) Time Encoded ResNet18 c) ResNet18.
NikonXLight dataset.
The results of the proposed models trained with the NikonXLight dataset from LiveCellMiner are discussed below. Fig 16 visualizes the label matrix of the ground-truth annotation and the labels generated by the proposed models on test sequences. The predicted label matrix of the proposed models compared to the user annotation labels looks comparable. This indicates that with the NikonXLight dataset, each of the three proposed models produced predictions close to the user annotation. By computing the frame-to-frame accuracy of the predictions on the test dataset, a better assessment of the performance is made possible.
Label matrices of user annotation and the predictions by proposed models for 51 sequences selected from the test data of the Zhong et al.’s dataset. The y-axis represents different cell trajectories, and the x-axis represents the length of each trajectory. Green, yellow, orange, violet, blue, and red represent interphase, prophase, prometaphase, metaphase, anaphase, and telophase classes respectively.
Table 8 demonstrates the frame-to-frame accuracy of the proposed models. For the NikonXLight dataset, the time encoded ResNet18 model has slightly lower frame-to-frame accuracy compared to the LiveCellMiner approach. The ResNet18 model has the least frame-to-frame accuracy. This indicates again that introducing the GRU layers to propagate features with time information between frames in a sequence helps in increasing the frame-to-frame accuracy. The precision, recall, and F-score of the three classes of various models for the NikonXLight dataset are presented in Table 9. These values help in the analysis of each model’s performance in predicting each class. It is evident from Table 9 that the time encoded ResNet18 model gave the highest F-scores except for mitosis class, in which LiveCellMiner approach outperforms. The precision and recall values are also highest for the time encoded ResNet18 model in most of the cases, with an exception of precision value in mitosis class and recall value in post-mitosis class in which LiveCellMiner approach slightly outperforms other models. Fig 17 visualizes the normalized confusion matrices for the proposed models. The diagonal values of the confusion matrix represent the true positive predictions for each of the three classes. With the NikonXLight dataset, the diagonal elements of the confusion matrix have higher values than the off-diagonal values. This implies that the proposed models mostly were able to classify the images into correct classes. The feature space embeddings shown in Fig 18 shows formation of very compact clusters of different classes for time encoded ResNet18 model, clearly confirming the advantage of the proposed model. We have also measured the classification accuracy of the proposed model in four datasets for two train test ratios, 0.85 and 0.5. With lower number of training data the performance only drops slightly indicating the advantage of the proposed model. Thus, even with less number of annotated data, the network will be able to given reasonably good performance. The results are given in Table 10.
The frame-to-frame accuracy value of various proposed models using the NikonXLight dataset. Boldface indicates the best performance.
Average precision, recall, and F-score for each stage of mitosis when trained using the NikonXLight dataset on various proposed models. Boldface indicates the best performance.
Classification accuracy in 4 datasets, for two different train test ratios.
Normalized confusion matrices of the prediction with the NikonXLight test dataset.
Illustration of the first three principal components of the embeddings of the proposed models. a) Base Model b) Time Encoded ResNet18 c) ResNet18.
Results on Zhong et al.’s dataset
This section compares the performance of models with the Zhong et al.’s dataset. As seen in results from the LiveCellMiner dataset, the base model, the time encoded ResNet18 model, and the ResNet18 models are evaluated. Fig 19 visualizes the label matrix of the ground-truth annotation and the labels generated by the proposed models on test sequences. The y-axis in this plot represents the 51 test sequences, and the x-axis is the length of each cell sequence. Each cell sequence in this dataset has 40 frames. The predicted label matrix of the proposed models compared to the user annotation labels looks comparable. This indicates that all the models were able to predict closer to ground-truth annotation.
Label matrices of user annotation and the predictions by proposed models for 51 sequences selected from the test data of the Zhong et al.’s dataset. The y-axis represents different cell trajectories, and the x-axis represents the length of each trajectory. Green, yellow, orange, violet, blue, and red represent interphase, prophase, prometaphase, metaphase, anaphase, and telophase classes respectively.
Table 11 shows the precision, recall, and F-score of the proposed models on each class of the Zhong et al.’s dataset. Zhong et al.’s dataset is labeled into six stages of cell-cycle. In Table 11, the first row shows the best model proposed by Zhong et al. [19]. This state-of-the-art method is not a deep learning approach. It is a combination of feature extraction and clustering algorithms. The results of this method compared with results from our proposed models that use deep learning techniques are shown in the table. Our models have slightly lower or higher performance compared to the method proposed by Zhong et al. Within our approaches, the maximum values for each class are distributed across various models. In general, it can be seen that, our proposed models with the GRU layers got higher scores compared to the ResNet18 model. A more precise evaluation can be done by analyzing the frame-to-frame accuracy of each model. Table 12 shows the frame-to-frame accuracy of the proposed models. From this table, it is evident that the time encoded ResNet18 model has the highest frame-to-frame accuracy. So it is clear that our proposed model with GRU layers is able to propagate features between different time instances and increase the performance of the model. The base model which is a very shallow network also has an accuracy higher than the ResNet18 model. The embedding space or the feature space gives characteristics related to some properties of the input data. The first three principal components from principal component analysis (PCA) [20] are used to visualize the embedding space of the proposed models. This is illustrated in Fig 20. The time encoded ResNet18 has the most separate features for each class. This explains the advantage of the proposed model compared to other approaches. The confusion matrix displays the correct predictions as well as the incorrect predictions. Fig 21 visualizes the normalized confusion matrices for the proposed models. The diagonal values of the confusion matrix represent the true-positive predictions for each of the three classes. For all the proposed models, the diagonal values are higher. This implies that the proposed models are able to predict most of the images into their correct classes. The prophase class has the lowest true-positive rate, which can be attributed to the prophase and interphase cell images having high visual similarities. The true-positive values are higher for the other classes. This denotes that our proposed models were able to give good classification results. We observe that there is a higher rate of misclassification between adjacent classes for the proposed model and the misclassfication value is quite low. This phenomenon comes as an additional advantage stemming from the inclusion of time information. In contrast, for the ResNet18 model without the incorporation of time information, misclassifications are evenly distributed across all classes, and the inclusion of time information mitigates this issue.
Average precision, recall, and F-score for each stage of mitosis when trained using the Zhong et al.’s dataset on various proposed models compared to the TC3 model proposed by Zhong et al. [19]. Boldface indicates the best performance.
The frame-to-frame accuracy value of various proposed models using the Zhong et al.’s dataset. Boldface indicates the best performance.
Illustration of the first three principal components of the embeddings of the proposed models a) Base Model b) Time Encoded ResNet18 c) ResNet18.
Normalized confusion matrices of the prediction for the six class dataset.
Tracking network results
We have also evaluated the reconstruction results from the network for all the datasets. Figs 22 and 23 shows the reconstruction of center-cell tracked by the proposed models at different stages of cell-splitting for the four LiveCellMiner datasets. The background information from other cells are suppressed in the tracker output. This ensures that the features for classification will not contain any background information. This is because, we train the model with a loss function which calculates the loss between the predicted output and a masked image. Fig 24 shows the output of the tracking network from the proposed models for the Zhong et al.’s dataset for six different mitotic phases. Here we could visualize that the neural tracking module of the proposed model is able to track the center-cell, avoid the background clutter from other cells, and extract the features of the cell undergoing mitosis. Training together with the tracker module boosts the performance of the classification model as the feature extraction for the classification happens from the neural features of the tracked cell and background clutter from other cells is completely suppressed, which indeed is the main advantage of our proposed model.
Center-cell tracking for NikonXLight and RecQL4 datasets.
Center-cell tracking for LSM710 and LSD1 datasets.
Illustration of the output of the tracking network of the proposed models for input images belonging to different classes. It can be seen that both models were able to reconstruct approximately the center-cell from the input image belonging to various classes. The base model performs the center-cell reconstruction more effectively than other model since the feature space is at a higher dimension.
Conclusions
In this work, we attempt to study the mitosis process in which the parent cell divides into two identical daughter cells. We also, track the cell division process automatically using deep learning methods. For training the neural network models, we have extracted, video sequences of the cells that undergo cell division and then trained an RNN model, to extract the feature vectors. The use of this RNN has proven to help in better extracting the feature vectors, compared to the conventional classifier models. We have carried out experiments to visualize the feature space to better understand why RNNs are better at identifying the mitosis states. This, in turn, showed that the feature space has a time continuity in the high-dimensional space and clustering is happening as well when trained using RNN model. In addition, we have measured the precision, recall, and F-score as well. This indicates the superiority of the proposed method compared to the ResNet18 baseline. By plotting the confusion matrix, we quantified the amount of misclassification in adjacent classes for both the 3-class and 6-class datasets. Furthermore, the reconstruction results were evaluated to understand how the neural network reconstructs the center-cell during training. Since the center cell is only reconstructed, the network is easily able to extract the features from the cell undergoing cell division, and suppress the clutter from the background cells. This is because during training, the loss is calculated between the predicted output and a masked image. A notable drawback of the proposed method is its inconsistency in outperforming the state-of-the-art across all phases of mitosis. In experiments with the 6-class dataset, the proposed method achieves a higher F-score only for the prophase and anaphase. However, for the other phases, the numbers are comparable to the state-of-the-art model. Similarly, in the case of the 3-class dataset, the F-score for the mitotic phase is lower. A possible reason for this might be the lower number of frames in this phase for training. As future work, we plan to investigate frame interpolation techniques to increase the number of data frames in this class to mitigate this issue.
References
- 1. Noatynska A, Gotta M, Meraldi P. Mitotic spindle (DIS) orientation and Disease: cause or consequence? Journal of Cell Biology. 2012;199(7):1025–35. pmid:23266953
- 2. Potapova T, Gorbsky GJ. The Consequences of chromosome segregation errors in mitosis and meiosis. Biology (Basel). 2017;6(1). pmid:28208750
- 3. Tijhuis AE, Johnson SC, McClelland SE. The emerging links between chromosomal instability (CIN), metastasis, inflammation and tumour immunity. Mol Cytogenet. 2019;12(1):17. pmid:31114634
- 4. Simonetti G, Bruno S, Padella A, Tenti E, Martinelli G. Aneuploidy: Cancer strength or vulnerability? Int J Cancer. 2019;144(1):8–25. pmid:29981145
- 5. Paweletz N. Walther Flemming: pioneer of mitosis research. Nat Rev Mol Cell Biol. 2001;2(1):72–5. pmid:11413469
- 6. er EG. Nuclear Morphology and the Biology of Cancer Cells. Acta Cytol. 2020;64(6):511–519.
- 7. Katayama A, Toss MS, Parkin M, Sano T, Oyama T, Quinn CM, et al. Nuclear morphology in breast lesions: refining its assessment to improve diagnostic concordance. Histopathology. 2021. pmid:34605058
- 8. Way GP, Kost-Alimova M, Shibue T, Harrington WF, Gill S, Piccioni F, et al. Predicting cell health phenotypes using image-based morphology profiling. Mol Biol Cell. 2021;32(9):995–1005. pmid:33534641
- 9. Neumann B, Walter T, Heriche JK, Bulkescher J, Erfle H, Conrad C, et al. Phenotypic profiling of the human genome by time-lapse microscopy reveals cell division genes. Nature. 2010;464(7289):721–7. pmid:20360735
- 10. Held M, Schmitz MH, Fischer B, Walter T, Neumann B, Olma MH, et al. CellCognition: time-resolved phenotype annotation in high-throughput live cell imaging. Nat Methods. 2010;7(9):747–54. pmid:20693996
- 11. Moreno-Andrés D, Bhattacharyya A, Scheufen A, Stegmaier J. LiveCellMiner: A new tool to analyze mitotic progression. PloS one. 2022;17(7):e0270923 pmid:35797385
- 12. Wang W, Douglas D, Zhang J, Kumari S, Enuameh MS, Dai Y, et al. Live-cell imaging and analysis reveal cell phenotypic transition dynamics inherently missing in snapshot data. Sci Adv. 2020;6(36). pmid:32917609
- 13. Cheng X, Dai C, Wen Y, Wang X, Bo X, He S, et al. NeRD: a multichannel neural network to predict cellular response of drugs by integrating multidimensional data. BMC medicine. 2022;20(1):1–16.
- 14. Xia F, Allen J, Balaprakash P, Brettin T, Garcia-Cardona C, Clyde A, et al. A cross-study analysis of drug response prediction in cancer cell lines. Briefings in Bioinformatics. 2021;23(1):bbab356.
- 15. Neumann B, Held M, Liebel U, Erfle H, Rogers P, Pepperkok R, et al. High-throughput RNAi screening by time-lapse imaging of live human cells. Nat Methods. 2006;3(5):385–390. pmid:16628209
- 16. Wählby C, Sintorn IM, Erlandsson F, Borgefors G, Bengtsson E. Combining intensity, edge and shape information for 2D and 3D segmentation of cell nuclei in tissue sections. J Microsc. 2004;215(1):67–76. pmid:15230877
- 17. Durbin R, Eddy SR, Krogh A, Mitchison G. Biological sequence analysis: probabilistic models of proteins and nucleic acids. Camb Univ Press. 1998.
- 18. Forney GD. The viterbi algorithm. Proceedings of the IEEE. 1973;61(3):268–278.
- 19. Zhong Q, Busetto AG, Fededa JP, Buhmann JM, Gerlich DW. Unsupervised modeling of cell morphology dynamics for time-lapse microscopy. Nat Methods. 2012;9(7):711–713. pmid:22635062
- 20. Wold S, Esbensen K, Geladi P. Principal component analysis. Chemometr Intell Lab Syst. 1987;2(1-3):37–52.
- 21.
Mikut R, Bartschat A, Doneit W, Ordiano JÁG, Schott B, Stegmaier J, et al. The MATLAB toolbox SciXMiner: User’s manual and programmer’s guide. arXiv preprint arXiv:1704.03298. 2017.
- 22. Otsu N. A threshold selection method from gray-level histograms. IEEE Trans Syst Man Cybern Syst. 1979;9(1):62–66.
- 23. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–1780. pmid:9377276
- 24. Medsker LR, Jain LC. Recurrent neural networks. Design and Applications. 2001;5:64–67.
- 25.
Chung J, Gulcehre C, Cho K, Bengio Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555. 2014.
- 26. Ondruska P, Posner I. Deep tracking: Seeing beyond seeing using recurrent neural networks. Thirtieth AAAI con on art intell. 2016.
- 27.
Ondruska P, Dequaire J, Wang DZ, Posner I. End-to-end tracking and semantic segmentation using recurrent neural networks. arXiv preprint arXiv:1604.05091. 2016.
- 28.
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2016. p. 770–778.
- 29. Ulman V, et al. An objective comparison of cell-tracking algorithms. Nat Methods. 2017;14(12):1141–1152. pmid:29083403
- 30.
Payer C, Štern D, Neff T, Bischof H, Urschler M. Instance segmentation and tracking with cosine embeddings and recurrent hourglass networks. In: Int Conf on Med Image Compu and Comp Inter. 2018. p. 3–11.
- 31.
Jose A, Roy R, Eschweiler D, Laube I, Azad R, Moreno-Andrés D, et al. End-To-End Classification Of Cell-Cycle Stages With Center-Cell Focus Tracker Using Recurrent Neural Networks. bioRxiv. 2022. Cold Spring Harbor Laboratory.
- 32. Ho Y, Wookey S. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access. 2019;8:4806–4813.
- 33. Schooley A, Moreno-Andrés D, De Magistris P, Vollmer B, Antonin W. The lysine demethylase LSD1 is required for nuclear envelope formation at the end of mitosis. J of cell sci. 2015;128(18):3466–3477. pmid:26224877
- 34. Yokoyama H, Moreno-Andrés D, Astrinidis SA, Hao Y, Weberruss M, Schellhaus AK, et al. Chromosome alignment maintenance requires the MAP RECQL4, mutated in the Rothmund–Thomson syndrome. Life sci all. 2019;2(1). pmid:30718377
- 35. Moreno-Andrés D, Yokoyama H, Scheufen A, Holzer G, Lue H, Schellhaus AK, et al. VPS72/YL1-mediated H2A.Z deposition is required for nuclear reassembly after mitosis. Cells. 2020;9(7):1702. pmid:32708675
- 36. Tharwat A. Classification assessment methods. Appl Comput and Informat. 2020.
- 37.
Brownlee J. How to grid search hyperparameters for deep learning models in python with keras. Disponible en línea. 2016.