Figures
Abstract
When grown on agar surfaces, microbes can produce distinct multicellular spatial structures called colonies, which contain characteristic sizes, shapes, edges, textures, and degrees of opacity and color. For over one hundred years, researchers have used these morphology cues to classify bacteria and guide more targeted treatment of pathogens. Advances in genome sequencing technology have revolutionized our ability to classify bacterial isolates and while genomic methods are in the ascendancy, morphological characterization of bacterial species has made a resurgence due to increased computing capacities and widespread application of machine learning tools. In this paper, we revisit the topic of colony morphotype on the within-species scale and apply concepts from image processing, computer vision, and deep learning to a dataset of 69 environmental and clinical Pseudomonas aeruginosa strains. We find that colony morphology and complexity under common laboratory conditions is a robust, repeatable phenotype on the level of individual strains, and therefore forms a potential basis for strain classification. We then use a deep convolutional neural network approach with a combination of data augmentation and transfer learning to overcome the typical data starvation problem in biological applications of deep learning. Using a train/validation/test split, our results achieve an average validation accuracy of 92.9% and an average test accuracy of 90.7% for the classification of individual strains. These results indicate that bacterial strains have characteristic visual ‘fingerprints’ that can serve as the basis of classification on a sub-species level. Our work illustrates the potential of image-based classification of bacterial pathogens and highlights the potential to use similar approaches to predict medically relevant strain characteristics like antibiotic resistance and virulence from colony data.
Author summary
Since the birth of microbiology, scientists have looked at the patterns of bacterial growth on agar (colony morphology) as a key tool for identifying bacterial species. We return to this traditional approach with modern tools of computer vision and deep learning and show that we can achieve high levels of classification accuracy on a within-species scale, despite what is considered a ‘data-starved’ dataset. Our results show that strains of the environmental generalist and opportunistic pathogen Pseudomonas aeruginosa have a characteristic morphological ‘fingerprint’ that enables accurate strain classification via a custom deep convolutional neural network. Our work points to extensions towards predicting phenotypes of interest (e.g. antibiotic resistance, virulence), and suggests that sample size limitations may be less restrictive than previously thought for deep learning applications in biology, given appropriate use of data augmentation and transfer-learning tools.
Citation: Rattray JB, Lowhorn RJ, Walden R, Márquez-Zacarías P, Molotkova E, Perron G, et al. (2023) Machine learning identification of Pseudomonas aeruginosa strains from colony image data. PLoS Comput Biol 19(12): e1011699. https://doi.org/10.1371/journal.pcbi.1011699
Editor: James O’Dwyer, University of Illinois at Urbana-Champaign, UNITED STATES
Received: October 14, 2022; Accepted: November 20, 2023; Published: December 13, 2023
Copyright: © 2023 Rattray et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper, its Supporting Information files, and on GitHub at https://github.com/GaTechBrownLab/Rattray-2023-PLOSCompBio.
Funding: This research was supported by the National Science Foundation Graduate Research Fellowship Program under grant no. DGE-1650044 (J.B.R), the Foundation for the National Institutes of Health grant no. 1R21AI143296 (S.P.B.), the Center for Disease Control Grant no. 75D30120C09782 (S.P.B.), and the Foundation for the National Institutes of Health grant no. 1R21AI156817 (S.P.B.). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors declare no competing interests.
Introduction
Since Semmelweis, Lister, and Koch began studying pure cultures of microorganisms, scientists have tried to organize the immense diversity of bacteria into orderly categories to understand their behavior, including pathogenesis. In particular, Koch pioneered the method of determining how specific microorganisms cause distinct diseases and highlighted the importance of laboratory culture for the identification and targeted treatment of pathogenic bacteria [1]. By identifying the specific bacterial cause of disease, we can better manage and eradicate the infection. When a sample from an environment or clinical source is plated on solid media at sufficient dilution, individual colonies can be observed, where each colony grows from a single founding bacterial cell (Fig 1A). These clonal groups have characteristic sizes, shapes, edges, textures, and degrees of opacity and color (Fig 1B) [2]. These characteristics describe the morphology of a single colony and formed the basis for the earliest attempts at categorization. In modern clinical laboratories, morphological tests play a role but are increasingly replaced by biochemical and genomic methods of classification [3].
A) Morphological identification of bacterial species from a mixed culture plated on Chocolate Agar: Rothia mucilaginosa (smallest white circles), Neisseria subflava (larger tan round circles), and Streptococcus mitis (large bullseye circles). B) A common example of 2D colony morphology features include the appearance of the colony edges. C) Morphological identification of bacterial strains, in this case called “smooth”, “wrinkly spreader”, and “fuzzy spreader” [4], from a culture containing only Pseudomonas fluorescens plated on King’s Medium B Agar.
While initial classification was morphology based, advances in genome sequencing technology have revolutionized our ability to classify bacterial isolates. Widespread use of 16S rRNA amplicon sequencing not only allows for robust taxon identification to the genus scale [5,6], but also fosters phylogenetic organization of bacterial taxa [7,8]. Marker gene or whole genome sequencing provides finer scale strain identification and allows specific focus on markers of pathogenicity (via identification of virulence factors), drug resistance or other characteristics of interest. Transcriptomic sequencing provides a more direct window into the behavior of pathogens, to assay for example whether virulence factor or drug resistance genes are actually being expressed [9].
While genomic methods are in the ascendancy, morphological characterization of bacterial species has made a resurgence. How a population of bacteria grow is a direct result of the interplay between their genotype, phenotype, and environment [10] and a priori contains critical phenotypic information on the behavior of individual microbes. For example, colony morphotype can indicate specific mutations in a bacteria, e.g. small colony variants and cyclic di-GMP regulation [11] or rugosity and phenazine production [12]. Morphological characterization and data science have collided to provide insights into image-based prediction of microbial taxa [13–18]. The earliest attempts introduced light-scattering for presence-absence detection of microorganisms on surfaces [19,20], which paved the way for image-based identification of Listeria species from light scattering patterns of Listeria colonies [13]. This developed into an automated light scattering sensor called BARDOT (BActerial Rapid Detection using Optical scattering Technology) that produced scatter patterns sensitive enough to distinguish between Escherichia coli serotypes [14], virulence mutants [15], and industrial Staphylococcus contaminants [16]. At the same time, imaged-based prediction diversified to use more accessible standard light microscopes instead of specialized equipment, putting the focus on new analysis methods like colony morphology ontology [18] and deep learning [17,21].
Deep learning algorithms have gained popularity [17,21,22] as machine learning becomes ubiquitous across fields. Several supervised deep learning models have demonstrated near human accuracy in classification of various image datasets [23] and in the case of bacterial biofilms, deep learning has outperformed human characterization of single and mixed species biofilms [24]. Zielinksi et al. [17] investigate several models for colony classification utilizing a deep convolutional neural network (D-CNN) on a rich dataset pre-trained on the ImageNet [25] dataset, achieving a very high validation accuracy of ~ 97%. However, existing work using a D-CNN approach has focused on classification across species, not across strains within a species. Additionally, a ubiquitous challenge in many biology applications (including the one in this paper) is the limited number of samples. For example, the hallmark ImageNet image-based dataset [26–28] contains more than 14 million images across 1000 classes. Outside a colloquial ‘rule of thumb’ that you need at least 1000 samples per class, there is no absolute way to determine a ‘minimum’ sample size [29].
We focus on within species classification of Pseudomonas aeruginosa, a gram-negative bacterium capable of causing severe chronic illness and a major cause of acute hospital-acquired infections [30–33]. While it is widely viewed as a generalist and can be isolated from a wide number of environments including water and soil, it is frequently found in human/animal-impacted environments [30]. P. aeruginosa has large variety of mechanisms that allow it to respond to its environment [34–36]. There are frequently differences in behavior across strains, like type and production of secreted products [37,38], which can result in observable morphological diversity when cultured on agar surfaces [39–43]. Classic examples of morphological variation in P. aeruginosa caused by underlying genetic differences include small colony variants related to cyclic di-GMP regulation [11], rugosity related to phenazine production [12], and mucoidy related to secreted polymer production [44].
In this paper, we use a collection of 69 clinical and environmental P. aeruginosa isolates which present a range of phenotypic features (e.g. exopolysaccharide production, virulence gene expression, and antibiotic resistance profiles) that have been previously described in [45–47]. From these 69 isolates we generate a library of 266 P. aeruginosa colony images in a training / validation dataset, plus a separately generated 69-image test dataset. As expected, we document a high level of variation in the physical appearance of colonies across strains. Additionally, we see that strain morphology in P. aeruginosa is consistent across replicates. We then use a D-CNN approach and a combination of data augmentation [48,49] and transfer learning [50] to classify strains of P. aeruginosa and achieve an average validation accuracy of 92.9% and an average test accuracy of 90.7%.
Results
Morphological variation across strains and replicates
Individual isolates from our 69-strain collection were spotted onto Luria-Bertani (LB) agar plates supplemented with Congo Red, a dye commonly used in microbiology to stain the biofilm matrix and extracellular polysaccharides (Fig 2). This was done in quadruplicate and generated a training and validation library of 266 colonies after quality control (10 colonies were removed due to debris or writing obscuring part of the image). Colonies were grown for 72 hours, which was determined during the pilot experiment to be enough time to reveal major morphological differences while minimizing outward expansion. This resulted in an imaged library of image colonies that ranged in size from 12M (approximately 0.74cm x 0.74cm) to 183M pixels (approximately 2.8cm x 2.8cm).
10 strains (in four-fold replication) were selected randomly without replacement from a complete list of 69 strains via the native sample() command in R. From top to bottom: (left) strain 25, strain, 38, strain 52, strain 113, strain 1; (right) strain 127, strain 316, strain 298, strain 121, strain 174. Colony size has been adjusted for viewing purposes.
In our first approach to characterizing morphological diversity, we use 8 simple descriptive metrics (Fig 3A). Classic morphological descriptive variables include area, radius, perimeter, centroid (the point in which the three medians of the triangle imposed within the image intersect), eccentricity (the ratio of the distance between the foci of the ellipse and its major axis length), circularity (deviation of the perimeter from a superimposed perfect circle), bounding disk center (center of mass of the brightest pixels), and caliper diameter (the caliper dimension obtained by averaging over all orientations).
Histograms are built from all replicates of all strains. A) classic metrics used to describe colony appearance. B) Derived metrics from image processing and computer vision to describe colony complexity including compression ratio (relative reduction in size of data) and 6 descriptive statistics derived from the Sobel–Feldman operator.
To expand on these classic descriptive variables, we next introduce complexity metrics (Fig 3B) used in image processing and computer vision [51,52]. Compression ratio is a single variable that describes how repetitive an image is, i.e. how repetitive are certain patterns in the pixel distribution and how many times do they occur? Specifically, it is a measurement of the reduction in size of pixel data produced by, in our case, the traditional JPEG compression algorithm. An alternate approach to quantifying image complexity is using the Sobel–Feldman operator, a popular method to detect edges in images due to its ability to detect gradients. At every pixel in an image, it calculates the pixel intensity of the vertical and horizontal adjacent pixels and expresses the results as vectors. You can then plot these gradient vectors creating a unique profile for every colony. These profiles can then be summarized by describing their distribution (mean, median, standard deviation, skew, root mean square, and kurtosis). Unsurprisingly, we find that many of these metrics are highly correlated (S1 Fig).
Fig 3 summarizes variation in our metrics across both strains and replicates. In Fig 4, we assess the extent of variation using coefficients of variation (CV = mean / standard deviation), across replicates (black bars) and across strains (grey bars). With the exception of the eccentricity and circularity metrics, we see coefficients of variation across replicates are low (standard deviation << mean), and less than the coefficient of variation across strains. This pattern of variation indicates that colony morphology under common laboratory conditions is a robust, repeatable phenotype on the level of individual strains, and therefore forms a potential basis for strain classification.
Coefficient of variation (mean / standard deviation) across replicates (black) and across strains (grey). With the exception of the eccentricity and circularity metrics, coefficients of variation across replicates are low (standard deviation << mean), and less than the coefficient of variation across strains.
Classifying strain identity from image data
Considering our initial results indicating robust morphological traits for each strain, we reasoned that morphological data could be used to classify colonies via a deep learning approach. Deep convolutional neural networks (D-CNNs) typically require larger datasets than those found in biology- usually a minimum of thousands of examples to train properly [53]. To address this issue, we applied various transfer learning [50] and data augmentation techniques [48,49] (see methods for details). This initial dataset was split into 90% training and 10% validation sets.
High performance on our validation dataset provides some confidence that the CNN approach can successfully classify previously unseen strain images. Yet validation results merit caution, as the validation images were all generated on the same agar plate, imaged on the same day and from the same overnight culture. These shared features raise the risk that a CNN is detecting technical batch effects of the experimental environment (e.g. variations in the lighting conditions) rather than strain-specific features of interest. To address this concern we generated a separate 69-image test dataset (1 image per strain) using distinct overnight cultures and produced by a different experimenter on a different day (following the same protocol).
In a first round of computational experiments, we sought to compare the performance of four transfer learning models (Fig 5 and S1 Table), using our standard data preprocessing and augmentation choices (normalized pixel intensities; shear / zoom / flip / rotation / brightness / scaling data augmentation–see methods). To determine the best approach for our dataset, our performance metrics are accuracy (the number of correct predictions divided by the total number of predictions x100) and loss (a summation of the errors, calculated via cross-entropy, made for each sample). We chose to use cross-entropy for loss instead of negative log-likelihood since it automatically transforms the raw outputs of the neural network into probabilities (via a softmax activation).
The performance of four trained transfer learning models (ResNet-50, VGG-19, MobileNetV2 and Xception, see methods) were evaluated on both validation (blue circles) and test (orange triangles) datasets. Each computational experiment was replicated five-fold (five separate training runs for each model), allowing statistical comparison of approaches. (A) Accuracy scores (the number of correct predictions divided by the total number of predictions x100). (B) Loss scores (a summation of the errors made for each sample). Statistical comparisons are summarized in S1 Table.
Fig 5 illustrates that model performance varies significantly between transfer learning models (e.g. for test accuracy: ANOVA, F = 124.2, df = 3, p < 0.05. See Table B in S1 Table).
MobileNetV2 performs significantly worse than the other transfer learning models on both accuracy and loss metrics (Tukey HSD, p < 0.05 on all comparisons, see Tables C-F in S1 Table), while VGG-19 performs significantly worse than ResNet and Xception on test loss (Table D in S1 Table). ResNet and Xception do not show significantly different performance in posthoc tests (Tables C-F in S1 Table). In terms of best individual model run performance, both the top accuracy and top loss was achieved by the ResNet model (Fig 5), so we use this as our baseline in our subsequent analyses.
In a second round of experiments, we take the best-performing model from Fig 5 (Resnet-50), and assess the contribution of data pre-processing, augmentation and training steps in generating the high levels of performance illustrated in Fig 5. Table 1 illustrates that removing each component of data augmentation leads to a significant drop in model accuracy (Tukey HSD, p < 0.05 on all comparisons, see Tables B-E in S2 Table), and the loss of all augmentation leads to a substantial drop in performance (38.4 +/- 7.4% validation accuracy, 34.7 +/- 7.4% test accuracy). The removal of pre-trained weights results in a similar drop in performance. Parameters for all models are detailed in S3 Table.
To assess contributions, we took the trained ResNet-50 model (Fig 5) as a baseline method and assessed the impact of removing components of our methods pipeline. Each computational experiment was replicated five-fold (five separate training runs for each model), allowing statistical comparison of approaches. Across replicates we report average (+/- standard deviation) accuracy (the number of correct predictions divided by the total number of predictions x100) and loss (a summation of the errors made for each sample). Statistical comparisons are summarized in S2 Table.
The performance of our best performing ResNet-50 model (Fig 5) indicates that while most classification calls are accurate, there remain a number of mistaken calls in our validation tests. To look more closely at these errors, we present in Fig 6 a reduced confusion matrix showing all classification errors across our five replicate runs of the ResNet-50 model. Of the 69 strains considered in this study, only 5 were not classified with perfect precision. Strains 2 and 84 were the most challenging for our model, attaining the lowest test accuracies of 40% and 20% respectively. In contrast, 64 out 69 strains were classified with perfect accuracy across 5 independent model runs.
Of the 69 strains in this study, only 5 strains were not classified with 100% precision (the 5 rows). For brevity, this matrix includes only the strains that were misclassified by our model instead of the whole 69 x 69 confusion matrix.
Our results in Tables 1 and 2 show that D-CNN models are capable of effective bacterial colony classification down to the sub-species (strain) level, and that performance is dependent on details of data pre-processing and augmentation. This success does not rule out that other simpler models can also effectively classify strains. To establish simple model benchmarks, we next evaluate the performance of ‘shallow learning’ classification models (support vector machines, SVMs), trained on our colony metric features (Figs 3 and 4), or on features extracted by our deep learning ResNet 50 model (Table 2). These benchmarking results highlight that SVM tools cannot match the performance of our D-CNN models (Fig 5), whether they are trained on a priori identified colony metric features (Figs 3 and 4) or on features extracted from our successful D-CNN models.
We contrast validation and test accuracy for our trained ResNet model (see also Fig 5 and S1 Table) with accuracy metrics for radial basis function (RBF) SVMs trained on colony metric data (Figs 3 and 4) and on features extracted from the trained ResNet model. See methods for details of SVM models and feature extraction. Statistical comparisons are summarized in S4 Table.
Discussion
Our results show that P. aeruginosa strains have a characteristic morphological ‘fingerprint’ (Figs 2–4) that enables accurate strain classification via a custom deep convolutional neural network (Fig 5). When trained on morphological data, our most successful model can classify strain identity with a test accuracy that exceeds 90% despite what is considered a “data starved” dataset. Our work points to extensions towards predicting phenotypes of interest (e.g. antibiotic resistance, virulence), and suggests that sample size limitations may be less restrictive than previously thought for deep learning applications in biology. In the following paragraphs we will review these implications, while addressing caveats and next steps for this research.
The existence of a morphological ‘fingerprint’ that can be classified via machine learning architecture opens the potential to relate strain identity to key phenotypes of interest, such as antimicrobial resistance (AMR) or virulence. While species differ in their levels and types of virulence [54], there are also significant differences in virulence between strains of the same species [55,56]. In so far as AMR or virulence is a stable and repeatable property of a strain, our classification model could simply “look up” the recorded values in a dataset based on the strain prediction. Yet this “look up” approach is vulnerable to variation within strains due for instance to gain or loss in virulence factor and AMR genes, or mutations in the regulatory control of these genes.
To address the challenge of predicting phenotypes such as virulence or AMR, a future direction would be to extend our deep learning approach to directly predict phenotypes of interest, either through a classification (qualitative trait) or regression (quantitative trait) framework. Our current analysis shows that morphological ‘fingerprints’ are sufficient to identify individual strains, but this does not directly imply that we can use image data to successfully predict disease traits (analogously, human fingerprints are sufficient to identify individuals, but are not considered predictive of human disease traits). We speculate that colony ‘fingerprints’ are likely to allow for trait prediction, due to paths of common causality from bacterial modes of growth to colony morphology and disease phenotypes. Given the established connection between biofilm growth and antibiotic tolerance [57], it is possible that our default growth settings (Fig 2) will reveal differences in patterning that correlate with differences in antibiotic resistance (and/or tolerance). We further anticipate that our approach would become more discriminatory if imaging data was also collected in the presence/absence of relevant stressors of interest (e.g. antibiotics or immune effectors).
Deep learning methods are commonly described as ‘black box’ methodologies [58], presenting challenges for model interrogation and mathematical justification for sample size requirements. Even when classification through deep learning has a monotonically increasing prediction accuracy, the nature of the image collection process in biological research will limit the total size of the dataset. This is a contrast to machine learning applications in other fields that have continually and often exponentially increasing datasets (e.g., spending habits and surveillance cameras). Considering these challenges, arguably our best alternatives for increasing the size of biological datasets for image-based prediction in microbiology are data augmentation and transfer learning, which we combined in this study. Our algorithm transfers learning from the canonical ImageNet dataset while also using standard data augmentation techniques (rotations, shifts, zooms, shears, flips, reflections, brightness.). Future tests to explore the amount of data needed for accurate classification could be run by reducing the size of the dataset by decreasing the amount of data augmentation performed or by reducing the number of replicate colonies used. Moving forward, generative adversarial networks, a class of machine learning frameworks that learns to generate new data by competing two neural networks, is promising technique that could supplement current data augmentation efforts [59,60].
Prior to our data augmentation, we downsampled the original dataset in order to match the image dimensions of the ImageNet dataset which we pretrained on and to minimize computational time. The original dataset contains colony images of 12 million to 183 million pixels (depending on colony size) while our downsized colony images each contained a standardized 60 thousand pixels. The success of strain classification based on these heavily down-sampled images suggests that classification is possible based only on more ‘macro’ scale morphological features that are still discriminable at this resolution. Yet this down-sampling is essentially throwing away a rich array of more micro-scale image features (Fig 7), and so opens the possibility for future analyses to explore these features. The inset panels in Fig 7 illustrate striking local and finer-scale features that are present across strains. These panels suggest an additional data augmentation option, where each colony image is processed into multiple 60K chunks, sampling at different locations within the image and at different spatial scales. Comparing model performance using 60K images taken at different spatial scales could provide valuable information on what are the most diagnostic spatial scales.
Insets reveal finer-scale features at smaller spatial scales.
Our analysis to date implicitly assumes that our image data represent independent samples of P. aeruginosa morphology, which due to phylogenetic relationships among these strains is not the case. A simple null expectation given these phylogenetic relationships is that strains that are most morphologically similar (i.e., similar morphological metrics, Fig 3; and more commonly confused in classification, Fig 6) will tend to have a shorter phylogenetic distance. Under this phylogenetic null model, strains with minimal phylogenetic differences will have overlapping morphological metrics and will not provide dependable classification results. We do not expect this pattern to hold uniformly, as we know that single-gene changes can generate large-scale changes to colony morphology [11,12,44]. Future work can systematically explore the impacts of single gene mutations through automated screening of transposon mutagenesis libraries [61], potentially aiding in the discovery of previously unrecognized genes involved in colony / biofilm formation. We also encourage future work applying our general methodology to sets of strains in other species, as we expect our approach will generalize beyond the model organism P. aeruginosa. In summary, the level of classification accuracy achieved in this work illustrates the potential of image-based prediction tools, deep learning, transfer learning, and data augmentation in the characterization of bacterial strains.
Methods
Strain culturing
This study uses a collection of 69 clinical and environmental P. aeruginosa isolates [45–47], each grown and imaged in quadruplicate. We used Luria-Bertani (LB) agar (1.5%) plates supplemented with 0.08% Congo Red, a dye commonly used to characterize biofilm formation which binds to compounds in the extracellular biofilm matrix. In an effort to minimize variation in plate preparation and the age of the plates, 20ml volume plates were made 24 hours before each experiment with a large, motorized pipette and kept in a sealed plastic sleeve at 4C. Strains were inoculated into LB broth and incubated shaking overnight at 37C. The day of the experiment, each plate was then sliced into 4 sections using a sterilized razor and tweezers, which both limited colony outgrowth and prevented the diffusion of small molecules between colonies. 5μl of the overnight culture was then spotted onto pre-prepared, sectioned LB agar Congo Red plates. After the spots dried, plates were parafilmed to retain moisture and placed in a 37C static incubator. All strains were incubated at 37C for 72 hours, which was determined during the pilot experiment to be enough time to reveal major morphological differences while minimizing outward expansion.
Colony imaging
After 72 hours, strains were imaged on a Nikon Eclipse Ti inverted microscope at 4X using the DS-Fi2 colored camera. This results in a pixel-to-size ratio is 2.07uM per pixel. Large, stitched images were required to scan an entire colony and were generated with Nikon’s built-in software. Some replicate colonies had to be excluded from analysis due to error (strains 13, 47, 78, 86, 122, 124, 132, 210 have 3 replicates and strain 329 has 2 replicates). This resulted in a dataset of 266 P. aeruginosa colony images that ranged in size from 12,644,652 pixels (approximately 0.74cm x 0.74cm) to 183,329,280 pixels (approximately 2.8cm x 2.8cm). The downsized images used in the manuscript can be found at https://github.com/GaTechBrownLab/Rattray-2023-PLOSCompBio/.
Morphological and complexity metrics
For the morphological and complexity measurements, colony images were analyzed in Mathematica. Images were scaled down by a factor of four, to make computations timelier, and the backgrounds were cropped out by using a watershed transformation, binarization, and selection of the contiguous non-zero area surrounding the colony. The morphological metrics (Area, EquivalentDiskRadius, PerimeterLength, MeanCentroidDistance, Eccentricity, Circularity, BoundingDiskCenter, and MeanCaliperDiameter) were all calculated using the native Mathematica commands. For the complexity metrics, CompRatio was calculated by comparing the number of bytes in the uncompressed and compressed (using the native Mathematica command) colony images. A Sobel-Freidman operator was applied in both x and y dimensions and Sobel histograms were generated. MeanSobel, MedianSobel, SDSobel, RootMSSobel, SkewSobel, and KurtosisSobel were all then calculated using the native Mathematica commands.
Data pre-processing for machine learning
Our experimental pipeline is an integration of several steps: preprocessing, transfer learning, and data augmentation techniques, all aimed at maximizing the predictive performance of our deep learning model. In the preprocessing phase, all images are first downsampled to a uniform size of 224x224 pixels, consistent with the image dimensions used in the pretrained ImageNet database. To preserve the original aspect ratio and minimize distortions during the resizing process, we employed inter-area interpolation.
In addition to resizing, another important step in our preprocessing is image normalization. This includes two primary aspects. The first is the correction of image backgrounds: the algorithm iterates over the images and replaces any black pixels with white ones. This process aids in standardizing the image backgrounds across our dataset, ensuring that the model does not get biased by variations in the background and thus enhancing its robustness to different imaging conditions.
The second aspect of image normalization involves rescaling the pixel intensities. Raw images are composed of pixels with intensity values that usually fall within a range of 0 to 255. These large values may slow down the learning process and make the model more susceptible to suboptimal solutions. To mitigate this, we applied a transformation to rescale the pixel values to a 0–1 range. This process, often termed pixel normalization, can help improve the efficiency of the learning process by preventing large input values from causing numerical instability and by ensuring that the input features have a similar data distribution, thus assisting the optimization algorithm to converge more quickly [62].
Overall, these preprocessing steps serve as an integral part of the pipeline, shaping the data into a more suitable form for the model to learn effectively and efficiently, while ensuring robustness to potential sources of variation.
Data augmentation
In order to further optimize our model’s performance and ensure its robustness, we utilized various data augmentation techniques. This approach, known to improve the generalization capability of deep learning models by artificially expanding the training dataset [63], involves making stochastic transformations to the training images, thereby mimicking variations likely to be seen in the real-world application of the model. One such transformation was brightness augmentation, where the brightness of the image is varied randomly within a range. Specifically, the brightness was altered by factors between 0.2 and 0.8, effectively simulating various lighting conditions that the images might be subjected to in real-world scenarios [64]. This rescaling process not only aids in normalizing the data but also helps in mitigating the risk of gradient explosion, a phenomenon that can hinder the learning process [65]. Random rotations within a range of -20 to 20 degrees and random shifts in image width and height by up to 20% of the original size are also used. These techniques simulate changes in the object’s orientation and position within the frame, thus allowing our model to learn invariance to such alterations [66]. The application of shear transformations and zoom operations, both by up to 20%, mimic the effects of perspective distortion and changes in the object-camera distance respectively. These techniques enhance the model’s ability to recognize the subject regardless of variations in the camera perspective or object scale [49]. Finally, we also incorporated horizontal flipping of images, a transformation especially useful for our dataset since it does not possess any inherent left-right bias. This method doubles the size of the dataset and helps the model generalize better by learning symmetrical features [67].
Train/validation/test split
Images from the initial 4 replicate imaging experiment were divided (following augmentation) into a 90–10 train/validation split. The imaging protocol was repeated by a separate experimenter to generate an independent test sample.
Transfer learning models
In our study, we employed four deep learning models to classify bacterial strains, leveraging the advantages of both transfer learning and architectural diversity. Each model used was pre-trained on the ImageNet dataset and further fine-tuned to our specific task. This section details the deep learning models used, the modifications we made, and our rationale for these choices. Parameters for all models are detailed in S3 Table. The code can be accessed at https://github.com/GaTechBrownLab/Rattray-2023-PLOSCompBio/.
Before the training of any models, the topmost layer of each pre-trained model was removed, and replaced by a new flatten layer. This was followed by two pairs of dense and dropout layers, and finally, a softmax dense output layer. The dense layers were designed using Rectified Linear Unit (ReLU) activations [68], while the final dense layer employs a softmax activation function. This transforms the output into a probability distribution across 69 neurons, each corresponding to a unique strain identification within our dataset. The resulting confidence vector is obtained when the output layer is activated.
The decision to downsize the new dense layers from the original size (e.g., from 2048 to 1024 in ResNet-50) was influenced by the size of our dataset in contrast to ImageNet. Dropout layers were also incorporated to prevent overfitting, following the transfer learning research conducted by [69].
ResNet-50
ResNet-50 [70] is known for its residual blocks, which mitigate the vanishing and exploding gradient issues common in deep learning models. In our application, the weights of the pre-existing layers were frozen, and only the newly added dense layers were trained—a method known as ’feature extraction.’ This method exploits the pre-trained model’s ability to extract generalized features, while the added layers learn to make predictions based on these features.
The use of ResNet-50 in bacterial strain identification tasks is well-documented. A study [71] employed a ResNet-based approach to classify three species of gram-positive bacteria, achieving an accuracy of 81%. Another study [72] compared the performance of various deep learning models, including ResNet, for bacilli detection, demonstrating the efficacy of these architectures in bacterial identification tasks. These previous successful uses of ResNet-50 provide a robust motivation for its selection in our study.
MobileNetV2
We employed the MobileNetV2 model for its compactness and efficiency, making it suitable for mobile and embedded vision applications [73]. Similar to ResNet-50, we used the feature extraction approach with this model.
VGG-19
The VGG-19 model [53] was also utilized in our study due to its proven efficacy in various classification tasks. We implemented the feature extraction strategy, leveraging the learned features from the model and fine-tuning our custom layers for our specific task.
Xception
The Xception model was incorporated due to its unique architecture, designed to recognize nuanced patterns [74]. For Xception, we adopted a ’fine-tuning’ strategy, training all layers of the network with a smaller learning rate. This allowed the model to adjust the learned weights more accurately to our problem.
Benchmark models
For comparison, two ’shallow learning’ methods were used as benchmarks. First, a Radial Basis Function (RBF) Support Vector Machine (SVM) was employed for strain classification using colony metric data (Fig 3), following the successful application demonstrated by Chen et al. [75] in bacterial classification.
Additionally, a hybrid approach was adopted to exploit the strengths of both deep and shallow learning. Here, we extracted feature vectors from our top-performing deep learning model, ResNet-50, which were then used as input for a RBF SVM. This process of feature extraction involves using ResNet-50’s pre-trained weights to generate a representative vector of the input image. These vectors essentially encapsulate the critical visual patterns recognized by the model, thus serving as robust predictors in the SVM. This hybrid approach involves loading the pre-trained ResNet-50 model with weights trained on the ImageNet dataset and setting the ’include_top’ parameter to False. This configuration returns a model that does not include the final dense layer, making it suitable for feature extraction. An image is loaded and preprocessed to match the input requirements of ResNet-50, and the model’s ’predict’ function is then used to generate the feature vector. This feature vector is then used as input to the SVM for classification. This transfer learning strategy has been successfully applied in numerous studies, including Rahmayuna et al. [76], who achieved a remarkable classification accuracy of 90.33%.
Supporting information
S1 Fig. Correlation matrix for all morphological and complexity metrics.
Morphological (green) and complexity (orange) metrics.
https://doi.org/10.1371/journal.pcbi.1011699.s001
(DOCX)
S1 Table. Statistical comparisons of transfer learning methods (Fig 5).
The performance of four trained transfer learning models (ResNet-50, VGG-19, MobileNetV2 and Xception, see methods) were evaluated on both validation and test datasets. Each computational experiment was replicated five-fold (five separate training runs for each model), allowing statistical comparison of approaches. Table A. Summary statistics. Across replicates we report average (+/- standard deviation) accuracy (the number of correct predictions divided by the total number of predictions x100) and loss (a summation of the errors made for each sample). Table B. ANOVA table. Tables C-F. Post-hoc pairwise tests (Tukey HSD with alpha = 0.05).
https://doi.org/10.1371/journal.pcbi.1011699.s002
(DOCX)
S2 Table. Performance contribution of data pre-processing, augmentation and training.
To assess contributions we took the trained ResNet-50 model (Fig 5 and Table A in S1 Table) as a baseline method, and assessed the impact of removing components of our methods pipeline. Five-fold replicated results are summarized in Table 1. Table A. ANOVA table of data in Table 1. Tables B-E. Post-hoc pairwise tests (Tukey HSD with alpha = 0.05).
https://doi.org/10.1371/journal.pcbi.1011699.s003
(DOCX)
S4 Table. Performance comparison with shallow learning (SVM) models.
We contrast validation and test accuracy for our trained ResNet model (see also Table 1) with accuracy metrics for radial basis function (RBF) SVMs trained on colony metric data (Fig 3 and 4) and on features extracted from the trained ResNet model. See methods for details of SVM models and feature extraction. Table A. ANOVA of Table 2. Tables B-C. Post-hoc pairwise tests (Tukey HSD with alpha = 0.05).
https://doi.org/10.1371/journal.pcbi.1011699.s005
(DOCX)
Acknowledgments
We thank Jennifer Farrell for culturing the mixed community in Fig 1 and members of the Center for Microbial Dynamics and Infection (CMDI), in particular Dr. Peter Yunker, Dr. Conan Zhao, and Dr. Yifei Wang, for valuable comments and discussion on this work.
References
- 1. Koch R. Ueber den augenblicklichen Stand der bakteriologischen Choleradiagnose. Zeitschrift für Hygiene und Infektionskrankheiten 1983 14:1 [Internet]. 1893 Dec [cited 2022 Jul 19];14(1):319–38. Available from: https://link.springer.com/article/10.1007/BF02284324
- 2.
Madigan MT, Brock TD. Brock biology of microorganisms. 12th editi. Biology of Microorganisms. San Francisco, CA: Pearson/Benjamin Cummings; 2009.
- 3.
Bergey D, Krieg NR, Holt JG. Bergey’s manual of systematic bacteriology. Baltimore, MD: Williams & Wilkins; 1984.
- 4. Rainey PB, Travisano M. Adaptive radiation in a heterogeneous environment. Nature 1998 394:6688 [Internet]. 1998 Jul 1 [cited 2022 Jul 18];394(6688):69–72. Available from: https://www.nature.com/articles/27900
- 5. Woo PCY, Lau SKP, Teng JLL, Tse H, Yuen KY. Then and now: use of 16S rDNA gene sequencing for bacterial identification and discovery of novel bacteria in clinical microbiology laboratories. Clin Microbiol Infect [Internet]. 2008 [cited 2022 Jul 19];14(10):908–34. Available from: https://pubmed.ncbi.nlm.nih.gov/18828852/ pmid:18828852
- 6. Yarza P, Yilmaz P, Pruesse E, Glöckner FO, Ludwig W, Schleifer KH, et al. Uniting the classification of cultured and uncultured bacteria and archaea using 16S rRNA gene sequences. Nat Rev Microbiol [Internet]. 2014 [cited 2022 Jul 19];12(9):635–45. Available from: https://pubmed.ncbi.nlm.nih.gov/25118885/ pmid:25118885
- 7. Woese CR. Bacterial evolution. Microbiol Rev [Internet]. 1987 Jun [cited 2022 Jul 19];51(2):221. Available from: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC373105/ pmid:2439888
- 8. Woese C. The universal ancestor. Proceedings of the National Academy of Sciences [Internet]. 1998 Jun 9 [cited 2022 Jul 19];95(12):6854–9. Available from: https://www.pnas.org/doi/abs/10.1073/pnas.95.12.6854 pmid:9618502
- 9. Güell M, Yus E, Lluch-Senar M, Serrano L. Bacterial transcriptomics: what is beyond the RNA horiz-ome? Nature Reviews Microbiology 2011 9:9 [Internet]. 2011 Aug 12 [cited 2022 Jul 19];9(9):658–69. Available from: https://www.nature.com/articles/nrmicro2620 pmid:21836626
- 10. Kalziqi A, Yanni D, Thomas J, Ng SL, Vivek S, Hammer BK, et al. Immotile Active Matter: Activity from Death and Reproduction. Phys Rev Lett [Internet]. 2018 Jan 5 [cited 2022 Jul 17];120(1):018101. Available from: https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.120.018101 pmid:29350941
- 11. Starkey M, Hickman JH, Ma L, Zhang N, de Long S, Hinz A, et al. Pseudomonas aeruginosa rugose small-colony variants have adaptations that likely promote persistence in the cystic fibrosis lung. J Bacteriol [Internet]. 2009 Jun [cited 2016 May 18];191(11):3492–503. Available from: http://www.ncbi.nlm.nih.gov/pubmed/19329647 pmid:19329647
- 12. Recinos DA, Sekedat MD, Hernandez A, Cohen TS, Sakhtah H, Prince AS, et al. Redundant phenazine operons in Pseudomonas aeruginosa exhibit environment-dependent expression and differential roles in pathogenicity. Proc Natl Acad Sci U S A [Internet]. 2012 Nov 20 [cited 2022 Jul 3];109(47):19420–5. Available from: www.pnas.org/cgi/doi/10.1073/pnas.1213901109 pmid:23129634
- 13. Banada PP, Guo S, Bayraktar B, Bae E, Rajwa B, Robinson JP, et al. Optical forward-scattering for detection of Listeria monocytogenes and other Listeria species. Biosens Bioelectron. 2007 Mar 15;22(8):1664–71. pmid:16949268
- 14. Tang Y, Kim H, Singh AK, Aroonnual A, Bae E, Rajwa B, et al. Light Scattering Sensor for Direct Identification of Colonies of Escherichia coli Serogroups O26, O45, O103, O111, O121, O145 and O157. PLoS One [Internet]. 2014 Aug 19 [cited 2022 Jul 17];9(8). Available from: /pmc/articles/PMC4138183/ pmid:25136836
- 15. Singh AK, Sun X, Bai X, Kim H, Abdalhaseib MU, Bae E, et al. Label-free, non-invasive light scattering sensor for rapid screening of Bacillus colonies. J Microbiol Methods. 2015 Feb 1;109:56–66. pmid:25533215
- 16. Alsulami TS, Zhu X, Abdelhaseib MU, Singh AK, Bhunia AK. Rapid detection and differentiation of Staphylococcus colonies using an optical scattering technology. Anal Bioanal Chem [Internet]. 2018 Sep 1 [cited 2022 Jul 17];410(22):5445–54. Available from: https://pubmed.ncbi.nlm.nih.gov/29796901/ pmid:29796901
- 17. Zieliński B, Plichta A, Misztal K, Spurek P, Brzychczy-Włoch M, Ochońska D. Deep learning approach to bacterial colony classification. PLoS One [Internet]. 2017 Sep 1 [cited 2022 Jul 17];12(9):e0184554. Available from: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0184554
- 18. Sousa AM, Pereira MO, Lourenço A. MorphoCol: An ontology-based knowledgebase for the characterisation of clinically significant bacterial colony morphologies. J Biomed Inform [Internet]. 2015 Jun 1 [cited 2022 Jul 17];55:55–63. Available from: https://pubmed.ncbi.nlm.nih.gov/25817920/ pmid:25817920
- 19. Fratamico PM, Strobaugh TR, Medina MB, Gehring AG. Detection of Escherichia coli 0157:H7 using a surface plasmon resonance biosensor. Biotechnology Techniques 1998 12:7 [Internet]. 1998 [cited 2022 Aug 23];12(7):571–6. Available from: https://link.springer.com/article/10.1023/A:1008872002336
- 20. Perkins EA, Squirrell DJ. Development of instrumentation to allow the detection of microorganisms using light scattering in combination with surface plasmon resonance. Biosens Bioelectron [Internet]. 2000 Jan [cited 2022 Aug 23];14(10–11):853–9. Available from: https://pubmed.ncbi.nlm.nih.gov/10945460/ pmid:10945460
- 21. Turra G, Arrigoni S, Signoroni A. CNN-Based Identification of Hyperspectral Bacterial Signatures for Digital Microbiology. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) [Internet]. 2017 [cited 2022 Jul 17];10485 LNCS:500–10. Available from: https://link.springer.com/chapter/10.1007/978-3-319-68548-9_46
- 22. Andreini P, Bonechi S, Bianchini M, Mecocci A, Scarselli F. A deep learning approach to bacterial colony segmentation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) [Internet]. 2018 [cited 2022 Jul 17];11141 LNCS:522–33. Available from: https://link.springer.com/chapter/10.1007/978-3-030-01424-7_51
- 23. Geirhos R, Janssen DHJ, Schütt HH, Rauber J, Bethge M, Wichmann FA. Comparing deep neural networks against humans: object recognition when the signal gets weaker. 2017 Jun 21 [cited 2022 Jul 17]; Available from: https://arxiv.org/abs/1706.06969v2
- 24. Buetti-Dinh A, Galli V, Bellenberg S, Ilie O, Herold M, Christel S, et al. Deep neural networks outperform human expert’s capacity in characterizing bioleaching bacterial biofilm composition. Biotechnology Reports. 2019 Jun 1;22:e00321. pmid:30949441
- 25. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, et al. ImageNet Large Scale Visual Recognition Challenge. Int J Comput Vis [Internet]. 2014 Sep 1 [cited 2022 Jul 17];115(3):211–52. Available from: https://arxiv.org/abs/1409.0575v3
- 26.
Deng J, Berg AC, Li K, Fei-Fei L. What Does Classifying More than 10,000 Image Categories Tell Us? In: Proceedings of the 11th European Conference on Computer Vision: Part V. Berlin, Heidelberg: Springer-Verlag; 2010. p. 71–84. (ECCV’10).
- 27. Russakovsky O, Fei-Fei L. Attribute learning in large-scale datasets. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) [Internet]. 2012 [cited 2022 Jul 18];6553 LNCS(PART 1):1–14. Available from: https://collaborate.princeton.edu/en/publications/attribute-learning-in-large-scale-datasets
- 28. Russakovsky O, Deng J, Huang Z, Berg AC, Fei-Fei L. Detecting avocados to Zucchinis: What have we done, and where are we going? Proceedings of the IEEE International Conference on Computer Vision [Internet]. 2013 [cited 2022 Jul 18];2064–71. Available from: https://collaborate.princeton.edu/en/publications/detecting-avocados-to-zucchinis-what-have-we-done-and-where-are-w
- 29. Balki I, Amirabadi A, Levman J, Martel AL, Emersic Z, Meden B, et al. Sample-Size Determination Methodologies for Machine Learning in Medical Imaging Research: A Systematic Review. Canadian Association of Radiologists Journal [Internet]. 2019 Nov 1 [cited 2022 Jul 17];70(4):344–53. Available from: https://journals.sagepub.com/doi/10.1016/j.carj.2019.06.002?url_ver=Z39.88-2003&rfr_id=ori%3Arid%3Acrossref.org&rfr_dat=cr_pub++0pubmed pmid:31522841
- 30. Diggle SP, Whiteley M. Microbe profile: Pseudomonas aeruginosa: Opportunistic pathogen and lab rat. Microbiology (United Kingdom) [Internet]. 2020 Oct 10 [cited 2022 Jul 17];166(1):30–3. Available from: https://www.microbiologyresearch.org/content/journal/micro/10.1099/mic.0.000860 pmid:31597590
- 31. Moradali MF, Ghods S, Rehm BHA. Pseudomonas aeruginosa Lifestyle: A Paradigm for Adaptation, Survival, and Persistence. Front Cell Infect Microbiol [Internet]. 2017 Feb 15 [cited 2022 Jul 17];7(FEB):39. Available from: /pmc/articles/PMC5310132/ pmid:28261568
- 32. Nathwani D, Raman G, Sulham K, Gavaghan M, Menon V. Clinical and economic consequences of hospital-acquired resistant and multidrug-resistant Pseudomonas aeruginosa infections: A systematic review and meta-analysis. Antimicrob Resist Infect Control [Internet]. 2014 Oct 20 [cited 2022 Jul 17];3(1):1–16. Available from: https://aricjournal.biomedcentral.com/articles/10.1186/2047-2994-3-32 pmid:25371812
- 33. Elborn JS. Cystic fibrosis. The Lancet. 2016 Nov 19;388(10059):2519–31.
- 34. Pohl S, Klockgether J, Eckweiler D, Khaledi A, Schniederjans M, Chouvarine P, et al. The extensive set of accessory Pseudomonas aeruginosa genomic components. FEMS Microbiol Lett [Internet]. 2014 Jul 1 [cited 2022 Jul 18];356(2):235–41. Available from: https://academic.oup.com/femsle/article/356/2/235/542446 pmid:24766399
- 35. Freschi L, Vincent AT, Jeukens J, Emond-Rheault JG, Kukavica-Ibrulj I, Dupont MJ, et al. The Pseudomonas aeruginosa Pan-Genome Provides New Insights on Its Population Structure, Horizontal Gene Transfer, and Pathogenicity. Genome Biol Evol [Internet]. 2019 Jan 1 [cited 2022 Jul 17];11(1):109–20. Available from: https://pubmed.ncbi.nlm.nih.gov/30496396/ pmid:30496396
- 36. Poulsen BE, Yang R, Clatworthy AE, White T, Osmulski SJ, Li L, et al. Defining the core essential genome of Pseudomonas aeruginosa. Proc Natl Acad Sci U S A [Internet]. 2019 May 14 [cited 2022 Jul 17];116(20):10072–80. Available from: /pmc/articles/PMC6525520/ pmid:31036669
- 37. Lebreton F, Snesrud E, Hall L, Mills E, Galac M, Stam J, et al. A panel of diverse Pseudomonas aeruginosa clinical isolates for research and development. JAC Antimicrob Resist [Internet]. 2021 Sep 30 [cited 2022 Jul 18];3(4). Available from: https://academic.oup.com/jacamr/article/3/4/dlab179/6458700 pmid:34909689
- 38. Kiyaga S, Kyany’a C, Muraya AW, Smith HJ, Mills EG, Kibet C, et al. Genetic Diversity, Distribution, and Genomic Characterization of Antibiotic Resistance and Virulence of Clinical Pseudomonas aeruginosa Strains in Kenya. Front Microbiol. 2022 Mar 14;13:699. pmid:35369511
- 39. Kirisits MJ, Prost L, Starkey M, Parsek MR. Characterization of Colony Morphology Variants Isolated from Pseudomonas aeruginosa Biofilms. Appl Environ Microbiol [Internet]. 2005 Aug [cited 2022 Jul 17];71(8):4809. Available from: /pmc/articles/PMC1183349/ pmid:16085879
- 40. Ikeno T, Fukuda K, Ogawa M, Honda M, Tanabe T, Taniguchi H. Small and rough colony pseudomonas aeruginosa with elevated biofilm formation ability isolated in hospitalized patients. Microbiol Immunol [Internet]. 2007 [cited 2022 Jul 17];51(10):929–38. Available from: https://pubmed.ncbi.nlm.nih.gov/17951982/ pmid:17951982
- 41. Rakhimova E, Munder A, Wiehlmann L, Bredenbruch F, Tümmler B. Fitness of Isogenic Colony Morphology Variants of Pseudomonas aeruginosa in Murine Airway Infection. PLoS One [Internet]. 2008 Feb 27 [cited 2022 Jul 17];3(2):1685. Available from: /pmc/articles/PMC2246019/
- 42. Azimi S, Roberts AEL, Peng S, Weitz JS, McNally A, Brown SP, et al. Allelic polymorphism shapes community function in evolving Pseudomonas aeruginosa populations. The ISME Journal 2020 14:8 [Internet]. 2020 Apr 27 [cited 2022 Jul 17];14(8):1929–42. Available from: https://www.nature.com/articles/s41396-020-0652-0 pmid:32341475
- 43. Vanderwoude J, Fleming D, Azimi S, Trivedi U, Rumbaugh KP, Diggle SP. The evolution of virulence in Pseudomonas aeruginosa during chronic wound infection. Proceedings of the Royal Society B [Internet]. 2020 Oct 28 [cited 2022 Jul 17];287(1937):20202272. Available from: https://royalsocietypublishing.org/doi/10.1098/rspb.2020.2272 pmid:33081616
- 44. Boucher JC, Yu H, Mudd MH, Deretic V. Mucoid Pseudomonas aeruginosa in cystic fibrosis: characterization of muc mutations in clinical isolates and analysis of clearance in a mouse model of respiratory infection. Infect Immun [Internet]. 1997 [cited 2022 Jul 20];65(9):3838–46. Available from: https://pubmed.ncbi.nlm.nih.gov/9284161/ pmid:9284161
- 45. Pirnay JP, Bilocq F, Pot B, Cornelis P, Zizi M, van Eldere J, et al. Pseudomonas aeruginosa Population Structure Revisited. PLoS One [Internet]. 2009 Nov 13 [cited 2022 Jul 17];4(11):e7740. Available from: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0007740 pmid:19936230
- 46. Dettman JR, Rodrigue N, Aaron SD, Kassen R. Evolutionary genomics of epidemic and nonepidemic strains of Pseudomonas aeruginosa. Proc Natl Acad Sci U S A [Internet]. 2013 Dec 24 [cited 2022 Jul 17];110(52):21065–70. Available from: www.pnas.org/cgi/doi/10.1073/pnas.1307862110 pmid:24324153
- 47. Shrestha SD, Guttman DS, Perron GG. Draft Genome Sequences of 10 Environmental Pseudomonas aeruginosa Strains Isolated from Soils, Sediments, and Waters. Genome Announc [Internet]. 2017 [cited 2022 Jul 17];5(34). Available from: https://journals.asm.org/doi/10.1128/genomeA.00804-17 pmid:28839021
- 48. Mikołajczyk A, Grochowski M. Data augmentation for improving deep learning in image classification problem. 2018 International Interdisciplinary PhD Workshop, IIPhDW 2018. 2018 Jun 18;117–22.
- 49. Wong SC, Gatt A, Stamatescu V, McDonnell MD. Understanding Data Augmentation for Classification: When to Warp? 2016 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2016. 2016 Dec 22;
- 50.
Pratt LY. Discriminability-Based Transfer between Neural Networks. In: Hanson S, Cowan J, Giles C, editors. Advances in Neural Information Processing Systems [Internet]. Morgan-Kaufmann; 1992. Available from: https://proceedings.neurips.cc/paper/1992/file/67e103b0761e60683e83c559be18d40c-Paper.pdf
- 51. Yu H, Winkler S. Image complexity and spatial information. 2013 5th International Workshop on Quality of Multimedia Experience, QoMEX 2013—Proceedings. 2013;12–7.
- 52. Danielsen J, Nordenfelt P. Computer vision-based image analysis of bacteria. Methods in Molecular Biology [Internet]. 2017 [cited 2022 Oct 11];1535:161–72. Available from: https://link.springer.com/protocol/10.1007/978-1-4939-6673-8_10 pmid:27914078
- 53. Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition. 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings [Internet]. 2014 Sep 4 [cited 2022 Jul 17]; Available from: https://arxiv.org/abs/1409.1556v6
- 54. Shaw LP, Wang AD, Dylus D, Meier M, Pogacnik G, Dessimoz C, et al. The phylogenetic range of bacterial and viral pathogens of vertebrates. Mol Ecol [Internet]. 2020 Sep 1 [cited 2022 Jul 17];29(17):3361–79. Available from: https://onlinelibrary.wiley.com/doi/full/10.1111/mec.15463 pmid:32390272
- 55. Subedi D, Vijay AK, Kohli GS, Rice SA, Willcox M. Comparative genomics of clinical strains of Pseudomonas aeruginosa strains isolated from different geographic sites. Scientific Reports 2018 8:1 [Internet]. 2018 Oct 23 [cited 2022 Jul 17];8(1):1–14. Available from: https://www.nature.com/articles/s41598-018-34020-7 pmid:30353070
- 56. Medina-Rojas M, Stribling W, Snesrud E, Garry BI, Li Y, Gann PM, et al. Comparison of Pseudomonas aeruginosa strains reveals that Exolysin A toxin plays an additive role in virulence. Pathog Dis [Internet]. 2020 Feb 1 [cited 2022 Jul 17];78(1):10. Available from: https://academic.oup.com/femspd/article/78/1/ftaa010/5804881 pmid:32167551
- 57. Sharma D, Misba L, Khan AU. Antibiotics versus biofilm: an emerging battleground in microbial communities. Antimicrobial Resistance & Infection Control 2019 8:1 [Internet]. 2019 May 16 [cited 2022 Jul 25];8(1):1–10. Available from: https://aricjournal.biomedcentral.com/articles/10.1186/s13756-019-0533-3 pmid:31131107
- 58. Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 2019 1:5 [Internet]. 2019 May 13 [cited 2022 Jul 20];1(5):206–15. Available from: https://www.nature.com/articles/s42256-019-0048-x pmid:35603010
- 59. Tolstikhin I, Bousquet O, Schölkopf B, Thierbach K, Bazin PL, de Back W, et al. Generative Adversarial Networks. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) [Internet]. 2014 Jun 10 [cited 2022 Jul 17];11046 LNCS(NeurIPS):1–9. Available from: https://arxiv.org/abs/1406.2661v1
- 60. Radford A, Metz L, Chintala S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. 4th International Conference on Learning Representations, ICLR 2016—Conference Track Proceedings [Internet]. 2015 Nov 19 [cited 2022 Jul 17]; Available from: https://arxiv.org/abs/1511.06434v2
- 61. Cabeen MT, Leiman SA, Losick R. Colony-morphology screening uncovers a role for the Pseudomonas aeruginosa nitrogen-related phosphotransferase system in biofilm formation. Mol Microbiol [Internet]. 2016 Feb 1 [cited 2022 Jul 17];99(3):557. Available from: /pmc/articles/PMC5130288/ pmid:26483285
- 62. LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proceedings of the IEEE. 1998;86(11):2278–323.
- 63. Perez L, Wang J. The Effectiveness of Data Augmentation in Image Classification using Deep Learning. 2017 Dec 13 [cited 2023 Jul 11]; Available from: https://arxiv.org/abs/1712.04621v1
- 64. Alomar K, Aysel HI, Cai X. Data Augmentation in Classification and Segmentation: A Survey and New Strategies. Journal of Imaging 2023, Vol 9, Page 46 [Internet]. 2023 Feb 17 [cited 2023 Jul 11];9(2):46. Available from: https://www.mdpi.com/2313-433X/9/2/46/htm pmid:36826965
- 65. Ioffe S, Szegedy C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. CoRR [Internet]. 2015;abs/1502.0. Available from: http://arxiv.org/abs/1502.03167
- 66. Shorten C, Khoshgoftaar TM. A survey on Image Data Augmentation for Deep Learning. J Big Data [Internet]. 2019 Dec 1 [cited 2023 Jul 11];6(1):1–48. Available from: https://journalofbigdata.springeropen.com/articles/10.1186/s40537-019-0197-0
- 67. Cubuk ED, Zoph B, Mané D, Vasudevan V, Le Q V. AutoAugment: Learning Augmentation Policies from Data. CoRR [Internet]. 2018;abs/1805.0. Available from: http://arxiv.org/abs/1805.09501
- 68. Nair V, Hinton GE. Rectified Linear Units Improve Restricted Boltzmann Machines. In: International Conference on Machine Learning. 2010.
- 69. Yosinski J, Clune J, Bengio Y, Lipson H. How transferable are features in deep neural networks? CoRR [Internet]. 2014;abs/1411.1. Available from: http://arxiv.org/abs/1411.1792
- 70. He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. CoRR [Internet]. 2015;abs/1512.0. Available from: http://arxiv.org/abs/1512.03385
- 71. Alhammad S, Zhao K, Jennings A, Hobson P, Smith DF, Baker B, et al. Efficient DNN-Based Classification of Whole Slide Gram Stain Images for Microbiology. In: 2021 Digital Image Computing: Techniques and Applications (DICTA). 2021. p. 1–8.
- 72. Shwetha V, Prasad K, Mukhopadhyay C, Banerjee B, Chakrabarti A. Automatic Detection of Bacilli Bacteria from Ziehl-Neelsen Sputum Smear Images. In: 2021 2nd International Conference on Communication, Computing and Industry 40 (C2I4). 2021. p. 1–5.
- 73. Sandler M, Howard AG, Zhu M, Zhmoginov A, Chen LC. Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation. CoRR [Internet]. 2018;abs/1801.0. Available from: http://arxiv.org/abs/1801.04381
- 74. Chollet F. Xception: Deep Learning with Depthwise Separable Convolutions. CoRR [Internet]. 2016;abs/1610.0. Available from: http://arxiv.org/abs/1610.02357
- 75. Chen Y, Yu G, Long Y, Teng J, You X, Liao BQ, et al. Application of radial basis function artificial neural network to quantify interfacial energies related to membrane fouling in a membrane bioreactor. Bioresour Technol [Internet]. 2019;293:122103. Available from: https://www.sciencedirect.com/science/article/pii/S0960852419313331 pmid:31505391
- 76. Rahmayuna N, Rahardwika D, Sari A, Setiadi DRIM, Rachmawanto E. Pathogenic Bacteria Genus Classification using Support Vector Machine. In 2018. p. 23–7.