<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">ResProt</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Res Protoc</journal-id>
      <journal-title>JMIR Research Protocols</journal-title>
      <issn pub-type="epub">1929-0748</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v9i7e16843</article-id>
      <article-id pub-id-type="pmid">32673276</article-id>
      <article-id pub-id-type="doi">10.2196/16843</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Protocol</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Protocol</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Deep Learning Frameworks for Rapid Gram Stain Image Data Interpretation: Protocol for a Retrospective Data Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Eysenbach</surname>
            <given-names>Gunther</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Chih</surname>
            <given-names>HuiJun</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Sheng</surname>
            <given-names>Bo</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Kim</surname>
            <given-names>Hee</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Heinrich-Lanz-Center for Digital Health</institution>
            <institution>Medical Faculty Mannheim</institution>
            <institution>Heidelberg University</institution>
            <addr-line>Theodor-Kutzer-Ufer 1-3</addr-line>
            <addr-line>Mannheim, 68167</addr-line>
            <country>Germany</country>
            <phone>49 015229530383</phone>
            <email>HeeEun.Kim@medma.uni-heidelberg.de</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2826-2796</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Ganslandt</surname>
            <given-names>Thomas</given-names>
          </name>
          <degrees>Prof Dr med</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-6864-8936</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Miethke</surname>
            <given-names>Thomas</given-names>
          </name>
          <degrees>Prof Dr med</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7451-5857</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Neumaier</surname>
            <given-names>Michael</given-names>
          </name>
          <degrees>Prof Dr med</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-0410-3931</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Kittel</surname>
            <given-names>Maximilian</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2580-5038</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Heinrich-Lanz-Center for Digital Health</institution>
        <institution>Medical Faculty Mannheim</institution>
        <institution>Heidelberg University</institution>
        <addr-line>Mannheim</addr-line>
        <country>Germany</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Institute of Medical Microbiology and Hygiene</institution>
        <institution>Medical Faculty Mannheim</institution>
        <institution>Heidelberg University</institution>
        <addr-line>Mannheim</addr-line>
        <country>Germany</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Institute for Clinical Chemistry</institution>
        <institution>Medical Faculty Mannheim</institution>
        <institution>Heidelberg University</institution>
        <addr-line>Mannheim</addr-line>
        <country>Germany</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Hee Kim <email>HeeEun.Kim@medma.uni-heidelberg.de</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>7</month>
        <year>2020</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>13</day>
        <month>7</month>
        <year>2020</year>
      </pub-date>
      <volume>9</volume>
      <issue>7</issue>
      <elocation-id>e16843</elocation-id>
      <history>
        <date date-type="received">
          <day>21</day>
          <month>1</month>
          <year>2020</year>
        </date>
        <date date-type="rev-request">
          <day>20</day>
          <month>2</month>
          <year>2020</year>
        </date>
        <date date-type="rev-recd">
          <day>14</day>
          <month>4</month>
          <year>2020</year>
        </date>
        <date date-type="accepted">
          <day>19</day>
          <month>4</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Hee Kim, Thomas Ganslandt, Thomas Miethke, Michael Neumaier, Maximilian Kittel. Originally published in JMIR Research Protocols (http://www.researchprotocols.org), 13.07.2020.</copyright-statement>
      <copyright-year>2020</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Research Protocols, is properly cited. The complete bibliographic information, a link to the original publication on http://www.researchprotocols.org, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="http://www.researchprotocols.org/2020/7/e16843/" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>In recent years, remarkable progress has been made in deep learning technology and successful use cases have been introduced in the medical domain. However, not many studies have considered high-performance computing to fully appreciate the capability of deep learning technology.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This paper aims to design a solution to accelerate an automated Gram stain image interpretation by means of a deep learning framework without additional hardware resources.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We will apply and evaluate 3 methodologies, namely fine-tuning, an integer arithmetic–only framework, and hyperparameter tuning.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The choice of pretrained models and the ideal setting for layer tuning and hyperparameter tuning will be determined. These results will provide an empirical yet reproducible guideline for those who consider a rapid deep learning solution for Gram stain image interpretation. The results are planned to be announced in the first quarter of 2021.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Making a balanced decision between modeling performance and computational performance is the key for a successful deep learning solution. Otherwise, highly accurate but slow deep learning solutions can add value to routine care.</p>
        </sec>
        <sec sec-type="registered-report">
          <title>International Registered Report Identifier (IRRID)</title>
          <p>DERR1-10.2196/16843</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>high performance computing</kwd>
        <kwd>rapid Gram stain classification</kwd>
        <kwd>image data analysis</kwd>
        <kwd>deep learning</kwd>
        <kwd>convolutional neural network</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>In recent years, remarkable progress has been made in deep learning due to the emergence of big data processing technology. Deep learning is a family of machine learning that consists of multiple neurons in multiple layers. A neuron is a mathematical function with weights and biases, known as parameters. It receives real numbers from the neurons in the previous layer, generates another real number, and transmits it to the neurons in the next layer. The parameters for each of these neurons are optimally determined by a backpropagation algorithm, such as stochastic gradient descent, that looks for the minimum of a function. This contributes to the success of deep learning of image data compared with conventional techniques because it is able to learn the intrinsic data features without handcrafted feature engineering.</p>
      <p>Gram stain is a laboratory procedure for the rapid classiﬁcation and identiﬁcation of microbial pathogens. Unlike other microbiology processes that can be fully automated [<xref ref-type="bibr" rid="ref1">1</xref>], the interpretation of Gram stain images still relies on human users, such as a physician or trained medical technical assistant. Although Gram stain seems like a medical image analysis problem, to the best of our knowledge, only 1 research paper [<xref ref-type="bibr" rid="ref2">2</xref>] has used the deep learning method to automate the Gram stain analysis. One of the challenges is that microbial pathogens, particularly gram-negative organisms, and the background material, such as bloodstains, look highly similar on a slide. Furthermore, in cases of a low density of bacteria in a clinical sample, the manual search by microscopy is tedious and will only examine a fraction of the microscope slide, and thus may be error-prone.</p>
      <p>Smith et al [<xref ref-type="bibr" rid="ref2">2</xref>] achieved 94.9% classification accuracy out of 468 Gram-stained slides from positive blood cultures. The respective sensitivities and specificities were 98.4% and 75.0% for gram-positive cocci in chains and pairs, 93.2% and 97.2% for gram-positive cocci in clusters, and 96.3% and 98.1% for Gram-negative rods. The authors reused the pretrained model called Inception-v3 [<xref ref-type="bibr" rid="ref3">3</xref>] and retrained the last layer with their image data (100,213 image crops of 146×146 pixels for training and testing) instead of constructing an end-to-end model from the scratch. This approach is called transfer learning [<xref ref-type="bibr" rid="ref4">4</xref>] because it reuses precomputed model parameters. The major advantage of transfer learning is that it is able to reduce computational costs for the model training.</p>
      <p>Despite the high accuracy achieved by Smith et al, there are still many open questions to be addressed. With regard to modeling, transfer learning could be improved with fine-tuning [<xref ref-type="bibr" rid="ref5">5</xref>] instead of modifying only the last layer. Fine-tuning is a type of transfer learning that allows one to adjust the ratio of retraining layers. This is highly relevant to Gram stain classification because Inception-v3 was constructed with ImageNet [<xref ref-type="bibr" rid="ref6">6</xref>], which contains 1.2 million nature images and 1000 image classes, such as dogs and cats, that are unrelated to Gram stain images. Therefore, increasing the number of unfrozen (retraining) layers (ie, decreasing the number of freezing layers) with Gram stain images is anticipated to yield better results. With respect to computational performance, it takes about 9 minutes to classify a whole-slide image of 28,032×28,032 pixels with a computer consisting of Intel Core i7 (Intel Corp) with 32 GB of RAM and a Nvidia GTX 1070 graphics processing unit (GPU). The turnaround time for multiple samples encountered in the medical laboratory would not provide timely decision-making, although this solution can run the job 24/7.</p>
      <p>This study aims to design a rapid deep learning solution for Gram stain interpretation without acquiring hardware resources, and it provides the optimal proportion for the fine-tuning. The hypothesis and the study design to evaluate the hypothesis will be explained in the following section.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <p>This section addresses the hypothesis, study design, data collection and description, study population, statistical considerations for nonbiased model construction and evaluation, and tools in detail.</p>
      <sec>
        <title>Study Design</title>
        <p>This study does not investigate a clinical hypothesis but performs an empirical evaluation of a deep learning framework for Gram stain image interpretation. The hypothesis to be tested is that the optimization of a deep learning framework will perform better than a scale-up strategy with a single GPU. In order to test this hypothesis, two strategies will be examined, as shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>. The scale-up strategy stacks up more computer capabilities (model A, highlighted in blue). On the other hand, the optimization strategy tunes the granular configuration of a deep learning framework (model B, depicted in green).</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Two strategies will be compared. The lineage of model A is the implementation of the scale-up strategy (highlighted in blue). On the other hand, the lineage of model B is the implementation of the optimization strategy (depicted in green). Model A is the base model with FPA framework, while model B replaces the floating-point arithmetic with IAO. Each model is built on top of a predecessor model. For instance, model A1 is empowered with a single GPU and model B1 is empowered with the optimal minibatch size. FPA: floating-point arithmetic. GPU: graphics processing unit. IAO: integer arithmetic only.</p>
          </caption>
          <graphic xlink:href="resprot_v9i7e16843_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>In order to avoid model bias, 4 pretrained models (Mobilenet [<xref ref-type="bibr" rid="ref7">7</xref>] with versions 1 and 2, and Inception [<xref ref-type="bibr" rid="ref3">3</xref>] with versions 3 and 4) will be evaluated. Mobilenet and Inception were originally trained with ImageNet [<xref ref-type="bibr" rid="ref6">6</xref>], which contains 1.2 million nature images and 1000 classes. Therefore, it is necessary to retrain them on the new data set, as described in the “Data Collection” section, because Gram stain interpretation is barely linked to the original task. The method that retrains a pretrained model without changing network architecture is called fine-tuning. Fine-tuning is a specific transfer learning technique that modifies more layers than the last layer, also known as the fully connected layer, and their corresponding weights and biases. The optimal proportion of the fine-tuning will be empirically determined in this study. Concretely, 10 implementations will be evaluated, with the proportions of frozen layer to unfrozen layer (retraining layer) at 9:1, 8:2, ongoing up to 0:10 (from the shallow strategy to the deep strategy). The computational performance will be measured by time to achieve the target accuracy metric [<xref ref-type="bibr" rid="ref8">8</xref>] of 95%.</p>
        <p>Once the base model is implemented with Tensorflow [<xref ref-type="bibr" rid="ref9">9</xref>] (Google Corp) as described above, model A will be accelerated with 2 approaches. Model A1 is the scale-up implementation built on top of model A with a single GPU. In contrast, model B is a mutated model of model A because it replaces floating-point arithmetic with an integer arithmetic–only deep learning framework called Tensorflow Lite [<xref ref-type="bibr" rid="ref10">10</xref>]. Model B1 aims to detect the optimal batch size for Gram stain classification. Model B2 and B3 activate batch normalization and weight normalization. Normalization penalizes large input numbers and weights. Finally, model B4 prunes the model network by adjusting the dropout rate, which speeds up the execution time. These hyperparameters—batch size, batch normalization, weight normalization, and dropout rate—are chosen based on the findings from previous studies [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref14">14</xref>]. The specification of the hyperparameter space is defined in <xref ref-type="table" rid="table1">Table 1</xref>. The boundary is chosen to be wide to account for many possible combinations of the hyperparameters.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Specification of the hyperparameter space for the model B family. Minibatch size and dropout rate are quantified to avoid an exhaustive search.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="100"/>
            <col width="300"/>
            <col width="300"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Model</td>
                <td>Hyperparameter</td>
                <td>Original value range</td>
                <td>Quantified values</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>B1</td>
                <td>Minibatch size</td>
                <td>{1-infinity}</td>
                <td>{32, 64, 128, 256, 512}</td>
              </tr>
              <tr valign="top">
                <td>B2</td>
                <td>Batch normalization</td>
                <td>{on, off}</td>
                <td>{on, off}</td>
              </tr>
              <tr valign="top">
                <td>B3</td>
                <td>Weight normalization</td>
                <td>{on, off}</td>
                <td>{on, off}</td>
              </tr>
              <tr valign="top">
                <td>B4</td>
                <td>Dropout rate</td>
                <td>{0-1}</td>
                <td>{0, 0.1, 0.2, 0.3, 0.4, 0.5}</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>The objective of this study is to understand the relation between computation time and each hyperparameter, not to create a hyperparameter optimization [<xref ref-type="bibr" rid="ref15">15</xref>] that searches for a global minimum or the local minima of hyperparameters.</p>
      </sec>
      <sec>
        <title>Data Collection</title>
        <p>This study will use 8728 Gram stain images from between 2015 and 2018 for modeling and images generated in 2019 for testing. Data are archived in a workstation at the Institute for Clinical Chemistry at the Medical Faculty Mannheim of Heidelberg University, Germany. Sample images and labels are shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>. The sizes of the cropped images vary from 800×600 pixels to 1920×1080 pixels. Note that the image is not a whole-slide image, but a crop of the interests—the microorganisms and the slide background.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>A sample image of Gram stain data. The image label does not have a link to personal information.</p>
          </caption>
          <graphic xlink:href="resprot_v9i7e16843_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>The label data corresponding to the image are stored in a central database for reporting purposes and extracted for this study. Each image is associated with 2 labels: (1) Gram stain class (ie, either gram-positive or gram-negative) and (2) a class for the genus. The genus label includes 5 of the most frequently encountered germs: (1) <italic>Staphylococcus</italic>, (2) <italic>Escherichia</italic>, (3) <italic>Streptococcus</italic>, (4) <italic>Enterococcus</italic>, and (5) others, for the rest of the germs that are rarely presented. This setting prevents a potential risk of identifying a patient with an extremely rare microorganism. Of the 8728 images, 446 images (5.11%) are associated with multiple classes. These images with multiple class labels are excluded because object recognition, that is, bacterial differentiation, is beyond the scope of this study.</p>
      </sec>
      <sec>
        <title>Study Population</title>
        <p>The population for this study is a group of sepsis patients, whose blood samples contain at least one harmful bacterium, such as <italic>Staphylococcus</italic>, <italic>Escherichia</italic>, or <italic>Streptococcus</italic>. This study does not recruit control and treatment groups. However, a set of gram-positive images would be regarded as one group, while gram-negative images would be regarded as a comparison group for this study. This is a retrospective data analysis reusing an archived image data set, as described in the “Data Collection” section.</p>
      </sec>
      <sec>
        <title>Statistical Analysis</title>
        <p>This section will address and describe the 3 underlying statistical considerations towards a solid study design: (1) the class balancing strategy for the input data set, (2) the proper split ratio for training and evaluation, and (3) the metric for the model evaluation.</p>
        <sec>
          <title>Class Balancing</title>
          <p>Imbalanced input data sets are a common limiting factor that degrades model quality. Chawla et al [<xref ref-type="bibr" rid="ref16">16</xref>] systematically proved that data augmentation can improve the imbalanced class problem and demonstrated the benefits. In the classification of bone lesions from x-ray images, Gupta et al [<xref ref-type="bibr" rid="ref17">17</xref>] mitigated the small number of positive samples by using data augmentation. In the given image data available for this study, gram-positive results are twice as frequent as gram-negative results. In order to balance the class proportion, this study will apply the data augmentation technique, which enriches the data set by cropping, rotating, zooming, and ﬂipping the given images.</p>
        </sec>
        <sec>
          <title>Split Ratio</title>
          <p>The data set will be split into a training set, a hold-out development set, and a test set. The hold-out development set is different from the test set, as the development set will only be used for tuning the model parameters in order to not bias classification. The training set for deep learning algorithms is increased to 99% of the entire data set when there are more than a million data points. However, this study will follow best practice in machine learning, in which the splitting ratio is 60%, 20%, and 20% [<xref ref-type="bibr" rid="ref18">18</xref>], because the available data points for this study are 8728 images.</p>
          <p>Cross-validation is not used in this study for model validation. Cross-validation estimates the performance of the model statistically, but it is not the chosen method for evaluating a deep learning model. For instance, a 10-fold cross-validation creates a model with 9 folds and tests the model with the hold-out data (1 fold) 10 times. When we evaluate the model with 100 whole-slide (28,032×28,032 pixels) images, each round will take at least 900 minutes with a workstation powered by Intel Core i7 with 32 GB of RAM and a Nvidia GTX 1070 GPU, which is the same hardware setting and the same image size used in the study by Smith et al [<xref ref-type="bibr" rid="ref2">2</xref>]. For the 10-fold cross-validation, it would take more than 6 days to evaluate 1 model, which is beyond the capacity of the workstation.</p>
        </sec>
        <sec>
          <title>Metric for Evaluation</title>
          <p>Despite considerable efforts that have been devoted to deep learning research, not many studies consider the computational efficiency, but focus solely on model evaluation. In order to provide more insightful information, this study will evaluate models with the classical metrics, such as accuracy, confusion matrix, and area under curve, as well as the training and testing times of models, to achieve the target accuracy proposed by the Stanford Data Analytics for What’s Next project team [<xref ref-type="bibr" rid="ref8">8</xref>].</p>
        </sec>
      </sec>
      <sec>
        <title>Apparatus</title>
        <p>This study will use Tensorflow [<xref ref-type="bibr" rid="ref9">9</xref>] (floating-point arithmetic framework) and Tensorflow Lite [<xref ref-type="bibr" rid="ref10">10</xref>] for deep learning solutions. Both of these frameworks are open source tools developed by the Google Brain team, and they are able to accelerate deep learning calculation by using multicore central processing units (CPUs) and GPUs. With regard to a model-debugging tool, TensorBoard will be used to graphically track all execution history.</p>
        <p>All solutions will be developed and deployed in the data center at the Heinrich-Lanz-Center for Digital Health. The hardware configuration for this study is one Intel Xeon Silver 4110 CPU (Intel Corp), one Tesla V100 PCIe 32 GB GPU (Nvidia Corp), and 189 GB memory. The server is virtualized by Docker technology (Docker Inc) [<xref ref-type="bibr" rid="ref19">19</xref>] for reproducible research.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <p>This study will provide an empirical guideline on how to accelerate a high-performance deep learning model without losing predictive power. Concretely, 2 results will be highlighted: (1) the performance improvement of an integer arithmetic–only deep learning framework for Gram stain image classification and (2) the optimal setting of fine-tuning and hyperparameters for 4 pretrained models (Mobilenet version 1 and 2 and Inception version 3 and 4). All models and the code for training and evaluation will be freely accessible in a public repository for reproducible research.</p>
      <p>As of October 2019, this study has been approved by the institutional review board of Medical Faculty Mannheim of Heidelberg University, and the image data for the retrospective data analysis are available. The results are planned to be announced in the first quarter of 2021.</p>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Limitations</title>
        <p>Distributed computing across multiple machines will not be covered in this study. Although it is the usual method to process big data, it is not always the most efficient choice to process the data. According to Boden et al [<xref ref-type="bibr" rid="ref20">20</xref>], distributed systems are surprisingly inefﬁcient in training machine learning models compared with a single workstation. The size of the input data for this study is 25 GB and it fits into the capacity of a standalone workstation. In future work, we would like to study Apache Spark [<xref ref-type="bibr" rid="ref20">20</xref>], which enables distributed machine learning model training when the data do not fit into the memory of a single computer.</p>
        <p>This study does not aim to propose novel neural network architecture, which requires many days of GPU processing time with state-of-the-art computational infrastructure that is not available within the scope of this project. Also, designing an outperforming architecture for image classification is a saturated topic, as many researchers have devoted their endeavors to this problem in the last decade. Nevertheless, for those who are interested in this topic, Elsken et al published a state-of-the-art review paper [<xref ref-type="bibr" rid="ref21">21</xref>] that provides an overview of existing works and categorizes them into 3 dimensions.</p>
      </sec>
      <sec>
        <title>Risk of Project Failure</title>
        <p>An insufficient amount of image data could lead to an underpowered deep learning solution. The proper input data size is still an open question in the computer vision community. The answer is, “it depends.” It depends on the number of classes, image size, image quality, and complexity of the problem. For instance, classifying a black image versus a white image demands fewer input data compared with classifying a gram-positive image versus a gram-negative image.</p>
        <p>In medical data analysis, power analysis is widely applied for determining the minimum sample size required. Unfortunately, power analysis is not applicable to unstructured data such as images. A rule of thumb for a good input size is 1000 images per class [<xref ref-type="bibr" rid="ref21">21</xref>], which was the basis of an object recognition competition that was part of the Pascal Visual Object Classes challenge [<xref ref-type="bibr" rid="ref22">22</xref>]. This study is anticipated to have low failure, since about 5800 gram-positive labeled images and 2700 gram-negative images are available for modeling.</p>
      </sec>
      <sec>
        <title>Data Protection Considerations</title>
        <p>Although this study does not use any personal information for data analysis, the name of the input data consists of a unique identifier for the experiment. This experiment identifier harbors a remote risk of linking back to personal information in the database. In the interest of data protection, this identifier is anonymized and securely stored at the Heinrich-Lanz-Center for Digital Health , which is protected by the hospital network ﬁrewalls. Unlike data pseudonymization, which transforms the identifier, data anonymization is an irreversible technique that removes the identiﬁer permanently. The anonymized data will be archived for reproducibility.</p>
        <p>The study will comply with the latest version of the Declaration of Helsinki [<xref ref-type="bibr" rid="ref23">23</xref>] and Professional Code for Physicians in Germany. Patient names and other personal data are subject to the legal requirements concerning conﬁdential medical communication. They comply with European Directive 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of individuals with regard to the processing of personal data and on the free movement of such data, the EU General Data Protection Regulation and the German Federal Data Protection Act, and the State Data Protection Act Baden-Württemberg.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">CPU</term>
          <def>
            <p>central processing unit</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">GPU</term>
          <def>
            <p>graphics processing unit</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bourbeau</surname>
              <given-names>PP</given-names>
            </name>
            <name name-style="western">
              <surname>Ledeboer</surname>
              <given-names>NA</given-names>
            </name>
          </person-group>
          <article-title>Automation in clinical microbiology</article-title>
          <source>J Clin Microbiol</source>
          <year>2013</year>
          <month>06</month>
          <volume>51</volume>
          <issue>6</issue>
          <fpage>1658</fpage>
          <lpage>65</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://jcm.asm.org/cgi/pmidlookup?view=long&#38;pmid=23515547"/>
          </comment>
          <pub-id pub-id-type="doi">10.1128/JCM.00301-13</pub-id>
          <pub-id pub-id-type="medline">23515547</pub-id>
          <pub-id pub-id-type="pii">JCM.00301-13</pub-id>
          <pub-id pub-id-type="pmcid">PMC3716104</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>Kenneth P</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>Anthony D</given-names>
            </name>
            <name name-style="western">
              <surname>Kirby</surname>
              <given-names>James E</given-names>
            </name>
          </person-group>
          <article-title>Automated Interpretation of Blood Culture Gram Stains by Use of a Deep Convolutional Neural Network</article-title>
          <source>J Clin Microbiol</source>
          <year>2018</year>
          <month>03</month>
          <volume>56</volume>
          <issue>3</issue>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://jcm.asm.org/cgi/pmidlookup?view=long&#38;pmid=29187563"/>
          </comment>
          <pub-id pub-id-type="doi">10.1128/JCM.01521-17</pub-id>
          <pub-id pub-id-type="medline">29187563</pub-id>
          <pub-id pub-id-type="pii">JCM.01521-17</pub-id>
          <pub-id pub-id-type="pmcid">PMC5824030</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <collab>Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A</collab>
          </person-group>
          <article-title>Going Deeper With Convolutions</article-title>
          <year>2015</year>
          <conf-name>Conference on Computer Vision and Pattern Recognition</conf-name>
          <conf-date>Jun 7-12, 2015</conf-date>
          <conf-loc>Boston</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.cv-foundation.org/openaccess/content_cvpr_2015/html/Szegedy_Going_Deeper_With_2015_CVPR_paper.html"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>Hoo-Chang</given-names>
            </name>
            <name name-style="western">
              <surname>Roth</surname>
              <given-names>Holger R</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>Mingchen</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>Le</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>Ziyue</given-names>
            </name>
            <name name-style="western">
              <surname>Nogues</surname>
              <given-names>Isabella</given-names>
            </name>
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>Jianhua</given-names>
            </name>
            <name name-style="western">
              <surname>Mollura</surname>
              <given-names>Daniel</given-names>
            </name>
            <name name-style="western">
              <surname>Summers</surname>
              <given-names>Ronald M</given-names>
            </name>
          </person-group>
          <article-title>Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning</article-title>
          <source>IEEE Trans Med Imaging</source>
          <year>2016</year>
          <month>05</month>
          <volume>35</volume>
          <issue>5</issue>
          <fpage>1285</fpage>
          <lpage>98</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/26886976"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/TMI.2016.2528162</pub-id>
          <pub-id pub-id-type="medline">26886976</pub-id>
          <pub-id pub-id-type="pmcid">PMC4890616</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tajbakhsh</surname>
              <given-names>Nima</given-names>
            </name>
            <name name-style="western">
              <surname>Shin</surname>
              <given-names>Jae Y</given-names>
            </name>
            <name name-style="western">
              <surname>Gurudu</surname>
              <given-names>Suryakanth R</given-names>
            </name>
            <name name-style="western">
              <surname>Hurst</surname>
              <given-names>R Todd</given-names>
            </name>
            <name name-style="western">
              <surname>Kendall</surname>
              <given-names>Christopher B</given-names>
            </name>
            <name name-style="western">
              <surname>Gotway</surname>
              <given-names>Michael B</given-names>
            </name>
          </person-group>
          <article-title>Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning?</article-title>
          <source>IEEE Trans Med Imaging</source>
          <year>2016</year>
          <month>05</month>
          <volume>35</volume>
          <issue>5</issue>
          <fpage>1299</fpage>
          <lpage>1312</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ieeexplore.ieee.org/abstract/document/7426826"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/TMI.2016.2535302</pub-id>
          <pub-id pub-id-type="medline">26978662</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Russakovsky</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Su</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Krause</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Satheesh</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Karpathy</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Khosla</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bernstein</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Berg</surname>
              <given-names>AC</given-names>
            </name>
            <name name-style="western">
              <surname>Fei-Fei</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>ImageNet Large Scale Visual Recognition Challenge</article-title>
          <source>Int J Comput Vis</source>
          <year>2015</year>
          <month>4</month>
          <day>11</day>
          <volume>115</volume>
          <issue>3</issue>
          <fpage>211</fpage>
          <lpage>252</lpage>
          <pub-id pub-id-type="doi">10.1007/s11263-015-0816-y</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Howard</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Kalenichenko</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Weyand</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Andreetto</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Adam</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications</article-title>
          <source>arXiv</source>
          <year>2017</year>
          <month>04</month>
          <day>16</day>
          <access-date>2020-06-17</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1704.04861.pdf">https://arxiv.org/pdf/1704.04861.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Coleman</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Narayanan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Nardi</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bailis</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Olukotun</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ré</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zaharia</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>DAWNBench: An End-to-End Deep Learning Benchmark and Competition</article-title>
          <source>SysML Conference</source>
          <year>2018</year>
          <conf-name>SysML Conference</conf-name>
          <conf-date>February 15-16, 2018</conf-date>
          <conf-loc>Stanford, CA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <collab>Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, Devin M, Ghemawat S, Irving G, Isard M, Kudlur M, Levenberg J, Monga R, Moore S, Murray DG, Steiner B, Tucker P, Vasudevan V, Warden P, Wicke M, Yu Y, Zheng X</collab>
          </person-group>
          <article-title>TensorFlow: A System for Large-Scale Machine Learning</article-title>
          <year>2016</year>
          <conf-name>12th USENIX Symposium on Operating Systems Design and Implementation</conf-name>
          <conf-date>November 2-4, 2016</conf-date>
          <conf-loc>Savannah, GA, USA</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.usenix.org/conference/osdi16/technical-sessions/presentation/abadi"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="web">
          <article-title>TensorFlow Lite</article-title>
          <source>TensorFlow</source>
          <access-date>2020-06-17</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.tensorflow.org/lite">https://www.tensorflow.org/lite</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ioffe</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Szegedy</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift</article-title>
          <year>2015</year>
          <conf-name>International Conference on Machine Learning</conf-name>
          <conf-date>July 7-9, 2015</conf-date>
          <conf-loc>Lille, France</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1502.03167.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goyal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Dollár</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Girshick</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Noordhuis</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Wesolowski</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Kyrola</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tulloch</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour</article-title>
          <source>arXiv</source>
          <access-date>2020-06-16</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1706.02677.pdf">https://arxiv.org/pdf/1706.02677.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Salimans</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Kingma</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks</article-title>
          <year>2016</year>
          <month>06</month>
          <day>3</day>
          <conf-name>Neural Information Processing Systems (NIPS)</conf-name>
          <conf-date>December 5-10, 2016</conf-date>
          <conf-loc>Barcelona, Spain</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1602.07868"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wen</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Learning Structured Sparsity in Deep Neural Networks</article-title>
          <year>2016</year>
          <month>10</month>
          <day>18</day>
          <conf-name>Neural Information Processing Systems (NIPS)</conf-name>
          <conf-date>December 5-10, 2016</conf-date>
          <conf-loc>Barcelona SPAIN</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1608.03665"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bergstra</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bardenet</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bengio</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kégl</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Algorithms for Hyper-Parameter Optimization</article-title>
          <year>2011</year>
          <conf-name>Neural Information Processing Systems (NIPS)</conf-name>
          <conf-date>December 12-17, 2011</conf-date>
          <conf-loc>Granada, Spain</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://papers.nips.cc/paper/4443-algorithms-for-hyper-parameter-optimization.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chawla</surname>
              <given-names>NV</given-names>
            </name>
            <name name-style="western">
              <surname>Bowyer</surname>
              <given-names>KW</given-names>
            </name>
            <name name-style="western">
              <surname>Hall</surname>
              <given-names>LO</given-names>
            </name>
            <name name-style="western">
              <surname>Kegelmeyer</surname>
              <given-names>WP</given-names>
            </name>
          </person-group>
          <article-title>SMOTE: Synthetic Minority Over-sampling Technique</article-title>
          <source>JAIR</source>
          <year>2002</year>
          <month>06</month>
          <day>01</day>
          <volume>16</volume>
          <fpage>321</fpage>
          <lpage>357</lpage>
          <pub-id pub-id-type="doi">10.1613/jair.953</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gupta</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Venkatesh</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chopra</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ledig</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Generative Image Translation for Data Augmentation of Bone Lesion Pathology</article-title>
          <year>2019</year>
          <month>02</month>
          <day>6</day>
          <conf-name>International Conference on Medical Imaging with Deep Learning (MIDL)</conf-name>
          <conf-date>July 8-10, 2019</conf-date>
          <conf-loc>London, United Kingdom</conf-loc>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1902.02248"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ng</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Improving deep neural networks: Hyperparameter tuning, regularization and optimization</article-title>
          <source>Coursera</source>
          <access-date>2020-06-20</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.coursera.org/learn/deep-neural-network">https://www.coursera.org/learn/deep-neural-network</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boettiger</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>An introduction to Docker for reproducible research</article-title>
          <source>SIGOPS Oper Syst Rev</source>
          <year>2015</year>
          <month>01</month>
          <day>20</day>
          <volume>49</volume>
          <issue>1</issue>
          <fpage>71</fpage>
          <lpage>79</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/2723872.2723882"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/2723872.2723882</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boden</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rabl</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Markl</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Distributed Machine Learning-but at what COST?</article-title>
          <year>2017</year>
          <conf-name>Neural Information Processing Systems (NIPS)</conf-name>
          <conf-date>December 4-9, 2017</conf-date>
          <conf-loc>Long Beach, CA, USA</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Srivastava</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Krizhevsky</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sutskever</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Salakhutdinov</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Improving neural networks by preventing co-adaptation of feature detectors</article-title>
          <source>arXiv</source>
          <year>2012</year>
          <month>07</month>
          <day>3</day>
          <access-date>2020-06-19</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1207.0580.pdf">https://arxiv.org/pdf/1207.0580.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Everingham</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Eslami</surname>
              <given-names>SMA</given-names>
            </name>
            <name name-style="western">
              <surname>Van Gool</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>CKI</given-names>
            </name>
            <name name-style="western">
              <surname>Winn</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zisserman</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The Pascal Visual Object Classes Challenge: A Retrospective</article-title>
          <source>Int J Comput Vis</source>
          <year>2014</year>
          <month>6</month>
          <day>25</day>
          <volume>111</volume>
          <issue>1</issue>
          <fpage>98</fpage>
          <lpage>136</lpage>
          <pub-id pub-id-type="doi">10.1007/s11263-014-0733-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <collab>World Medical Association.</collab>
          </person-group>
          <article-title>World Medical Association Declaration of Helsinki. Ethical principles for medical research involving human subjects</article-title>
          <source>Bull World Health Organ</source>
          <year>2001</year>
          <volume>79</volume>
          <issue>4</issue>
          <fpage>373</fpage>
          <lpage>4</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/11357217"/>
          </comment>
          <pub-id pub-id-type="medline">11357217</pub-id>
          <pub-id pub-id-type="pii">S0042-96862001000400016</pub-id>
          <pub-id pub-id-type="pmcid">PMC2566407</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
