pith. machine review for the scientific record. sign in

arxiv: 2604.20030 · v1 · submitted 2026-04-21 · 💻 cs.CV

Recognition: unknown

Learning to count small and clustered objects with application to bacterial colonies

Authors on Pith no claims yet

Pith reviewed 2026-05-10 01:58 UTC · model grok-4.3

classification 💻 cs.CV
keywords bacterial colony countingobject countingFamNetmulti-head attentionresidual connectionscomputer visionimage analysisACFamNet Pro
0
0 comments X

The pith

A neural network extension called ACFamNet Pro counts small and clustered bacterial colonies with a mean error of 9.64 percent.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper addresses the problem of automated bacterial colony counting from images, which is needed for vaccine and antibiotic development but challenged by small object sizes, clustering, annotation costs, and generalization across species. Building on FamNet, which handles clustered objects with few labels, the authors introduce ACFamNet with a new region of interest pooling that aligns features to manage small colonies better. ACFamNet Pro then adds multi-head attention to weigh objects dynamically and residual connections to improve learning. Under 5-fold cross-validation, this yields a mean normalised absolute error of 9.64%, which is 2.23% better than ACFamNet and 12.71% better than the original FamNet. A sympathetic reader would care because this could speed up lab work by replacing tedious manual counts with reliable automation.

Core claim

The authors establish that augmenting FamNet with region of interest pooling with alignment, optimised feature engineering, multi-head attention, and residual connections produces ACFamNet Pro, which counts bacterial colonies more accurately than prior versions, reaching a mean normalised absolute error of 9.64% in cross-validation tests on colony images.

What carries the argument

ACFamNet Pro, which applies multi-head attention and residual connections to FamNet along with aligned region of interest pooling to process small clustered objects in images.

If this is right

  • Counts of bacterial colonies can be obtained automatically with lower error for vaccine development data.
  • The approach reduces the impact of high annotation costs by building on FamNet's few-shot capabilities.
  • Dynamic weighting via attention improves performance on varying cluster densities and sizes.
  • Residual connections allow better training for these dense small-object scenes.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same architecture changes could apply to counting other small clustered items, such as cells in medical imaging.
  • Testing on images from more bacterial species would strengthen claims of cross-species generalization.
  • Combining this with newer vision transformers might push errors lower still.

Load-bearing premise

The improvements in error rates result directly from the added architectural components rather than from dataset-specific optimizations or the particular choice of cross-validation folds.

What would settle it

An experiment that applies the models to a fresh collection of bacterial colony images from unseen species and measures whether ACFamNet Pro still shows lower error than FamNet.

Figures

Figures reproduced from arXiv: 2604.20030 by Allen Donald, Minghua Zheng, Na Helian, Peter C. R. Lane, Yi Sun.

Figure 1
Figure 1. Figure 1: Four example plate images with colonies of different species, sizes, and colours. [PITH_FULL_IMAGE:figures/full_fig_p009_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Distribution of colony counts in the training and test sets. [PITH_FULL_IMAGE:figures/full_fig_p010_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Overall structure of ACFamNet. The feature correlation and regression modules [PITH_FULL_IMAGE:figures/full_fig_p011_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Illustration of the ACFamNet feature correlation module. [PITH_FULL_IMAGE:figures/full_fig_p012_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Illustration of ACFamNet regression module. [PITH_FULL_IMAGE:figures/full_fig_p013_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Overall structure of ACFamNet Pro. Details of the residual feature enhancement [PITH_FULL_IMAGE:figures/full_fig_p015_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Feature extractor. mechanism allows the model to control the mixing of information between elements, i.e. support feature, to enrich feature representations. The detailed design of the residual feature enhancement module is illus￾trated in [PITH_FULL_IMAGE:figures/full_fig_p016_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Residual feature enhancement module. Feature correlation block. The aim of feature correlation block is to produce a similarity map to robustly highlight regions in the query feature fQ that are similar to the support feature fS. It has three steps: learnable feature projection, feature comparison, and score normalisation. Learnable feature projection The useful features from the support feature fS and que… view at source ↗
Figure 9
Figure 9. Figure 9: Illustration of kernel flipping in FEM. Its purpose is to preserve the spatial structure from the projected support feature fP S. In this illustration, the K dimension is removed from R, fP S, and fR for simplicity, meaning only a support image is used in this example. The motivation behind this design is as follows: suppose that the feature in the projected query feature fP Q corresponding to the position… view at source ↗
Figure 10
Figure 10. Figure 10: Regression module. passed through a 1×1 convolution, and added to the input of the third con￾volutional layer. Finally, the input to the third convolutional layer is added to the output of the final convolutional layer to produce the density map D. The number of convolutional kernels and the kernel size in each convolutional layer are detailed in [PITH_FULL_IMAGE:figures/full_fig_p021_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Counting results for an image with 83 colonies. Left: OpenCFU detects 66 [PITH_FULL_IMAGE:figures/full_fig_p030_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: Counting results for an image with 83 colonies. ACFamNet detects 89.58 [PITH_FULL_IMAGE:figures/full_fig_p030_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: Four plate images with colonies that are completely different from the Synoptics [PITH_FULL_IMAGE:figures/full_fig_p031_13.png] view at source ↗
Figure 14
Figure 14. Figure 14: Illustration of ACFamNet Pro’s prediction. The predicted count and ground [PITH_FULL_IMAGE:figures/full_fig_p037_14.png] view at source ↗
read the original abstract

Automated bacterial colony counting from images is an important technique to obtain data required for the development of vaccines and antibiotics. However, bacterial colonies present unique machine vision challenges that affect counting, including (1) small physical size, (2) object clustering, (3) high data annotation cost, and (4) limited cross-species generalisation. While FamNet is an established object counting technique effective for clustered objects and costly data annotation, its effectiveness for small colony sizes and cross-species generalisation remains unknown. To address the first three challenges, we propose ACFamNet, an extension of FamNet that handles small and clustered objects using a novel region of interest pooling with alignment and optimised feature engineering. To address all four challenges above, we introduce ACFamNet Pro, which augments ACFamNet with multi-head attention and residual connections, enabling dynamic weighting of objects and improved gradient flow. Experiments show that ACFamNet Pro achieves a mean normalised absolute error (MNAE) of 9.64% under 5-fold cross-validation, outperforming ACFamNet and FamNet by 2.23% and 12.71%, respectively.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The paper introduces ACFamNet as an extension of FamNet that uses region-of-interest pooling with alignment and optimised feature engineering to handle small object sizes and clustering in bacterial colony images while mitigating high annotation costs. It further proposes ACFamNet Pro, which adds multi-head attention and residual connections to enable dynamic object weighting and improved gradient flow for better cross-species generalisation. Experiments under 5-fold cross-validation report that ACFamNet Pro attains a mean normalised absolute error (MNAE) of 9.64%, outperforming ACFamNet by 2.23% and FamNet by 12.71%.

Significance. If the reported gains are shown to stem from the architectural changes rather than dataset-specific tuning, the work could provide a targeted improvement for automated colony counting in microbiology, directly supporting data acquisition for vaccine and antibiotic development by reducing manual effort on small, clustered objects across species.

major comments (2)
  1. [Abstract] Abstract: The claim that ACFamNet Pro addresses the core challenge of limited cross-species generalisation is not supported by the described evaluation. The manuscript supplies no information on the number of species, image counts per species, species balance, or whether the 5-fold CV splits hold out entire species (as opposed to random image-level partitioning). Without species-stratified folds, the 2.23% and 12.71% MNAE reductions cannot be interpreted as evidence of improved cross-species robustness.
  2. [Experiments] Experiments section: The headline quantitative result (MNAE 9.64%) is presented without dataset size, total number of images, baseline re-implementation details for FamNet and ACFamNet, or ablation studies that isolate the contribution of ROI pooling with alignment, multi-head attention, and residual connections. These omissions leave the central performance claim only moderately supported and prevent attribution of gains to the proposed components.
minor comments (1)
  1. The abstract would be clearer if it briefly stated the dataset characteristics (e.g., number of images and species) alongside the MNAE figures to allow immediate assessment of the scale of the evaluation.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback, which identifies key areas where additional details and clarifications will strengthen the manuscript. We address each major comment below and will incorporate revisions to improve transparency and support for our claims.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claim that ACFamNet Pro addresses the core challenge of limited cross-species generalisation is not supported by the described evaluation. The manuscript supplies no information on the number of species, image counts per species, species balance, or whether the 5-fold CV splits hold out entire species (as opposed to random image-level partitioning). Without species-stratified folds, the 2.23% and 12.71% MNAE reductions cannot be interpreted as evidence of improved cross-species robustness.

    Authors: We agree that the abstract and evaluation description lack explicit details on dataset composition and split strategy, which limits interpretation of the cross-species generalization claim. The experiments use a multi-species bacterial colony dataset, with 5-fold CV performed at the image level. In the revision, we will expand the abstract and add a dataset subsection specifying the number of species, image counts per species, and balance. We will also clarify that the splits are not species-holdout and adjust the language to indicate that the results show improved performance on the multi-species collection rather than proving explicit cross-species robustness. This addresses the concern without overstating the evidence. revision: yes

  2. Referee: [Experiments] Experiments section: The headline quantitative result (MNAE 9.64%) is presented without dataset size, total number of images, baseline re-implementation details for FamNet and ACFamNet, or ablation studies that isolate the contribution of ROI pooling with alignment, multi-head attention, and residual connections. These omissions leave the central performance claim only moderately supported and prevent attribution of gains to the proposed components.

    Authors: We concur that these omissions reduce the strength of the central claims. The revised manuscript will report the total number of images and dataset size. We will detail the re-implementation of FamNet and ACFamNet, including hyperparameters, training protocols, and any dataset-specific adaptations. We will also include ablation studies that add components sequentially (ROI pooling with alignment, then multi-head attention, then residual connections) to isolate their contributions to the MNAE improvement. These additions will enable better attribution of gains to the proposed elements. revision: yes

Circularity Check

0 steps flagged

No circularity: empirical results on held-out CV folds are independent of model equations

full rationale

The paper's central claims rest on proposing ACFamNet and ACFamNet Pro as architectural extensions to FamNet, followed by reporting MNAE under 5-fold cross-validation. These performance numbers are computed directly on held-out image folds and do not reduce, via any equation in the paper, to quantities defined solely by fitted parameters or by the model's own definitions. No self-definitional loops, fitted-input-as-prediction steps, or load-bearing self-citations appear in the reported derivation or evaluation chain. The result is therefore self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests on standard deep-learning training assumptions and the prior effectiveness of FamNet for clustered objects; no new physical entities or ad-hoc constants are introduced.

free parameters (1)
  • neural network hyperparameters
    Learning rate, attention heads, residual scaling, and pooling parameters are chosen or tuned during training; exact values not stated in abstract.
axioms (1)
  • domain assumption FamNet is effective for clustered objects and costly annotation scenarios
    Invoked as the established baseline whose limitations the new models address.

pith-pipeline@v0.9.0 · 5509 in / 1122 out tokens · 39883 ms · 2026-05-10T01:58:19.177742+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

96 extracted references · 21 canonical work pages · 8 internal anchors

  1. [1]

    Abozeid, A., Alanazi, R., Elhadad, A., Taloba, A.I., Abd El-Aziz, R.M.,

  2. [2]

    Computational Intelligence and Neuroscience 2022, 1549842

    ALarge-ScaleDatasetandDeepLearningModelforDetectingand Counting Olive Trees in Satellite Imagery. Computational Intelligence and Neuroscience 2022, 1549842

  3. [3]

    Deep Learning using Rectified Linear Units (ReLU)

    Agarap, A.F., 2018. DeepLearningusingRectifiedLinearUnits(ReLU). arXiv:1803.08375 [cs, stat]arXiv:1803.08375

  4. [4]

    Automated Counting of Colony Forming Units Using Deep Transfer Learning From a Model for Congested Scenes Analysis

    Albaradei, S.A., Napolitano, F., Uludag, M., Thafar, M., Napolitano, S., Essack, M., Bajic, V.B., Gao, X., 2020. Automated Counting of Colony Forming Units Using Deep Transfer Learning From a Model for Congested Scenes Analysis. IEEE Access 8, 164340–164346

  5. [5]

    A- CCNN: Adaptive CCNN for Density Estimation and Crowd Count- ing, in: 2018 25th IEEE International Conference on Image Processing (ICIP), pp

    Amirgholipour, S., He, X., Jia, W., Wang, D., Zeibots, M., 2018. A- CCNN: Adaptive CCNN for Density Estimation and Crowd Count- ing, in: 2018 25th IEEE International Conference on Image Processing (ICIP), pp. 948–952

  6. [6]

    A Deep Learning Approach to Bacterial Colony Segmentation, Springer International Publishing, Cham

    Andreini, P., Bonechi, S., Bianchini, M., Mecocci, A., Scarselli, F., 2018. A Deep Learning Approach to Bacterial Colony Segmentation, Springer International Publishing, Cham. volume 11141, pp. 522–533

  7. [7]

    An image-processing based automated bac- teria colony counter, in: 2009 24th International Symposium on Com- puter and Information Sciences, IEEE

    Ates, H., Gerek, O.N., 2009. An image-processing based automated bac- teria colony counter, in: 2009 24th International Symposium on Com- puter and Information Sciences, IEEE. pp. 18–23

  8. [8]

    Automated counting of mammalian cell colonies

    Barber, P.R., Vojnovic, B., Kelly, J., Mayes, C.R., Boulton, P., Wood- cock, M., Joiner, M.C., 2001. Automated counting of mammalian cell colonies. Physics in Medicine and Biology 46, 63–76

  9. [9]

    Deep Learning to Detect Bacterial Colonies for the Production of Vaccines arXiv:2009.00926

    Beznik, T., Smyth, P., de Lannoy, G., Lee, J.A., 2020. Deep Learning to Detect Bacterial Colonies for the Production of Vaccines arXiv:2009.00926

  10. [10]

    YOLOv4: Optimal Speed and Accuracy of Object Detection

    Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M., 2020. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXivarXiv:2004.10934. 49

  11. [11]

    Boominathan, L., Kruthiventi, S.S.S., Babu, R.V., 2016. CrowdNet: A Deep Convolutional Network for Dense Crowd Counting, in: Proceed- ings of the 24th ACM International Conference on Multimedia, Associ- ation for Computing Machinery, New York, NY, USA. pp. 640–644

  12. [12]

    Automated Counting of Bacterial Colony Forming Units on Agar Plates

    Brugger, S.D., Baumberger, C., Jost, M., Jenni, W., Brugger, U., Müh- lemann, K., 2012. Automated Counting of Bacterial Colony Forming Units on Agar Plates. PLoS ONE 7, e33695

  13. [13]

    U2-Net and ResNet50-Based Automatic Pipeline for Bacterial Colony Counting

    Cao, L., Zeng, L., Wang, Y., Cao, J., Han, Z., Chen, Y., Wang, Y., Zhong, G., Qiao, S., 2024. U2-Net and ResNet50-Based Automatic Pipeline for Bacterial Colony Counting. Microorganisms 12, 201

  14. [14]

    Bayesian Poisson regression for crowd counting, in: 2009 IEEE 12th International Conference on Com- puter Vision, pp

    Chan, A.B., Vasconcelos, N., 2009. Bayesian Poisson regression for crowd counting, in: 2009 IEEE 12th International Conference on Com- puter Vision, pp. 545–551

  15. [15]

    An automated bacterial colony counting and classification system

    Chen, W.B., Zhang, C., 2009. An automated bacterial colony counting and classification system. Information Systems Frontiers 11, 349–368

  16. [16]

    Automated count- ing of bacterial colonies by image analysis

    Chiang, P.J., Tseng, M.J., He, Z.S., Li, C.H., 2015. Automated count- ing of bacterial colonies by image analysis. Journal of Microbiological Methods 108, 74–82.arXiv:quant-ph/0312207

  17. [17]

    High-Throughput Method for Automated Colony and Cell Counting by Digital Image Analysis Based on Edge Detection

    Choudhry, P., 2016. High-Throughput Method for Automated Colony and Cell Counting by Digital Image Analysis Based on Edge Detection. PLOS ONE 11, e0148469

  18. [18]

    Low-cost, high-throughput, automated counting of bacterial colonies

    Clarke, M.L., Burton, R.L., Hill, A.N., Litorja, M., Nahm, M.H., Hwang, J., 2010. Low-cost, high-throughput, automated counting of bacterial colonies. Cytometry Part A 77A, 790–797

  19. [19]

    Natural Language Processing (Almost) from Scratch

    Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P., 2011. Natural Language Processing (Almost) from Scratch. The Journal of Machine Learning Research 12, 2493–2537

  20. [20]

    Ballard, 1981

    Dana H. Ballard, 1981. Generalizing the Hough Transform to Detect Arbitrary Shapes 13, 111–122

  21. [21]

    Crowd monitoring us- ing image processing

    Davies, A.C., Yin, J.H., Velastin, S.A., 1995. Crowd monitoring us- ing image processing. Electronics & Communication Engineering Journal 7, 37–47. 50

  22. [22]

    An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

    Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N., 2020. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv:2010.11929 [cs] arXiv:2010.11929

  23. [23]

    U-Net: Deep learning for cell counting, detection, and morphometry

    Falk, T., Mai, D., Bensch, R., Çiçek, Ö., Abdulkadir, A., Marrakchi, Y., Böhm, A., Deubner, J., Jäckel, Z., Seiwald, K., Dovzhenko, A., Tietz, O., Dal Bosco, C., Walsh, S., Saltukoglu, D., Tay, T.L., Prinz, M., Palme, K., Simons, M., Diester, I., Brox, T., Ronneberger, O., 2019. U-Net: Deep learning for cell counting, detection, and morphometry. Nature Me...

  24. [24]

    A survey of crowd counting and density estimation based on convolutional neural network

    Fan, Z., Zhang, H., Zhang, Z., Lu, G., Zhang, Y., Wang, Y., 2022. A survey of crowd counting and density estimation based on convolutional neural network. Neurocomputing 472, 224–251

  25. [25]

    Bacterial colony counting by Convolutional Neural Networks, in: 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), IEEE

    Ferrari, A., Lombardi, S., Signoroni, A., 2015. Bacterial colony counting by Convolutional Neural Networks, in: 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), IEEE. pp. 7458–7461

  26. [26]

    Bacterial colony counting with Convolutional Neural Networks in Digital Microbiology Imaging

    Ferrari, A., Lombardi, S., Signoroni, A., 2017. Bacterial colony counting with Convolutional Neural Networks in Digital Microbiology Imaging. Pattern Recognition 61, 629–640

  27. [27]

    Learning to Count Cells: Applications to lens-free imaging of large fields , 1–6

    Flaccavento, G., Lempitsky, V., Pope, I., Barber, P., Zisserman, A., Noble, J., Vojnovic, B., 2011. Learning to Count Cells: Applications to lens-free imaging of large fields , 1–6

  28. [28]

    OpenCFU, a New Free and Open-Source Software to Count Cell Colonies and Other Circular Objects

    Geissmann, Q., 2013. OpenCFU, a New Free and Open-Source Software to Count Cell Colonies and Other Circular Objects. PLoS ONE 8, 1–10

  29. [29]

    Fast r-cnn

    Girshick, R., 2015. Fast R-CNN. Proceedings of the IEEE In- ternational Conference on Computer Vision 2015 Inter, 1440–1448. arXiv:1504.08083

  30. [30]

    Rich feature hierarchies for accurate object detection and semantic segmentation

    Girshick, R., Donahue, J., Darrell, T., Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition , 580–587arXiv:1311.2524. 51

  31. [31]

    Generative adversarial networks

    Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y., 2020. Generative adversarial networks. Communications of the ACM 63, 139–144

  32. [32]

    Self- normalized density map (SNDM) for counting microbiological objects

    Graczyk, K.M., Pawlowski, J., Majchrowska, S., Golan, T., 2022. Self- normalized density map (SNDM) for counting microbiological objects. Scientific Reports 12, 10583

  33. [33]

    The Elements of Statis- tical Learning: Data Mining, Inference, and Prediction

    Hastie, T., Tibshirani, R., Friedman, J., 2009. The Elements of Statis- tical Learning: Data Mining, Inference, and Prediction. Second edi ed., Springer New York, New York

  34. [34]

    [He et al

    He, K., Gkioxari, G., Dollar, P., Girshick, R., 2017. Mask R-CNN. Proceedings of the IEEE International Conference on Computer Vision 2017-Octob, 2980–2988.arXiv:1703.06870

  35. [35]

    Deep Residual Learning for Image Recognition

    He, K., Zhang, X., Ren, S., Sun, J., 2015. Deep Residual Learning for ImageRecognition. 2016IEEEConferenceonComputerVisionandPat- tern Recognition (CVPR) 2016-Decem, 770–778.arXiv:1512.03385

  36. [36]

    FastER: A User- Friendly tool for ultrafast and robust cell segmentation in large-scale microscopy

    Hilsenbeck, O., Schwarzfischer, M., Loeffler, Di., DImopoulos, S., Has- treiter, S., Marr, C., Theis, F.J., Schroeder, T., 2017. FastER: A User- Friendly tool for ultrafast and robust cell segmentation in large-scale microscopy. Bioinformatics 33, 2020–2028

  37. [37]

    Hu, Y., Jiang, X., Liu, X., Zhang, B., Han, J., Cao, X., Doermann, D., 2020. NAS-Count: Counting-by-Density with Neural Architecture Search, in: Computer Vision – ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII, Springer- Verlag, Berlin, Heidelberg. pp. 747–766

  38. [38]

    Body Structure Aware Deep Crowd Counting

    Huang, S., Li, X., Zhang, Z., Wu, F., Gao, S., Ji, R., Han, J., 2018. Body Structure Aware Deep Crowd Counting. IEEE Transactions on Image Processing 27, 1049–1059

  39. [39]

    Ioffe, S., Szegedy, C., 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift, in: Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37, JMLR.org, Lille, France. pp. 448–456. 52

  40. [40]

    Bac- terial Colonies in Solid Media and Foods: A Review on Their Growth and Interactions with the Micro-Environment

    Jeanson, S., Floury, J., Gagnaire, V., Lortal, S., Thierry, A., 2015. Bac- terial Colonies in Solid Media and Foods: A Review on Their Growth and Interactions with the Micro-Environment. Frontiers in Microbiology 6

  41. [41]

    doi:10.5281/zenodo.7347926 , url =

    Jocher, G., Chaurasia, A., Stoken, A., Borovec, J., NanoCode012, Kwon, Y., Michael, K., TaoXie, Fang, J., Imyhxy, 2022. ultr- alytics/yolov5: v7.0 - yolov5 sota realtime instance segmentation. doi:10.5281/zenodo.7347926

  42. [42]

    Arraycount, an algorithm for automatic cell counting in microwell arrays

    Kachouie, N.N., Kang, L., Khademhosseini, A., 2009. Arraycount, an algorithm for automatic cell counting in microwell arrays. BioTechniques 47, x–xvi

  43. [43]

    AutoCellSeg: Ro- bust automatic colony forming unit (CFU)/cell analysis using adaptive image segmentation and easy-to-use post-editing techniques

    Khan, A.U.M., Torelli, A., Wolf, I., Gretz, N., 2018. AutoCellSeg: Ro- bust automatic colony forming unit (CFU)/cell analysis using adaptive image segmentation and easy-to-use post-editing techniques. Scientific Reports 8, 7302

  44. [44]

    Adam: A Method for Stochastic Optimization

    Kingma, D.P., Ba, J., 2014. Adam: A Method for Stochastic Optimiza- tion. 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings , 1–15arXiv:1412.6980

  45. [45]

    ImageNet classifica- tion with deep convolutional neural networks, in: NIPS’12: Proceedings of the 25th International Conference on Neural Information Processing Systems, pp

    Krizhevsky, A., Sutskever, I., Hinton, G.E., 2012. ImageNet classifica- tion with deep convolutional neural networks, in: NIPS’12: Proceedings of the 25th International Conference on Neural Information Processing Systems, pp. 1097–1105

  46. [46]

    Counting Cows: Tracking Illegal Cattle Ranching From High-Resolution Satellite Imagery.arXiv:2011.07369

    Laradji, I., Rodriguez, P., Kalaitzis, F., Vazquez, D., Young, R., Davey, E., Lacoste, A., 2020. Counting Cows: Tracking Illegal Cattle Ranching From High-Resolution Satellite Imagery.arXiv:2011.07369

  47. [47]

    Lewis, M., Yarats, D., Dauphin, Y., Parikh, D., Batra, D., 2017. Deal or No Deal? End-to-End Learning of Negotiation Dialogues, in: Pro- ceedings of the 2017 Conference on Empirical Methods in Natural Lan- guage Processing, Association for Computational Linguistics, Copen- hagen, Denmark. pp. 2443–2453

  48. [48]

    Yolov6: A single-stage object detection framework for industrial applications

    Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., Ke, Z., Li, Q., Cheng, M., Nie, W., Li, Y., Zhang, B., Liang, Y., Zhou, L., Xu, X., 53 Chu, X., Wei, X., Wei, X., 2022. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications.arXiv:2209.02976

  49. [49]

    High temporalfrequencyvehiclecountingfromlow-resolutionsatelliteimages

    Liao, L., Xiao, J., Yang, Y., Ma, X., Wang, Z., Satoh, S., 2023. High temporalfrequencyvehiclecountingfromlow-resolutionsatelliteimages. ISPRS Journal of Photogrammetry and Remote Sensing 198, 45–59

  50. [50]

    A Novel Cell Detection Method Using Deep Convolutional Neural Network and Maximum-Weight Independent Set, in: AdvancesinComputerVisionandPatternRecognition.volume9351, pp

    Liu, F., Yang, L., 2017. A Novel Cell Detection Method Using Deep Convolutional Neural Network and Maximum-Weight Independent Set, in: AdvancesinComputerVisionandPatternRecognition.volume9351, pp. 63–72

  51. [51]

    Liu, N., Long, Y., Zou, C., Niu, Q., Pan, L., Wu, H., 2019. AD- CrowdNet: An Attention-Injective Deformable Convolutional Network forCrowdUnderstanding, in: 2019IEEE/CVFConferenceonComputer Vision and Pattern Recognition (CVPR), IEEE, Long Beach, CA, USA. pp. 3220–3229

  52. [52]

    An image analysis-based approach for automated counting of cancer cell nuclei in tissue sections

    Loukas, C.G., Wilson, G.D., Vojnovic, B., Linney, A., 2003. An image analysis-based approach for automated counting of cancer cell nuclei in tissue sections. Cytometry 55A, 30–42

  53. [53]

    Majchrowska, S., Pawlowski, J., Gula, G., Bonus, T., Hanas, A., Loch, A., Pawlak, A., Roszkowiak, J., Golan, T., Drulis-Kawa, Z.,

  54. [54]

    arXiv:2108.01234 [cs, q-bio]arXiv:2108.01234

    AGAR a microbial colony dataset for deep learning detection. arXiv:2108.01234 [cs, q-bio]arXiv:2108.01234

  55. [55]

    As- sessing microbial colony counting: A deep learning approach with the AGAR image dataset

    Majchrowska, S., Pawłowski, J., Guła, G., Bonus, T., Hanas, A., Loch, A., Pawlak, A., Roszkowiak, J., Golan, T., Drulis-Kawa, Z., 2025. As- sessing microbial colony counting: A deep learning approach with the AGAR image dataset. Neurocomputing 630, 129654

  56. [56]

    YOLO-Based Deep Learning Frame- work for Olive Fruit Fly Detection and Counting

    Mamdouh, N., Khattab, A., 2021. YOLO-Based Deep Learning Frame- work for Olive Fruit Fly Detection and Counting. IEEE Access 9, 84252– 84262

  57. [57]

    Marstaller, J., Tausch, F., Stock, S., 2019. DeepBees - Building and Scaling Convolutional Neuronal Nets For Fast and Large-Scale Visual Monitoring of Bee Hives, in: 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 271–278. 54

  58. [58]

    Semi-automatic prototype system for bacterial colony counting, in: 2016 International Conference on Smart Systems and Technologies (SST), IEEE

    Matic, T., Vidovic, I., Siladi, E., Tkalec, F., 2016. Semi-automatic prototype system for bacterial colony counting, in: 2016 International Conference on Smart Systems and Technologies (SST), IEEE. pp. 205– 210

  59. [59]

    Detection and Counting of Corn Plants in the Presence of Weeds with Convolutional Neural Net- works

    Mota-Delfin, C., López-Canteñs, G.d.J., López-Cruz, I.L., Romantchik- Kriuchkova, E., Olguín-Rojas, J.C., 2022. Detection and Counting of Corn Plants in the Presence of Weeds with Convolutional Neural Net- works. Remote Sensing 14, 4892

  60. [60]

    Counting colonies of clonogenic assays by using densitometric software

    Niyazi, M., Niyazi, I., Belka, C., 2007. Counting colonies of clonogenic assays by using densitometric software. Radiation Oncology 2, 3–5

  61. [61]

    Dense Crowd Count- ing Convolutional Neural Networks with Minimal Data using Semi- Supervised Dual-Goal Generative Adversarial Networks

    Olmschenk, G., Chen, J., Tang, H., Zhu, Z., 2019. Dense Crowd Count- ing Convolutional Neural Networks with Minimal Data using Semi- Supervised Dual-Goal Generative Adversarial Networks. IEEE Con- ference on Computer Vision and Pattern Recognition: Learning with Imperfect Data Workshop

  62. [62]

    U2-Net: Going deeper with nested U-structure for salient object detection

    Qin, X., Zhang, Z., Huang, C., Dehghan, M., Zaiane, O.R., Jagersand, M., 2020. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognition 106, 107404

  63. [63]

    Extended-maxima trans- form watershed segmentation algorithm for touching corn kernels

    Qin, Y., Wang, W., Liu, W., Yuan, N., 2013. Extended-maxima trans- form watershed segmentation algorithm for touching corn kernels. Ad- vances in Mechanical Engineering 2013, 268046

  64. [64]

    Learning To Count Everything, in: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp

    Ranjan, V., Sharma, U., Nguyen, T., Hoai, M., 2021. Learning To Count Everything, in: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3393–3402

  65. [65]

    You Only Look Once: Unified, Real-Time Object Detection, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE

    Redmon, J., Divvala, S., Girshick, R., Farhadi, A., 2016. You Only Look Once: Unified, Real-Time Object Detection, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE. pp. 779– 788

  66. [66]

    Yolo9000: Better, faster, stronger.CoRR, abs/1612.08242, 2016

    Redmon, J., Farhadi, A., 2016. YOLO9000: Better, Faster, Stronger. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017-Janua, 6517–6525.arXiv:1612.08242. 55

  67. [67]

    YOLOv3: An Incremental Improvement

    Redmon, J., Farhadi, A., 2018. YOLOv3: An Incremental Improvement arXiv:1804.02767

  68. [68]

    Girshick, and Jian Sun

    Ren, S., He, K., Girshick, R., Sun, J., 2017. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence 39, 1137– 1149.arXiv:1506.01497

  69. [69]

    Rodriguez, I.F., Megret, R., Acuna, E., Agosto-Rivera, J.L., Giray, T.,

  70. [70]

    Recognition of Pollen-Bearing Bees from Video Using Convolu- tional Neural Network, in: 2018 IEEE Winter Conference on Applica- tions of Computer Vision (WACV), pp. 314–322

  71. [71]

    U-Net: Convolutional Networks for Biomedical Image Segmentation

    Ronneberger, O., Fischer, P., Brox, T., 2015. U-net: Convolutional networksforbiomedicalimagesegmentation. LectureNotesinComputer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 9351, 234–241.arXiv:1505.04597

  72. [72]

    Au- tomated Training of Deep Convolutional Neural Networks for Cell Seg- mentation

    Sadanandan, S.K., Ranefall, P., Le Guyader, S., Wählby, C., 2017. Au- tomated Training of Deep Convolutional Neural Networks for Cell Seg- mentation. Scientific Reports 7, 1–7

  73. [73]

    Generative Adversarial Networks (GANs): Challenges, Solutions, and Future Directions

    Saxena, D., Cao, J., 2021. Generative Adversarial Networks (GANs): Challenges, Solutions, and Future Directions. ACM Computing Surveys 54, 63:1–63:42

  74. [74]

    A Standard Driven Software Architecture for Fully Autonomous Vehicles, in: 2018 IEEE Interna- tional Conference on Software Architecture Companion (ICSA-C), pp

    Serban, A.C., Poll, E., Visser, J., 2018. A Standard Driven Software Architecture for Fully Autonomous Vehicles, in: 2018 IEEE Interna- tional Conference on Software Architecture Companion (ICSA-C), pp. 120–127

  75. [75]

    Active learning strategies for phenotypic profiling of high-content screens

    Smith, K., Horvath, P., 2014. Active learning strategies for phenotypic profiling of high-content screens. Journal of Biomolecular Screening 19, 685–695

  76. [76]

    EfficientDet: Scalable and Efficient Object Detection, in: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp

    Tan, M., Pang, R., Le, Q.V., 2020. EfficientDet: Scalable and Efficient Object Detection, in: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10778–10787

  77. [77]

    Attention is all you need, in: 56 Proceedings of the 31st International Conference on Neural Information Processing Systems, Curran Associates Inc., Red Hook, NY, USA

    Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I., 2017. Attention is all you need, in: 56 Proceedings of the 31st International Conference on Neural Information Processing Systems, Curran Associates Inc., Red Hook, NY, USA. pp. 6000–6010

  78. [78]

    Adaptive Density Map Generation for Crowd Counting, in: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), IEEE, Seoul, Korea (South)

    Wan, J., Chan, A., 2019. Adaptive Density Map Generation for Crowd Counting, in: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), IEEE, Seoul, Korea (South). pp. 1130–1139

  79. [79]

    YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors

    Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M., 2022. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv:2207.02696

  80. [80]

    A Convolutional Neural Network-Based Method for Corn Stand Counting in the Field

    Wang, L., Xiang, L., Tang, L., Jiang, H., 2021. A Convolutional Neural Network-Based Method for Corn Stand Counting in the Field. Sensors 21, 507

Showing first 80 references.