Generative AI methods for synthesis of image data to train AI for automated scene understanding in a military context: a review of opportunities

Open Access
Authors
  • E.P. Fokkinga
  • T.A. Eker
  • J.E. van Woerden
  • J.-M. Witon
  • S.O.B. Stallinga
  • A. Visser ORCID logo
  • K. Schutte
  • F.G. Heslinga
Publication date 2025
Host editors
  • K.E. Manser
  • C.L. Howell
  • R.M. Rao
  • C. De Melo
  • K.F. Prussing
Book title Synthetic Data for Artificial Intelligence and Machine Learning: Tools, Techniques, and Applications III
Book subtitle 14-17 April 2025, Orlando, Florida, United States
ISBN
  • 9781510687073
ISBN (electronic)
  • 9781510687080
Series Proceedings of SPIE
Event Synthetic Data for Artificial Intelligence and Machine Learning: Tools, Techniques, and Applications
Article number 13459 05
Number of pages 23
Publisher Bellingham, Washington: SPIE
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
The rapid increase in sensors on manned and unmanned military platforms has led to a significant rise in image data (e.g., visible, infrared, sonar, radar), enabling extensive scene analysis. Thorough and real-time understanding of these scenes requires automated image analysis tools, for e.g. object detection, traversability analysis, and threat classification. However, the development of artificial intelligence (AI) models for automated scene understanding is constrained by limited access to relevant military training data due to its restricted nature, high acquisition costs, and evolving threat signatures. Several studies highlight the potential of synthetic data as an alternative to measured training data, for example by utilizing physics-based modeling of scenes and objects of interest.

Recent advances in generative AI (GenAI), particularly in diffusion-based models, offer opportunities to synthesize data with variations beyond what was previously possible, improving performance in various nonmilitary image analysis tasks. Despite this, the lack of military-relevant data used for GenAI model development suggests that non-specialized models may produce military scenes with limited quality and variation. In this review, we explore the opportunities of state-of-the-art GenAI methods for creating high-quality training data for military AI systems. We identify three key strategies: (1) full-image generation by fine-tuning with application-specific data; (2) inpainting, where objects of interest can be placed in existing image data; and (3) image-to-image translation which is used to augment image conditions or translate between image modalities. Visual results of each of these methods are promising. Some studies have already shown benefits of these data synthesis methods as data augmentation to improve downstream AI models. Further research shall determine the value for operationalization in a wide set of use-cases.
Document type Conference contribution
Language English
Published at https://doi.org/10.1117/12.3053494
Downloads
Permalink to this page
Back