The Joint Force’s ability to integrate data and systems at speed and scale is of utmost importance to maintain a competitive edge in the data-driven arena of 21st century warfare. The current methods employed by developers to find system integration points, however, are ad-hoc, predominantly manual, and too costly – consider the explosion in volume and heterogeneity of the underlying databases that support modern-day operations. In this work, we present a data archeology and alignment workbench, known as Semantic Lifting of Integrated Messages (SLIM), that allows developers to quickly discover and document commonalities among disparate data. Our novel workbench uses background knowledge to constrain the geometric orientation of unfamiliar data in ways that are meaningful to developers. To accomplish this, SLIM uses a variety of Machine Learning algorithms to unify and characterize the meaning of data, and a dashboard of linked visualizations to convey semantic commonalities from different perspectives and dimensions. Additionally, the workbench maintains a web of execution provenance that developers can use to trace and probe results in order to assess relevancy and meaningfulness. As a demonstration, we describe two studies that use our process to quickly find non-trivial commonalities among multiple datasets, including the Universal Command and Control Interface (UCI) and an ontology about the arrival and departure of vehicles.
DevOps engineers take many factors into account when assessing the suitability of AI/ML algorithms for operations. They rely on documentation about data requirements, parameter settings, theoretical limits, and operating characteristics to anticipate how well an algorithm will perform in the field prior to writing a single line of code. However, the USAF currently lacks code quality and documentation standards for AI/ML, which forces downstream consumers to make assumptions about the behavior of algorithms that often lead to unexpected results and wasted efforts. Therefore, we present a preliminary set of criteria by which to judge the maturity of word embedding algorithms in terms of reproducibility, testability, and documentation quality. We hope the criteria will grow into a set of quality requirements that govern how the DoD procures AI/ML capabilities, and eventually motivate the need for an ML development maturity model. Our nascent evaluation criteria surfaced during our own struggles with trying to replicate and compare the performance between a well-known word embedding algorithm (Word2Vec) and a custom graph embedding algorithm (IRI2Vec), procured specifically for the purpose of connecting missions across ATOs. We walk through two case studies in which we applied our evaluation criteria to assess the maturity of both Word2Vec and IRI2Vec
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.