<?xml version="1.0" encoding="UTF-8"?>
<rdf:RDF xmlns="http://purl.org/rss/1.0/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:dc="http://purl.org/dc/elements/1.1/">
<channel rdf:about="https://hdl.handle.net/1721.1/145263">
<title>AIA</title>
<link>https://hdl.handle.net/1721.1/145263</link>
<description/>
<items>
<rdf:Seq>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165294"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165232"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165231"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165230"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165229"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/165228"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164903"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164902"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164901"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164900"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164899"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164898"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164897"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164896"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/164868"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162653"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162634"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162632"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162631"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162630"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162629"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162628"/>
<rdf:li rdf:resource="https://hdl.handle.net/1721.1/162627"/>
</rdf:Seq>
</items>
<dc:date>2026-04-03T20:57:08Z</dc:date>
</channel>
<item rdf:about="https://hdl.handle.net/1721.1/165294">
<title>Synthetic Network Data Generation for Analyst Training</title>
<link>https://hdl.handle.net/1721.1/165294</link>
<description>Synthetic Network Data Generation for Analyst Training
Smith, Liam; Wright, Matthew
Rapidly evolving cyber threats demand continuous,&#13;
high-fidelity training for defense analysts. However, generating&#13;
realistic network traffic datasets creates a significant barrier&#13;
to entry, often requiring extensive virtualization infrastructure,&#13;
specialized hardware, and knowledge in cyber range administration.&#13;
This paper introduces a streamlined architecture, called&#13;
Generative Packet Captures (GenCap), built upon the foundational&#13;
capabilities of the FOSR benign traffic generator and&#13;
the ID2T attack injector. By abstracting these complex tools&#13;
behind an automated orchestration layer, it enables users to&#13;
generate scenario-specific PCAP files on demand. This approach&#13;
democratizes access to training data, allowing analysts to create&#13;
rigorous network defense scenarios without the need for complex&#13;
provisioning or systems engineering knowledge.
</description>
<dc:date>2026-04-01T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/165232">
<title>Office AI Automation using Existing DAF-Approved Software</title>
<link>https://hdl.handle.net/1721.1/165232</link>
<description>Office AI Automation using Existing DAF-Approved Software
Cui, Wei; Kennedy, Laura
The Department of the Air Force (DAF) continues&#13;
to face mounting administrative workloads that hinder mission&#13;
focus and operational efficiency. Executive officers and staff&#13;
officers spend substantial time generating reports, managing&#13;
emails, routing documents, and organizing taskers across multiple&#13;
systems. This paper presents the Smart Executive Assistant, an&#13;
office AI initiative to automate repetitive administrative tasks using&#13;
existing DAF-approved technologies without a new Authority-&#13;
To-Operate (ATO). By integrating DAF 365 applications, Power&#13;
Automate, and approved large language models (LLMs) within&#13;
secure IL5 and IL6 environments, this solution seeks to reduce&#13;
time spent on low-value administrative processes by 90% while&#13;
maintaining compliance and data security.
</description>
<dc:date>2026-03-20T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/165231">
<title>AI for Scalable Defensive Cyber Log Analysis</title>
<link>https://hdl.handle.net/1721.1/165231</link>
<description>AI for Scalable Defensive Cyber Log Analysis
Schofield, Catherine; Jananthan, Hayden; Kepner, Jeremy
Centralized cyber logging platforms ingest large&#13;
volumes of heterogeneous telemetry, yet high dimensionality&#13;
and query-driven workflows often limit scalable analytic insight&#13;
on these systems. This work presents an automated pipeline&#13;
for ingesting, characterizing, and analyzing large-scale hostbased&#13;
logs using sparse representations and distribution-aware&#13;
statistics. A systematic dimensional analysis reduces hundreds of&#13;
raw log fields to a small set of informative dimensions suitable&#13;
for aggregation across extended time windows. Temporal analysis&#13;
of the reduced representation reveals coordinated deviations&#13;
in activity volume and distributional behavior that are not&#13;
apparent in individual log streams. The results demonstrate that&#13;
dimensional reduction enables scalable, interpretable analysis&#13;
of enterprise cyber telemetry. Furthermore, these results were&#13;
obtained using host-based sensors designed for event-oriented&#13;
point-defense and demonstrate the feasibility of integrating such&#13;
sensors to enable long-range, long-duration area defense.
</description>
<dc:date>2026-03-20T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/165230">
<title>Cross-Aircraft Flight Phase Classification Using ADS-B Data and Transfer Learning</title>
<link>https://hdl.handle.net/1721.1/165230</link>
<description>Cross-Aircraft Flight Phase Classification Using ADS-B Data and Transfer Learning
Kiefer, Jacob; Alemany, Sheila
Flight phase identification (FPI) approaches that&#13;
apply traditional machine learning techniques are expensive to&#13;
scale, difficult to generalize across platforms, and frequently&#13;
unavailable in permissive or distributed training environments.&#13;
We propose a scalable, data-driven pipeline for automatic FPI&#13;
using open-source Automatic Dependent Surveillance-Broadcast&#13;
(ADS-B) data, with an emphasis on cross-aircraft generalization&#13;
through transfer learning. Leveraging ADS-B telemetry from&#13;
USAF Initial Flight Training aircraft, a neural network classifier&#13;
is trained on Diamond DA-20 flight data and evaluated on Texan&#13;
T-6 aircraft under zero-shot and fine-tuned transfer learning&#13;
conditions. We describe a robust ADS-B preprocessing pipeline&#13;
integrating digital elevation model (DEM) data, a data labeling&#13;
strategy using unsupervised learning, and a transfer learning&#13;
approach enabling adaptation across aircraft types with limited&#13;
labeled data. Our results demonstrate that transfer learning significantly&#13;
improves classification accuracy for flight phases with&#13;
limited data, highlighting the potential of ADS-B-based models&#13;
to support scalable, behavior-aware airspace intelligence across&#13;
heterogeneous fleets and permissive environments. This research&#13;
advances FPI capabilities for USAF training analysis and broader&#13;
operational priorities in autonomy, situational awareness, and&#13;
data-driven decision support.
</description>
<dc:date>2026-03-20T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/165229">
<title>Neural Networks for Stress Intensity Factor Vertex Prediction</title>
<link>https://hdl.handle.net/1721.1/165229</link>
<description>Neural Networks for Stress Intensity Factor Vertex Prediction
Hokaj, Ian; Ghanem, Janelle
Structural fatigue in aging metallic aircraft is a&#13;
primary driver of sustainment costs for the U.S. Air Force,&#13;
significantly impacting fleet readiness. Fatigue life prediction tools&#13;
like AFGROW depend on interpolating between computationally&#13;
expensive stress intensity factors (K-solutions) to approximate&#13;
unknown values. However, interpolation errors in the current&#13;
approach introduce uncertainty and force overly conservative&#13;
maintenance schedules. This paper investigates the use of a&#13;
machine learning surrogate to replace AFGROW’s dimensionreduction&#13;
interpolation for the finite-width c orner-cracked hole&#13;
geometry. We developed a robust data processing pipeline for a&#13;
large FEA dataset and trained a neural network model.&#13;
Our results reveal a critical insight: the surrogate model offers&#13;
substantial performance gains over AFGROW’s interpolation&#13;
in low-data regimes, emphasizing both the potential of the&#13;
model and its sensitivity to dataset size. For the original, sparse&#13;
dataset—which is characteristic of computationally expensive&#13;
problems—the neural network significantly o utperformed the&#13;
baseline interpolation, reducing the mean absolute percentage&#13;
error (MAPE) by over 40% (from 2.77% to 1.60%) and achieving&#13;
an R² value exceeding 0.99. However, experiments on synthetically&#13;
generated dense datasets showed that the traditional interpolation&#13;
method becomes more accurate as the data grid becomes less&#13;
sparse.&#13;
This study concludes that while neural network surrogates&#13;
offer a powerful, high-fidelity solution for computationally intensive&#13;
engineering problems, their adoption should be guided by&#13;
a careful analysis of data density after dataset has been cleaned&#13;
of outliers. It also highlights the necessity of employing rigorous,&#13;
application-relevant validation strategies that move beyond&#13;
simplistic random splits to accurately assess model performance.
</description>
<dc:date>2026-03-20T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/165228">
<title>Evaluating Adaptive AI for Contracting Officer Readiness: Design and Pedagogical Proposal for the Warrant Board RAG Chatbot</title>
<link>https://hdl.handle.net/1721.1/165228</link>
<description>Evaluating Adaptive AI for Contracting Officer Readiness: Design and Pedagogical Proposal for the Warrant Board RAG Chatbot
Mullen, Julia; Grosvenor, Sarah
The United States Air Force (USAF) requires a&#13;
sustained and expanding pool of warranted Contracting Officers&#13;
(COs) to meet growing operational and fiscal demands across&#13;
its global enterprise. The authority to obligate funds and bind&#13;
the government contractually—granted through the issuance&#13;
of a warrant—requires successful completion of a multi-stage&#13;
evaluation process culminating in a scenario-based oral board.&#13;
This final interview assesses a candidate’s ability to interpret&#13;
and apply acquisition policy under complex and ambiguous&#13;
conditions.&#13;
This paper proposes the design of an adaptive artificial&#13;
intelligence (AI) training system—the Warrant Board Retrieval-&#13;
Augmented Generation (RAG) Chatbot—to serve as a simulated&#13;
board-preparation environment. This chatbot inverts the common&#13;
’user question, AI answer’ model, and instead has the&#13;
chatbot ask the learner a series of critical thinking scenariobased&#13;
questions. The prototype design adopts a model-agnostic&#13;
LLM gateway capable of operation through either commercial&#13;
APIs (e.g., OpenAI) or secure, government-hosted environments&#13;
such as GenAI.mil, ensuring accessibility within unclassified Air&#13;
Force networks. This research contributes to the emerging field of&#13;
AI-assisted professional education by developing a transparent,&#13;
auditable, and pedagogically grounded framework for formative&#13;
learning in acquisition training.
</description>
<dc:date>2026-03-20T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164903">
<title>AgentNexus: Accelerating AI Agent Development and Enhancing Interoperability with MCP</title>
<link>https://hdl.handle.net/1721.1/164903</link>
<description>AgentNexus: Accelerating AI Agent Development and Enhancing Interoperability with MCP
Yae, Jung; Hamilton, Lei
The DoD faces significant challenges in its pursuit&#13;
of AI superiority, as disparate data and development platforms&#13;
create redundant efforts and limit interoperability. Additionally,&#13;
existing DoD systems are ill-equipped to handle the recent&#13;
paradigm shift toward agentic AI, which requires modern standards&#13;
and tools. To address these gaps, this paper introduces&#13;
AgentNexus, an application designed to streamline the development,&#13;
deployment, and servicing of AI agents. AgentNexus&#13;
provides an application featuring an advanced agents processing&#13;
backend, a scalable service layer, and an intuitive user interface.&#13;
It provides pre-built toolkits, sophisticated RAG pipeline, and&#13;
MCP for enhanced interoperability. The successful development&#13;
of an Education Assistant agent validates the application’s capacity&#13;
to support the rapid implementation of multi-agent workflows.&#13;
By fostering a collaborative and standardized environment,&#13;
AgentNexus mitigates critical barriers of interoperability and&#13;
duplicated effort, accelerating the delivery of multi-agent AI to&#13;
warfighters.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164902">
<title>Intelligent C-17 Load Planning for Flight Optimization</title>
<link>https://hdl.handle.net/1721.1/164902</link>
<description>Intelligent C-17 Load Planning for Flight Optimization
McAlister, Catherine; Jones, Mathew; McConville, Sean
C-17 Globemaster III cargo capacity is significantly&#13;
underutilized, with many sorties transporting only a few pallets&#13;
despite the aircraft’s 170,900-pound payload capability. Historical&#13;
flight data analysis reveals inefficient scheduling practices that&#13;
increase operational costs, crew workload, and overall negatively&#13;
effect mission capability. This paper details the development&#13;
of an AI-powered optimization model to improve C-17 cargo&#13;
utilization and reduce required flight operations. We analyzed&#13;
historical C-17 transportation data and created both traditional&#13;
optimization algorithms and predictive AI models to determine&#13;
optimal flight scheduling for 3-week operational periods. The AI&#13;
model achieved 97.9% accuracy in predicting optimal flight count&#13;
requirements and 89.3% accuracy in predicting optimal flight&#13;
assignment for specific cargo, representing a 23% reduction in&#13;
total flights and a 15% increase in average cargo utilization.&#13;
These results demonstrate that data-driven flight scheduling&#13;
can significantly improve C-17 operational efficiency, reduce&#13;
costs across the airlift community, and enabling additional time&#13;
towards advanced training, contingency support, and critical&#13;
warfighter operations, ultimately increasing the lethality and&#13;
readiness of the Department of Defense.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164901">
<title>Securing Intelligence: The Strategic Necessity of Air-Gapped AI Systems in the Age of Cloud-Based LLMs</title>
<link>https://hdl.handle.net/1721.1/164901</link>
<description>Securing Intelligence: The Strategic Necessity of Air-Gapped AI Systems in the Age of Cloud-Based LLMs
Viggh, Herbert; Tsagaratos, Jennifer
The increasing use of large language models (LLMs)&#13;
in applications, from military strategy to customer service, raises&#13;
concerns about data sovereignty, security, and privacy. Cloudbased&#13;
API models, created by companies such as OpenAI, pose&#13;
significant risks due to training data exposure and prompt&#13;
injection attacks, which can compromise sensitive information&#13;
and hidden biases that could influence reporting or executive&#13;
decision-making processes. Real-world incidents, such as the&#13;
leakage of Samsung’s proprietary source code through ChatGPT,&#13;
highlight the dangers of relying on cloud providers with complete&#13;
visibility into client queries. Furthermore, data localization laws&#13;
and regulations, such as the General Data Protection Regulation&#13;
(GDPR), underscore the risks associated with outsourcing&#13;
intelligence and decision support systems to foreign entities. Airgapped&#13;
AI solutions, which run on isolated networks disconnected&#13;
from the outside world, offer a secure alternative for sensitive&#13;
environments such as national defense, research laboratories,&#13;
and critical infrastructure. By maintaining control over AI&#13;
processes, organizations can ensure information safety, comply&#13;
with regulations, and mitigate risks associated with cloud-based&#13;
AI infrastructure, ultimately safeguarding their data integrity,&#13;
privacy, and operational independence.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164900">
<title>RAIMOND Requirements AI for Military Operational Needs Development</title>
<link>https://hdl.handle.net/1721.1/164900</link>
<description>RAIMOND Requirements AI for Military Operational Needs Development
Garcia, Fabio; Steilberg, Jackson
The Joint Capabilities Integration and Development&#13;
System (JCIDS) was created as a means to overhaul military&#13;
procurement processes. Ideally, the requirements development&#13;
process is meant to take a total of 2-4 years from concept&#13;
to manufacturing. However the actual length of concept development&#13;
is much longer. As a result, technologies that are&#13;
conceptualized through the analytical process often enter the&#13;
acquisition too late to need for the warrior. To reduce the&#13;
lengthy timeline in requirements development, we used Large&#13;
Language Models (LLMs) to conduct the necessary research&#13;
and synthesize documents that abide by strict JCIDS guidelines.&#13;
Prompt engineering can achieve these results as a proof of&#13;
concept. However, the output responses lack the content length&#13;
and depth necessary to pass through the requirements validation&#13;
process. Therefore, a combination of agentic workflows, prompt&#13;
engineering, and sufficient context is needed to achieve the desired&#13;
outcomes. This project utilizes a novel framework to derive&#13;
Capabilities Based Assessments (CBAs) at an approximate 80&#13;
percent readiness level requiring the final steps of validation and&#13;
verification by subject matter experts.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164899">
<title>Machine Learning for the Enhancement of Adaptive Optics</title>
<link>https://hdl.handle.net/1721.1/164899</link>
<description>Machine Learning for the Enhancement of Adaptive Optics
Hall, Robert; Chen, Justin
Optical systems (telescopes, lasers, microscopes,&#13;
etc.) have degraded performance over long distances&#13;
due to scintillation caused by Earth’s atmosphere,&#13;
where adaptive optics (AO) is often used to enhance&#13;
its signal-to-noise (SNR) ratio or image quality. Astronomers&#13;
have found success in laser-based adaptive&#13;
optics where they survey the atmosphere with a laser&#13;
and subtract its effects on the resultant image. Although&#13;
effective in most cases, these systems can be extremely&#13;
costly, are computationally intensive in real time, and&#13;
fall short in some edge cases. We propose an autoencoder/&#13;
decoder and a generalized sequence to sequence&#13;
model (LSTM) as a cost-effective method to off-load&#13;
computational complexity from real time and enhance&#13;
performance in edge cases. This study utilizes four&#13;
simulated datasets of wavefront sensor frames for a&#13;
variety of atmospheric conditions, done in collaboration&#13;
with MIT Lincoln Laboratory [1]–found auto-encoding&#13;
performance just shy of traditional methodology and&#13;
found LSTM performance that predicts well the general&#13;
shape on the WFS, but suffers from scaling issues.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164898">
<title>From Hype to Reality: Real-World Lessons and Recommendations for AI in Military Applications</title>
<link>https://hdl.handle.net/1721.1/164898</link>
<description>From Hype to Reality: Real-World Lessons and Recommendations for AI in Military Applications
Lynch, Joshua; Niss, Laura
The current use cases, limitations, and future capacity&#13;
of large language models (LLMs) as assistants to military&#13;
personnel remain an open question. This paper presents a case&#13;
study of an Airman’s interaction with and trust calibration of&#13;
LLMs over three months, both as an everyday assistant and&#13;
for development of ROMAD-AI, a tactical military application.&#13;
Through intuitive, AI-generated software development, an approach&#13;
relying on iterative code generation through natural&#13;
language prompting of LLMs from a technical novice rather&#13;
than human generated programming from a technical expert,&#13;
the research reveals significant gaps between industry curated&#13;
AI capability demonstrations and operational reality, requiring&#13;
systematic trust calibration and realistic scope management.&#13;
Outcomes are analyzed through operational and technical expertise&#13;
perspectives to provide practical guidance for both military&#13;
service members seeking effective AI integration and researchers&#13;
developing military-focused AI systems.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164897">
<title>Large Language Models and Defense Strategy: Escalation Risks and National Security Challenges</title>
<link>https://hdl.handle.net/1721.1/164897</link>
<description>Large Language Models and Defense Strategy: Escalation Risks and National Security Challenges
Hou, Jonathan; Lax, Edwin
This literature review examines the strategic vulnerabilities&#13;
posed by Large Language Models (LLMs) in military&#13;
and national security contexts. It synthesizes recent research&#13;
on their propensity for escalatory reasoning, cultural misalignment,&#13;
semantic manipulation, and dual-use ambiguity. Findings&#13;
from conflict s imulations a nd c oalition p lanning m odels reveal&#13;
how LLMs may default to aggressive or biased outputs under&#13;
ambiguity. These tendencies threaten alliance cohesion, distort&#13;
decision-making, and undermine trust in AI-enabled operations.&#13;
The review concludes by advocating for safeguards such as culturally&#13;
calibrated training, rigorous output verification, a nd the&#13;
integration of human-AI intermediaries to prevent destabilizing&#13;
outcomes.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164896">
<title>Synchronization-Aware Diffusion Models for Intra-Family RF Signal Classification</title>
<link>https://hdl.handle.net/1721.1/164896</link>
<description>Synchronization-Aware Diffusion Models for Intra-Family RF Signal Classification
Hayden, Hunter; Botero, Joey
Classification of radio frequency (RF) signals in the&#13;
presence of channel-induced synchronization errors remains a&#13;
critical challenge in spectrum awareness systems. Traditional&#13;
classification pipelines generally rely on fixed synchronization&#13;
algorithms or assume aligned signals, which limits robustness&#13;
under real world timing, phase, and frequency distortions.&#13;
We introduce SyncDiff, a novel encoder-only diffusion model&#13;
architecture that predicts synchronization parameters through&#13;
iterative denoising steps prior to classification. By replacing&#13;
conventional synchronization algorithms with a learned datadriven&#13;
correction mechanism, our approach enables adaptive&#13;
signal alignment based on current channel distortions in unsynchronized&#13;
input data. SyncDiff employs a UNet based encoder&#13;
to refine synchronization parameters across multiple inference&#13;
steps, dynamically reducing channel-induced alignment errors&#13;
while preserving the inherit modulation specific characteristics&#13;
that allow these signals to be discriminable. Evaluations of the&#13;
RadioML2018 RF standard benchmark data set [1] demonstrates&#13;
improved classification accuracy across varying SNRs, modulation&#13;
schemes and synchronization impairments. Our findings&#13;
highlight the potential of diffusion-based synchronization learning&#13;
to improve downstream RF classification without reliance on&#13;
expert-engineered synchronization routines.
</description>
<dc:date>2026-02-17T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/164868">
<title>ML Prediction Models to Identify Novel Beyond Visual Range Tactics and Error Analysis for DARPA AIR Agents</title>
<link>https://hdl.handle.net/1721.1/164868</link>
<description>ML Prediction Models to Identify Novel Beyond Visual Range Tactics and Error Analysis for DARPA AIR Agents
Li, William; Castor, Jeremy
This paper investigates the utility of using machine&#13;
learning models to predict the outcome of simulated 2 vs. 2&#13;
Tactical Intercept engagements flown by autonomous agents in&#13;
support of the DARPA Artificial Intelligence Reinforcements&#13;
(AIR) program. We investigated the performance of four models:&#13;
Feed Forward Neural Network, Random Forest, Extreme&#13;
Gradient Boost, and Long Short Term Memory (LSTM). We&#13;
examined their ability to successfully predict the outcomes of&#13;
simulated engagements, tactical errors, and the execution of novel&#13;
game plans by autonomous agents. The models were trained on&#13;
53 features pertaining to the agents including distance between&#13;
aircraft, altitude, speed, missile availability, and other eventbased&#13;
features from simulated runs. The LSTM model had the&#13;
best performance towards the beginning of a run and was able to&#13;
predict the correct winner with 87.8% accuracy only one minute&#13;
into a run while the XGBoost model achieved the best overall&#13;
performance with a 91.7% classification accuracy and an R² of&#13;
0.712. The XGB model was also able to correctly predict the&#13;
winner of 84.7% of the runs after only seven minutes into the&#13;
simulated engagement. These results demonstrate the utility and&#13;
need for further investigation into other ML models potential&#13;
to identify unique attributes and predictive analysis of more&#13;
complex multi-agent scenarios that include additional criteria&#13;
such as varying rules of engagement, incorporating acceptable&#13;
levels of risk as well as other requirements fighter pilots must take&#13;
into account during offensive and defensive operations needed to&#13;
gain air superiority and support the objectives of the Joint Forces&#13;
Commander.
</description>
<dc:date>2026-02-12T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162653">
<title>Artificial Intelligence for Tactical Network Troubleshooting</title>
<link>https://hdl.handle.net/1721.1/162653</link>
<description>Artificial Intelligence for Tactical Network Troubleshooting
Jaimes, Rafael; Mendez, Maximillian
The tactical network is a key component of most&#13;
United States Marine Corps missions. It is critical to expeditiously&#13;
stand up a robust communications architecture for both voice&#13;
and data transmissions across a variety of classification levels.&#13;
However, when there are unforeseen or induced faults in network&#13;
configurations, the establishment time can increase by hours&#13;
if not days. The research described in this report sought to&#13;
determine if a large language model (LLM), when provided&#13;
the correct baseline network configurations, would be able to&#13;
identify errors in active working network configurations and&#13;
reduce network establishment time. A/B testing was conducted to&#13;
see whether teams assisted by artificial intelligence (AI) or control&#13;
teams with no AI assistance could establish the network faster.&#13;
The LLM hosted by NIPRGPT decreased the establishment time&#13;
by 50 percent (p &lt;0.05) compared to warfighters unaided by AI.&#13;
The results conclude that AI agents such as LLMs can be useful&#13;
in providing commanders with a course of action to establish&#13;
command, control, communications, and computers (C4) faster.
The Department of the Air Force Artificial Intelligence Accelerator
</description>
<dc:date>2025-09-12T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162634">
<title>A KNOWLEDGE GRAPH IS ALL YOU NEED</title>
<link>https://hdl.handle.net/1721.1/162634</link>
<description>A KNOWLEDGE GRAPH IS ALL YOU NEED
Streilen, William; Brooks, Nicholas; Burill, Daniel; Smith, Corey
The Department of the Air Force (DAF) faces&#13;
unique challenges in adopting Large Language Models&#13;
(LLMs). Commercially available models often lack the&#13;
domain-specific knowledge necessary to support airmen,&#13;
as this information is not inherently embedded. To maintain&#13;
a competitive edge, the integration of LLMs to&#13;
improve efficiency and decision making is a critical priority.&#13;
This presentation explores two innovative methodologies&#13;
designed to better integrate domain-specific knowledge&#13;
into language models and improve the discovery of&#13;
relevant information. The first is EntiGraph Continuous&#13;
Pretraining, which leverages continuous training to embed&#13;
specialized knowledge into language models. The second&#13;
is the GFM-RAG Graph RAG Framework, a novel approach&#13;
to knowledge retrieval and synthesis that enhances&#13;
model performance by improving multi-hop retrieval and&#13;
complex information connections.&#13;
Through both quantitative and qualitative evaluations, we&#13;
assess their impact on retrieval accuracy and response&#13;
relevance. Our findings demonstrate the potential of these&#13;
customized approaches to streamline information access,&#13;
improve decision making, and better support the operational&#13;
needs of the DAF.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162632">
<title>LLM-Based Entity Extraction for Cyber Threat Reports</title>
<link>https://hdl.handle.net/1721.1/162632</link>
<description>LLM-Based Entity Extraction for Cyber Threat Reports
Alperin, Kenneth; de Silva, Alexis
As the cyber threat landscape and capabilities&#13;
of advance persistent threats continue to expand, applying&#13;
cutting-edge technology to the domain of cyber intelligence&#13;
is necessary for the United States Space Force&#13;
to keep pace in the Great Power Competition. Cyber&#13;
intelligence analysts spend an estimated time of nearly&#13;
840 man-hours annually on the extraction and validation&#13;
of relevant intelligence from cyber threat reports (CTRs).&#13;
Named entity recognition (NER) is a natural language&#13;
processing technique capable of automatically extracting&#13;
and labeling all relevant information from a given text.&#13;
Although not a novel idea, this paper aims to expand&#13;
the current but limited research on the applications of&#13;
NER to the domain of cyber intelligence. This study&#13;
uses a new openly-licensed dataset, AnnoCTR, to finetune&#13;
a cybersecurity-specific, transformers-based model,&#13;
CYBERT. The performance of the model is compared&#13;
to the models from the derived literature. Although the&#13;
results showed an F1 score of 0.733 – a less optimal&#13;
performance compared to previous models – there is&#13;
still more work to explore to reduce the production time&#13;
of intelligence analysis by half.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162631">
<title>Democratizing Data: An Intelligent Querying System for Marine Corps Data</title>
<link>https://hdl.handle.net/1721.1/162631</link>
<description>Democratizing Data: An Intelligent Querying System for Marine Corps Data
Johnson, Lane; Nam, Kevin
This research presents the development and implementation&#13;
of a text-to-Structured Query Language (SQL)&#13;
system tailored for Marine Corps logistics, capitalizing upon&#13;
the proven capabilities of Large Language Models (LLMs). By&#13;
fine-tuning an open-source LLM on a curated Global Combat&#13;
Support System - Marine Corps supply and maintenance dataset,&#13;
we demonstrate how non-technical users can intuitively interact&#13;
with Marine Corps data through natural language queries,&#13;
enhancing data accessibility and operational decision-making.&#13;
Our approach assumes a resource-constrained environment,&#13;
demonstrating that fine-tuning and deploying the model on a&#13;
single NVIDIA A100 graphics processing unit (GPU) is not&#13;
only feasible, but also highlights the potential for local or edgebased&#13;
artificial intelligence (AI) solutions. We further identify the&#13;
critical importance of high-quality, representative datasets and&#13;
propose a hybrid approach combining prompt engineering with&#13;
fine-tuning to improve performance. Our findings culminate in&#13;
concrete recommendations for the Marine Corps regarding data&#13;
governance, AI integration, and workforce development.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162630">
<title>Pixels to Places: Improving Zero-Shot Image Geolocalization using Prior Knowledge</title>
<link>https://hdl.handle.net/1721.1/162630</link>
<description>Pixels to Places: Improving Zero-Shot Image Geolocalization using Prior Knowledge
Cha, Miriam; Borg, Trent
The ability to predict the geographic origin of&#13;
a photo is critical for open-source investigation applications.&#13;
However, image geolocalization is highly challenging due to&#13;
the vast diversity of images captured worldwide. While vision&#13;
transformer-based approaches have demonstrated success—&#13;
even outperforming grandmasters in geolocation games like&#13;
GeoGuessr—their performance does not generalize well to unseen&#13;
locations. Prior methods rely solely on visual cues, neglecting&#13;
broader contextual knowledge that image analysts typically&#13;
employ. To bridge this gap, our research integrates the contextual&#13;
understanding of geographic regions that imagery analysts&#13;
possess into the geolocalization model. Specifically, we develop a&#13;
variant of StreetCLIP, which embeds CLIP within geolocalization&#13;
tasks and facilitates the incorporation of user-supplied prior&#13;
knowledge such as continental or national boundaries. Our&#13;
results on the IM2GPS3K benchmark dataset demonstrate a&#13;
10.66% improvement in regional prediction (within 200 km)&#13;
and a 15.27% improvement in country-level prediction (within&#13;
750 km) over baseline models. Our results suggest that humanprovided&#13;
supervision can enhance image geolocalization accuracy,&#13;
highlighting the potential of interactive systems where human&#13;
expertise and AI work collaboratively to refine predictions.&#13;
Index Terms—image geolocalization, CLIP, human-machine&#13;
teaming, vision transformers
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162629">
<title>Improved Automatic Electronic Intelligence Collection System for Internal and External Forward Fusion and Collaborative Geolocation of Adversary Emitters</title>
<link>https://hdl.handle.net/1721.1/162629</link>
<description>Improved Automatic Electronic Intelligence Collection System for Internal and External Forward Fusion and Collaborative Geolocation of Adversary Emitters
Botero, Joey; Benge, Arianne; Heisey, Curtis
With the 2022 National Defense Strategy shifting&#13;
focus from counterinsurgency operations to near-peer adversaries,&#13;
airborne ISR platforms within the USAF and DoD must&#13;
be improved for effectiveness in a near-peer conflict. They&#13;
need to be able to operate quickly and effectively in contested&#13;
environments with longer-range threats, act as a forward edge&#13;
intelligence node for blue forces and provide DoD Research&#13;
and Development efforts with cutting-edge data regarding new&#13;
adversary signals and technology.&#13;
To aid in tackling these challenges, this project introduces a&#13;
Machine Learning (ML)-driven approach that revamps the Automatic&#13;
Electronic Intelligence Collection System (ACS) on U.S.&#13;
Airborne ISR platforms in four ways: First, by providing nodal&#13;
analysis to the user in real time by automatically aggregating&#13;
existing data across the aircraft to the user for decreased operator&#13;
cognitive load. Second, increasing internal aircraft database&#13;
information with external intelligence database information to&#13;
increase confidence in targeting. Third, by providing automatic&#13;
signal anomaly detection to the operator utilizing a support&#13;
vector machines algorithm that cues operators to potential&#13;
signals of interest based on previous activity and pattern of life&#13;
prediction. Lastly, by providing better surface against airborne&#13;
identification through utilization of cone angle to the system&#13;
to help operators with faster threat warning and situational&#13;
awareness of the environment.&#13;
Findings include Support Vector Machines being the most&#13;
effective tested binary classifier for predicting single signal&#13;
anomaly detection at 84% AUC and a rule-based method of&#13;
averages successfully classifying 1089 surface versus air ELINT&#13;
samples with a success rate of 89% compared to other tested&#13;
methods, such as Gaussian Mixture Models at 68% and KNearest&#13;
Neighbor at 66%.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162628">
<title>Artificial Intelligence for Derivative Security Classification: Applications to DoD</title>
<link>https://hdl.handle.net/1721.1/162628</link>
<description>Artificial Intelligence for Derivative Security Classification: Applications to DoD
Gelbard, Andrew; Hamilton, Lei
The accurate classification of government&#13;
documents according to their sensitivity (e.g., UNCLASSIFIED,&#13;
SECRET, TOP SECRET) is critical for national&#13;
security, yet historically has relied on time-intensive&#13;
manual review. The current manual classification process&#13;
consumes millions of labor hours annually within the&#13;
U.S. government, significantly diverting skilled personnel&#13;
from essential analytical tasks. This research explores&#13;
automating this security classification task using recently&#13;
available declassified materials from the DISC&#13;
dataset [1], addressing practical challenges such as&#13;
noisy Optical Character Recognition (OCR) output,&#13;
imbalanced data distributions, and potential leakage&#13;
of explicit classification markers within document text.&#13;
This dataset contains declassified government documents&#13;
sourced from the Digital National Security Archive, providing&#13;
authentic textual examples representative of actual&#13;
classification scenarios. We evaluate both traditional&#13;
machine learning approaches and advanced transformerbased&#13;
language models to classify documents accurately&#13;
across multiple sensitivity levels. Our results highlight&#13;
that transformer-based models, particularly DeBERTa,&#13;
effectively improve identification of the minority but&#13;
critical TOP SECRET class, achieving recall over 70%&#13;
and an overall balanced performance (macro F1 score of&#13;
0.75), while traditional methods exhibit similar overall&#13;
accuracy but struggle with minority class recall. Despite&#13;
promising findings, we caution that conclusions drawn&#13;
here remain constrained by limited training data size&#13;
and inherent uncertainties in human-labeled documents.&#13;
We emphasize the need for larger, rigorously preprocessed&#13;
datasets and suggest future research integrating&#13;
authoritative classification guidelines directly into model&#13;
training, potentially via retrieval-augmented methods.&#13;
This work thus contributes a foundational, reproducible&#13;
framework that demonstrates significant potential for&#13;
machine-assisted security classification, guiding future&#13;
research and practical applications in the information&#13;
security domain.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
<item rdf:about="https://hdl.handle.net/1721.1/162627">
<title>The Area-of-Measurable-Performance (AOMP) Method Standard as a Foundational Archetype for the Cyclical Enhancement of the State of the Art Joint Simulation Environment (JSE) Technology</title>
<link>https://hdl.handle.net/1721.1/162627</link>
<description>The Area-of-Measurable-Performance (AOMP) Method Standard as a Foundational Archetype for the Cyclical Enhancement of the State of the Art Joint Simulation Environment (JSE) Technology
Li, William; Johnson, Kevin; Picardo, Christopher; Ambion, Francis
The Department of the Air Force (DAF) envisions the need to incorporate Artificial Intelligence and Machine Learning (AI/ML) models into novel systems it develops for the purpose of enhancing them to meet its primary goal of maintaining total air superiority [2]. There is currently a need for developing a standard process for the design of successful AI/ML models capable of enhancing the novel systems the DAF develops. In this white paper we introduce the Area of Measurable Performance (AOMP) Method Standard and apply it to the Joint Simulation Environment (JSE) Technology, a state of the art system of systems under test, to identify AOMPS and their modular requirements [3] and metrics that lead to the accurate characterization of modular AI/ML models through a process that offers a high degree of trust and reuse, resulting in a method standard that organically promotes the development of successful modular AI/ML models for use in the performance improvement of the JSE technology or other system of system(s) [4] under test.
</description>
<dc:date>2025-09-10T00:00:00Z</dc:date>
</item>
</rdf:RDF>
