Dr. Jesse Ables Receives $155,000 NSF Grant for Explainable Intrusion Detection Research
Posted on August 29, 2025 by Keith Lynn

Dr. Jesse Ables has been awarded a $155,000 grant from the National Science Foundation
(NSF) under the Computer and Information Science and Engineering Research Initiation
Initiative (CRII). His project was specifically accepted for the NSF Secure and Trustworthy
Computing (SaTC) CRII.
His funded project, titled “Temporal Eclectic Rule Extraction (TERE) and Rule Extraction
Optimization Strategies for Explainable Intrusion Detection Systems (X-IDS),” combines
two critical areas of cybersecurity and artificial intelligence: Intrusion Detection
Systems (IDS), which safeguard computer networks, and Explainable Artificial Intelligence
(XAI).
“Right now, many powerful AI systems are like a ‘black box’—they can detect issues in a network, such as hacking attempts or virus activity, but we often don’t know how or why they reached that conclusion. This makes it harder to fully trust their results. My research is focused on opening up that black box,” explained Dr. Ables.
Dr. Ables is developing a new algorithm called Temporal Eclectic Rule Extraction (TERE). Its goal is to create "white-box" explanations for a specific type of powerful AI called a Recurrent Neural Network (RNN). Essentially, the algorithm will allow us to see the reasoning behind an RNN's predictions.
In addition, the research will examine optimization strategies for eclectic rule extraction, as this process currently exhibits poor runtime scalability.
In short: “I'm building a way to make advanced network security systems not only powerful, but also transparent and understandable. This will help us better defend against cyber threats by knowing exactly why and how an AI identifies a threat.”
Dr. Ables’ interest in Explainable AI began during his Ph.D. research, where his dissertation challenged prevailing notions of what constitutes a trustworthy explanation. He argues that most current XAI solutions are neither truly explainable nor inherently trustworthy, and that new methods must prioritize both qualities from the ground up. His work explores philosophical and technical questions such as:
- How do we generate useful explanations?
- What form will these explanations take (visual, textual, resource-backed)?
- Who are the explanations for?
- How will we use the explanations?
-
Dr. Jesse Ables Receives $155,000 NSF Grant for Explainable Intrusion Detection Research
Dr. Jesse Ables has been awarded a $155,000 grant from the National Sc...
August 29, 2025 -
The School of Computing has launched a new fully online bachelor's degree program in Information Systems
The University of South Alabama School of Computing has launched a new...
May 8, 2025 -
SoC faculty member featured at 3rd Annual Research and Technology Showcase
The Office of Research and Economic Development hosted its 3rd Annual ...
November 1, 2024 -
SoC student Destin Hinkel places second in the Three Minute Thesis competition.
The University of South Alabama Graduate School hosted its annual 3MT ...
October 29, 2024