Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

Soumya Baddham

Battling Toxicity: A Comparative Analysis of Machine Learning Models for Content Moderation

When & Where:


Eaton Hall, Room 2001B

Committee Members:

David Johnson, Chair
Prasad Kulkarni
Hongyang Sun


Abstract

With the exponential growth of user-generated content, online platforms face unprecedented challenges in moderating toxic and harmful comments. Due to this, Automated content moderation has emerged as a critical application of machine learning, enabling platforms to ensure user safety and maintain community standards. Despite its importance, challenges such as severe class imbalance, contextual ambiguity, and the diverse nature of toxic language often compromise moderation accuracy, leading to biased classification performance.

This project presents a comparative analysis of machine learning approaches for a Multi-Label Toxic Comment Classification System using the Toxic Comment Classification dataset from Kaggle.  The study examines the performance of traditional algorithms, such as Logistic Regression, Random Forest, and XGBoost, alongside deep architectures, including Bi-LSTM, CNN-Bi-LSTM, and DistilBERT. The proposed approach utilizes word-level embeddings across all models and examines the effects of architectural enhancements, hyperparameter optimization, and advanced training strategies on model robustness and predictive accuracy.

The study emphasizes the significance of loss function optimization and threshold adjustment strategies in improving the detection of minority classes. The comparative results reveal distinct performance trade-offs across model architectures, with transformer models achieving superior contextual understanding at the cost of computational complexity. At the same time, deep learning approaches(LSTM models) offer efficiency advantages. These findings establish evidence-based guidelines for model selection in real-world content moderation systems, striking a balance between accuracy requirements and operational constraints.


Manu Chaudhary

Utilizing Quantum Computing for Solving Multidimensional Partial Differential Equations

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Esam El-Araby, Chair
Perry Alexander
Tamzidul Hoque
Prasad Kulkarni
Tyrone Duncan

Abstract

Quantum computing has the potential to revolutionize computational problem-solving by leveraging the quantum mechanical phenomena of superposition and entanglement, which allows for processing a large amount of information simultaneously. This capability is significant in the numerical solution of complex and/or multidimensional partial differential equations (PDEs), which are fundamental to modeling various physical phenomena. There are currently many quantum techniques available for solving partial differential equations (PDEs), which are mainly based on variational quantum circuits. However, the existing quantum PDE solvers, particularly those based on variational quantum eigensolver (VQE) techniques, suffer from several limitations. These include low accuracy, high execution times, and low scalability on quantum simulators as well as on noisy intermediate-scale quantum (NISQ) devices, especially for multidimensional PDEs.

 In this work, we propose an efficient and scalable algorithm for solving multidimensional PDEs. We present two variants of our algorithm: the first leverages finite-difference method (FDM), classical-to-quantum (C2Q) encoding, and numerical instantiation, while the second employs FDM, C2Q, and column-by-column decomposition (CCD). Both variants are designed to enhance accuracy and scalability while reducing execution times. We have validated and evaluated our proposed concepts using a number of case studies including multidimensional Poisson equation, multidimensional heat equation, Black Scholes equation, and Navier-Stokes equation for computational fluid dynamics (CFD) achieving promising results. Our results demonstrate higher accuracy, higher scalability, and faster execution times compared to VQE-based solvers on noise-free and noisy quantum simulators from IBM. Additionally, we validated our approach on hardware emulators and actual quantum hardware, employing noise mitigation techniques. This work establishes a practical and effective approach for solving PDEs using quantum computing for engineering and scientific applications.


Alex Manley

Taming Complexity in Computer Architecture through Modern AI-Assisted Design and Education

When & Where:


Nichols Hall, Room 250 (Gemini Room)

Committee Members:

Heechul Yun, Chair
Tamzidul Hoque
Prasad Kulkarni
Mohammad Alian

Abstract

The escalating complexity inherent in modern computer architecture presents significant challenges for both professional hardware designers and students striving to gain foundational understanding. Historically, the steady improvement of computer systems was driven by transistor scaling, predictable performance increases, and relatively straightforward architectural paradigms. However, with the end of traditional scaling laws and the rise of heterogeneous and parallel architectures, designers now face unprecedented intricacies involving power management, thermal constraints, security considerations, and sophisticated software interactions. Prior tools and methodologies, often reliant on complex, command-line driven simulations, exacerbate these challenges by introducing steep learning curves, creating a critical need for more intuitive, accessible, and efficient solutions. To address these challenges, this thesis introduces two innovative, modern tools.

The first tool, SimScholar, provides an intuitive graphical user interface (GUI) built upon the widely-used gem5 simulator. SimScholar significantly simplifies the simulation process, enabling students and educators to more effectively engage with architectural concepts through a visually guided environment, both reducing complexity and enhancing conceptual understanding. Supporting SimScholar, the gem5 Extended Modules API (gEMA) offers streamlined backend integration with gem5, ensuring efficient communication, modularity, and maintainability.

The second contribution, gem5 Co-Pilot, delivers an advanced framework for architectural design space exploration (DSE). Co-Pilot integrates cycle-accurate simulation via gem5, detailed power and area modeling through McPAT, and intelligent optimization assisted by a large language model (LLM). Central to Co-Pilot is the Design Space Declarative Language (DSDL), a Python-based domain-specific language that facilitates structured, clear specification of design parameters and constraints.

Collectively, these tools constitute a comprehensive approach to taming complexity in computer architecture, offering powerful, user-friendly solutions tailored to both educational and professional settings.


Past Defense Notices

Dates

NAGA ANUSHA BOMMIDI

The Comparison of Performance and Complexity of Rule Sets induced from Incomplete Data

When & Where:


317 Nichols Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Andy Gill
Prasad Kulkarni


Abstract

The main focus of this project is to identify the best interpretation of missing attribute values in terms of performance and complexity of rule sets. This report summarizes the experimental comparison of the performance and the complexity of rule sets induced from incomplete data sets with three interpretations of missing attribute values: lost values, attribute-concept values, and “do not care” conditions. Furthermore, it details the experiments conducted using MLEM2 rule induction system on 176 data sets, using three kinds of probabilistic approximations: lower, middle and upper. The performance was evaluated using the error rate computed by ten-fold cross validation, and the complexity of rule sets was evaluated based the size of the rule sets and the number of conditions in the rule sets. The results showed that lost values were better in terms of the performance in 10 out of 24 combinations. In addition, attribute-concept values were better in 5 out of 24 combinations, and “do not care” conditions were better in 1 combination in terms of the complexity of rule sets. Furthermore, there was not even one combination of dataset and type of approximation for which both performance and complexity of rule sets were better for one interpretation of missing attributes compared to the other two.


BLAKE BRYANT

Hacking SIEMS to Catch Hackers: Decreasing the Mean Time to Respond to Security Incidents with a Novel Threat Ontology in SIEM Software

When & Where:


2012 BEST

Committee Members:

Hossein Saiedian, Chair
Bo Luo
Gary Minden


Abstract

Information security is plagued with increasingly sophisticated and persistent threats to communication networks. The development of new threat tools or vulnerability exploits often outpaces advancements in network security detection systems. As a result, detection systems often compensate by over reporting partial detections of routine network activity to security analysts for further review. Such alarms seldom contain adequate forensic data for analysts to accurately validate alerts to other stakeholders without lengthy investigations. As a result, security analysts often ignore the vast majority of network security alarms provided by sensors, resulting in security breaches that may have otherwise been prevented. 

Security Information and Event Management (SIEM) software has been introduced recently in an effort to enable data correlation across multiple sensors, with the intent of producing a lower number of security alerts with little forensic value and a higher number of security alerts that accurately reflect malicious actions. However, the normalization frameworks found in current SIEM systems do not accurately depict modern threat activities. As a result, recent network security research has introduced the concept of a "kill chain" model designed to represent threat activities based upon patterns of action, known indicators, and methodical intrusion phases. Such a model was hypothesized by many researchers to result in the realization of the desired goals of SIEM software. 

The focus of this thesis is the implementation of a "kill chain" framework within SIEM software. A novel "Kill chain" model was developed and implemented within a commercial SIEM system through modifications to the existing SIEM database. These modifications resulted in a new log ontology capable of normalizing security sensor data in accordance with modern threat research. New SIEM correlation rules were developed using the novel log ontology compared to existing vendor recommended correlation rules using the default model. The novel log ontology produced promising results indicating improved detection rates, more descriptive security alarms, and a lower number of false positive alarms. These improvements were assessed to provide improved visibility and more efficient investigation processes to security analysts ultimately reducing the mean time required to detect and escalate security incidents. 


SHAUN CHUA

Implementation of a Multichannel Radar Waveform Generator System Controller

When & Where:


317 Nichols Hall

Committee Members:

Carl Leuschen, Chair
Chris Allen
Fernando Rodriguez-Morales


Abstract

Waveform generation is crucial in a radar system operation. There is a recent need for an 8 channel transmitter with high bandwidth chirp signals (100 MHz – 600 MHz). As such, a waveform generator (WFG) hardware module is required for this purpose. The WFG houses 4 Direct Digital Synthesizers (DDS), and an ALTERA Cyclone V FPGA that acts as its controller. The DDS of choice is the AD9915, because its Digital to Analog Converter can be clocked at a maximum rate of 2.5 GHz, allowing plenty of room to produce the high bandwidth and high frequency chirp signals desired, and also because it supports synchronization between multiple AD9915s. 

The brains behind the DDS operations are the FPGA and the radar software developed in NI LabVIEW. These two aspects of the digital systems grants the WFG highly configurable waveform capabilities. The configurable inputs that can be controlled by the user include: number of waveforms in a playlist, start and stop frequency (bandwidth of chirp signal), zero-pi mode, and waveform amplitude and phase control. 

The FPGA acts as a DDS controller that directly configures and control the DDS operations, while also managing and synchronizing the operations of all DDS channels. This project details largely the development of such a controller, named Multichannel Waveform Generator (MWFG) Controller, and the necessary modifications and development in the NI LabVIEW software, so that they complement each other.


DEEPIKA KOTA

Automatic Color Detection of Colored Wires In Electric Cables

When & Where:


2001B Eaton Hall

Committee Members:

Jim Stiles, Chair
Ron Hui
James Rowland


Abstract

An automatic Color detection system checks for the sequence of colored wires in electric cables which are ready to get crimped together. The system inspects for flat connectors with differs in type and number of wires.This is managed in an automatic way with a self learning system without any requirement of manual input from the user to load new data to the machine. The system is coupled to a connector crimping machine and once the system learns the actual sample of cable order , it automatically inspects each cable assembled by the machine. There are three methodologies based on which this automatic detection takes place 1) A self learning system 2) An algorithm for wire segmentation to extract colors from the captured images 3) An algorithm for color recognition to cope up with wires with different illuminations and insulation .The main advantage of this system is when the cables are produced in large batches ,it provides high level of accuracy and prevents false negatives in order to guarantee defect free production.


MOHAMMED ZIAUDDIN

Open Source Python Widget Application to Synchronize Bibliographical References Between Two BibTeX Repositories

When & Where:


246 Nichols Hall

Committee Members:

Andy Gill, Chair
Perry Alexander
Prasad Kulkarni


Abstract

Bibtex is a tool to edit and manage bibliographical references in a document.Researchers face a common problem that they have one copy of their bibliographical reference databases for a specific project and a master bibliographical database file that holds all their bibliographical references. Syncing these two files is an arduous task as one has to search and modify each reference record individually. Most of the bibtex tools available either provide help in maintaining bibtex bibliographies in different file formats or searching for references in web databases but none of them provide a way to synchronize the fields of the same reference record in the two different bibtex database files. 
The intention of this project is to create an application that helps academicians to keep their bibliographical references in two different databases in sync. We have created a python widget application that employs the Tkinter library for GUI and unQLite database for data storage. This application is integrated with Github allowing users to modify bibtex files present on Github. 


HARISH ROHINI

Using Intel Pintools to Analyze Memory Access Patterns

When & Where:


246 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Andy Gill
Heechul Yun


Abstract

Analysis of large benchmark programs can be very difficult because of their changes in memory state for every run and with billions of instructions the simulation of a whole program in general can be extremely slow. The solution for this is to simulate only some selected regions which are the most representative parts of a program, So that we can focus our analysis and optimizations on those particular regions which represent more part of the execution of a program. In order to accomplish that, we use intel’s pintool, a binary instrumentation framework which performs program analysis at run time, simpoint to get the most representative regions of a program and pinplay for the reproducible analysis of the program. This project uses these frameworks to simulate and analyze programs to provide various statistics about the memory allocations, memory reference traces, allocated memory usage across the most representative regions of the program and also the cache simulations of the representative regions.


GOVIND VEDALA

Iterative SSBI Compensation in Optical OFDM Systems and the Impact of SOA Nonlinearities MS Project Defense (EE)

When & Where:


246 Nichols Hall

Committee Members:

Ron Hui, Chair
Chris Allen
Erik Perrins


Abstract

Multicarrier modulation using Orthogonal Frequency Division Multiplexing (OFDM) is a best fit candidate for the next generation long-haul optical transmission systems, offering high degree of spectral efficiency and easing the compensation of linear impairments such as chromatic dispersion and polarization mode dispersion, at the receiver. Optical OFDM comes in two flavors – coherent optical OFDM (CO-OFDM) and direct detection optical OFDM (DD-OFDM), each having its own share of pros and cons. CO-OFDM is highly robust to fiber impairments and imposes a relaxation on the electronic component bandwidth requirements, but requires narrow linewidth lasers, optical hybrids and local oscillators. On the other hand DD-OFDM has relaxed laser linewidth requirement and low complexity receiver making it an attractive multicarrier system. However, DD-OFDM system suffers from signal-signal beat interference (SSBI), caused by mixing among the sub-carriers in the photo detector, which deteriorates the system performance. Previously, to mitigate the effect of SSBI, a guard band was used between optical carrier and data sideband. In this project, we experimentally demonstrate a linearly field modulated virtual single sideband OFDM (VSSB-OFDM) transmission with direct detection and digitally compensate for the SSBI using an iterative SSBI compensation algorithm. 
Semiconductor optical amplifiers (SOA), with their small footprint, ultra-high gain bandwidth, and ease of integration, are attracting the attention of optical telecommunication engineers for their use in high speed transmission systems as inline amplifiers. However, the SOA gain saturation induced nonlinearities cause pulse distortion and induce nonlinear cross talk effects such as cross gain modulation especially in Wavelength Division Multiplexed systems. In this project, we also evaluate the performance of iterative SSBI compensation in an optical OFDM system, in the presence of these SOA induced nonlinearities. 

 


KEERTHI GANTA

TCP Illinois Protocol Implementation in ns-3

When & Where:


250 Nichols Hall

Committee Members:

James Sterbenz, Chair
Victor Frost
Bo Luo


Abstract

The choice of congestion control algorithm has an impact on the performance of a network. The congestion control algorithm should be selected and implemented based on the network scenario in order to achieve better results. Congestion control in high speed networks and networks with large BDP is proved to be more critical due to the high amount of data at risk. There are problems in achieving better throughput with conventional TCP in the above mentioned scenario. Over the years conventional TCP is modified to pave way for TCP variants that could address the issues in high speed networks. TCP Illinois is one such protocol for high speed networks. It is a hybrid version of a congestion control algorithm as it uses both packet loss and delay information to decide on the window size. The packet loss information is used to decide on whether to increase or decrease the congestion window and delay information is used to assess the amount of increase or decrease that has to be made.


ADITYA RAVIKANTI

sheets-db: Database powered by Google Spreadsheets

When & Where:


2001B Eaton Hall

Committee Members:

Andy Gill, Chair
Perry Alexander
Prasad Kulkarni


Abstract

The sheets-db library is a Haskell binding to Google Sheets API. sheets-db allows Haskell users to utilize google spread sheets as a light weight database. It provides various functions to create, read, update and delete rows in spreadsheets along with a way to construct simple structured queries. 


NIRANJAN PURA VEDAMURTHY

Testing the Accuracy of Erlang Delay Formula for Smaller Number of TCP Flows

When & Where:


246 Nichols Hall

Committee Members:

Victor Frost, Chair
Gary Minden
Glenn Prescott


Abstract

The Erlang delay formula for dimensioning different networks is used to calculate the probability of congestion. Testing the accuracy of a probability of congestion found using the Erlang formula against the simulation for probability of packet loss is demonstrated in this project. The simulations are done when TCP traffic is applied through one bottleneck node. Three different source traffic models having small number of flows is considered. Simulations results for three different source traffic models is shown in terms of probability of packet loss and load supplied to the topology. Various traffic parameters are varied in order to show the impact on the probability of packet loss and to compare with the Erlang prediction for probability of congestion.