Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

Md Mashfiq Rizvee

Hierarchical Probabilistic Architectures for Scalable Biometric and Electronic Authentication in Secure Surveillance Ecosystems

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Sumaiya Shomaji, Chair
Tamzidul Hoque
David Johnson
Hongyang Sun
Alexandra Kondyli

Abstract

Secure and scalable authentication has become a primary requirement in modern digital ecosystems, where both human biometrics and electronic identities must be verified under noise, large population growth and resource constraints. Existing approaches often struggle to simultaneously provide storage efficiency, dynamic updates and strong authentication reliability. The proposed work advances a unified probabilistic framework based on Hierarchical Bloom Filter (HBF) architectures to address these limitations across biometric and hardware domains. The first contribution establishes the Dynamic Hierarchical Bloom Filter (DHBF) as a noise-tolerant and dynamically updatable authentication structure for large-scale biometrics. Unlike static Bloom-based systems that require reconstruction upon updates, DHBF supports enrollment, querying, insertion and deletion without structural rebuild. Experimental evaluation on 30,000 facial biometric templates demonstrates 100% enrollment and query accuracy, including robust acceptance of noisy biometric inputs while maintaining correct rejection of non-enrolled identities. These results validate that hierarchical probabilistic encoding can preserve both scalability and authentication reliability in practical deployments. Building on this foundation, Bio-BloomChain integrates DHBF into a blockchain-based smart contract framework to provide tamper-evident, privacy-preserving biometric lifecycle management. The system stores only hashed and non-invertible commitments on-chain while maintaining probabilistic verification logic within the contract layer. Large-scale evaluation again reports 100% enrollment, insertion, query and deletion accuracy across 30,000 templates, therefore, solving the existing problem of blockchains being able to authenticate noisy data. Moreover, the deployment analysis shows that execution on Polygon zkEVM reduces operational costs by several orders of magnitude compared to Ethereum, therefore, bringing enrollment and deletion costs below $0.001 per operation which demonstrate the feasibility of scalable blockchain biometric authentication in practice. Finally, the hierarchical probabilistic paradigm is extended to electronic hardware authentication through the Persistent Hierarchical Bloom Filter (PHBF). Applied to electronic fingerprints derived from physical unclonable functions (PUFs), PHBF demonstrates robust authentication under environmental variations such as temperature-induced noise. Experimental results show zero-error operation at the selected decision threshold and substantial system-level improvements as well as over 10^5 faster query processing and significantly reduced storage requirements compared to large scale tracking.


Fatima Al-Shaikhli

Optical Measurements Leveraging Coherent Fiber Optics Transceivers

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Rongqing Hui, Chair
Shannon Blunt
Shima Fardad
Alessandro Salandrino
Judy Wu

Abstract

Recent advancements in optical technology are invaluable in a variety of fields, extending far beyond high-speed communications. These innovations enable optical sensing, which plays a critical role across diverse applications, from medical diagnostics to infrastructure monitoring and automotive systems. This research focuses on leveraging commercially available coherent optical transceivers to develop novel measurement techniques to extract detailed information about optical fiber characteristics, as well as target information. Through this approach, we aim to enable accurate and fast assessments of fiber performance and integrity, while exploring the potential for utilizing existing optical communication networks to enhance fiber characterization capabilities. This goal is investigated through three distinct projects: (1) fiber type characterization based on intensity-modulated electrostriction response, (2) coherent Light Detection and Ranging (LiDAR) system for target range and velocity detection through different waveform design, including experimental validation of frequency modulation continuous wave (FMCW) implementations and theoretical analysis of orthogonal frequency division multiplexing (OFDM) based approaches and (3) birefringence measurements using a coherent Polarization-sensitive Optical Frequency Domain Reflectometer (P-OFDR) system.

Electrostriction in an optical fiber is introduced by interaction between the forward propagated optical signal and the acoustic standing waves in the radial direction resonating between the center of the core and the cladding circumference of the fiber. The response of electrostriction is dependent on fiber parameters, especially the mode field radius. We demonstrated a novel technique of identifying fiber types through the measurement of intensity modulation induced electrostriction response. As the spectral envelope of electrostriction induced propagation loss is anti-symmetrical, the signal to noise ratio can be significantly increased by subtracting the measured spectrum from its complex conjugate. We show that if the field distribution of the fiber propagation mode is Gaussian, the envelope of the electrostriction-induced loss spectrum closely follows a Maxwellian distribution whose shape can be specified by a single parameter determined by the mode field radius.        

We also present a self-homodyne FMCW LiDAR system based on a coherent receiver. By using the same linearly chirped waveform for both the LiDAR signal and the local oscillator, the self-homodyne coherent receiver performs frequency de-chirping directly in the photodiodes, significantly simplifying signal processing. As a result, the required receiver bandwidth is much lower than the chirping bandwidth of the signal. Simultaneous multi-target of range and velocity detection is demonstrated experimentally. Furthermore, we explore the use of commercially available coherent transceivers for joint communication and sensing using OFDM waveforms.

In addition, we demonstrate a P-OFDR system utilizing a digital coherent optical transceiver to generate a linear frequency chirp via carrier-suppressed single-sideband modulation. This method ensures linearity in chirping and phase continuity of the optical carrier. The coherent homodyne receiver, incorporating both polarization and phase diversity, recovers the state of polarization (SOP) of the backscattered optical signal along the fiber, mixing with an identically chirped local oscillator. With a spatial resolution of approximately 5 mm, a 26 GHz chirping bandwidth, and a 200 us measurement time, this system enables precise birefringence measurements. By employing three mutually orthogonal SOPs of the launched optical signal, we measure relative birefringence vectors along the fiber.


Past Defense Notices

Dates

RUXIN XIE

Single-fiber-laser-based-multimodal coherent Raman System

When & Where:


250 Nichols Hall

Committee Members:

Ron Hui, Chair
Chris Allen
Shannon Blunt
Victor Frost
Carey Johnson

Abstract

Coherent Raman scattering (CRS) is an appealing technique for spectroscopy and microscopy, due to its selectivity and sensitivity. We designed and built single-fiber-laser-based coherent Raman scattering spectroscopy and microscopy system which can automatically maintain frequency synchronization between pump and Stokes beam. The Stokes frequency shift is generated by soliton self-frequency shift (SSFS) through a photonic crystal fiber. The impact of pulse chirping on the signal power reduction of coherent anti-Stokes Raman scattering (CARS) and stimulated Raman scattering (SRS) have been investigate through theoretical analysis and experiment. 

Our multimodal system provides measurement diversity among CARS, SRS and photothermal, which can be used for comparison and offering complementary information. Distribution of hemoglobin in human red blood cells and lipids in sliced mouse brain sample have been imaged. Frequency and power dependency of photothermal signal is characterized. 
Based on the polarization dependency of the third-order susceptibility of the material, the polarization switched SRS method is able to eliminate the nonresonant photothermal signal from the resonant SRS signal. Red blood cells and sliced mouse brain samples were imaged to demonstrate the capability of the proposed technique. The result shows that polarization switched SRS removes most of the photothermal signal. 


MAHITHA DODDALA

Properties of Probabilistic Approximations Applied to Incomplete Data

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Man Kong
Bo Luo


Abstract

The main focus of the project is to discuss mining of incomplete data which we find frequently in real-life records. For this, I considered the probabilistic approximations as they have a direct application to mining incomplete data. I have examined the results obtained from the experiments conducted on eight real-life data sets taken from University of California at Irvine Machine Learning Repository. I also investigated the properties of singleton, subset, and concept approximations and corresponding consistencies. The main objective was to compare the global and local approximations and generalize the consistency definition for incomplete data with two interpretations of missing attribute values: lost values and "do not care" conditions. In addition to this comparison, the most useful approach among singleton, subset and concept approximations is also tested for which the conclusion is the best approach would be selected with the help of tenfold cross validation after applying all three approaches. Also it’s shown that even if there exist six types of consistencies, there are only four distinct consistencies of incomplete data as two pairs of such consistencies are equivalent.


ROHIT YADAV

Automatic Text Summarization of Email Corpus Using Importance of Sentences

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Prasad Kulkarni
Bo Luo


Abstract

With the advent of Internet, the data being added online have been increasing at an enormous rate. Though search engines use information retrieval (IR) techniques to facilitate the search requests from users, the results may not always be effective or the efficiency of results according to a search query may not be high. The user has to go through certain web pages before getting at the web page he/she needs. This problem of information overload can be solved using automatic text summarization. Summarization is a process of obtaining an abridged version of documents so that user can have a quick understanding of the document. A new technique to produce a summary of an original text is investigated in this project. 
Email threads from the World Wide Web consortium’s sites (W3C) corpus are used in this system.Our system is based on identification and extraction of important sentences from the input document. Apart from common IR features like term frequency and inverse document frequency, novel features such as Term Frequency-Inverse Document Frequency,subject words, sentence position and thematic words have also been implemented. The model consists of four stages. The pre-processing stage converts the unstructured (all those things that can't be so readily classified) text into structured (any data that resides in a fixed field within a record or file). In the first stage each sentence is partitioned into the list of tokens and stop words are removed. The second stage is to extract the important key phrases in the text by implementing a new algorithm through ranking the candidate words. The system uses the extracted keywords/key phrases to select the important sentence. Each sentence is ranked depending on many features such as the existence of the keywords/key phrase in it, the relation between the sentence and the title by using a similarity measurement and other many features. The third stage of the proposed system is to extract the sentences with the highest rank. The fourth stage is the filtering stage where sentences from email threads are ranked as per features and summaries are generated. This system can be considered as a framework for unsupervised learning in the field of text summarization. 


ARJUN MUTHALAGU

Flight Search Application

When & Where:


250 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Andy Gill
Jerzy Grzymala-Busse


Abstract

“Flight-search” application is an Angular JS application implemented in a client side architecture. The application displays the flight results from different airline companies based on the input parameters. The application also has custom filtering conditions and custom pagination, which a user can interact with to filter the result and also limit the results displayed in the browser. The application uses QPX Express API to pull data for the flight searches.


SATYA KUNDETI

A comparison of Two Decision Tree Generating Algorithms: C4.5 and CART Based on Numerical Data

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Luke Huan
Bo Luo


Abstract

In Data Mining, classification of data is a challenging task. One of the most popular techniques for classifying data is decision tree induction. In this project, two decision tree generating algorithms CART and C4.5, using their original implementations, are compared on different numerical data sets, taken from University of California Irvine (UCI). The comparative analysis of these two implementations is carried out in terms of accuracy and decision tree complexity. Results from experiments show that there is statistically insignificant difference(5% level of significance, two-tailed test)between C4.5 and CART in terms of accuracy. On the other hand, decision trees generated by C4.5 and CART have significant statistical difference in terms of their complexity. 

 


NAGA ANUSHA BOMMIDI

The Comparison of Performance and Complexity of Rule Sets induced from Incomplete Data

When & Where:


317 Nichols Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Andy Gill
Prasad Kulkarni


Abstract

The main focus of this project is to identify the best interpretation of missing attribute values in terms of performance and complexity of rule sets. This report summarizes the experimental comparison of the performance and the complexity of rule sets induced from incomplete data sets with three interpretations of missing attribute values: lost values, attribute-concept values, and “do not care” conditions. Furthermore, it details the experiments conducted using MLEM2 rule induction system on 176 data sets, using three kinds of probabilistic approximations: lower, middle and upper. The performance was evaluated using the error rate computed by ten-fold cross validation, and the complexity of rule sets was evaluated based the size of the rule sets and the number of conditions in the rule sets. The results showed that lost values were better in terms of the performance in 10 out of 24 combinations. In addition, attribute-concept values were better in 5 out of 24 combinations, and “do not care” conditions were better in 1 combination in terms of the complexity of rule sets. Furthermore, there was not even one combination of dataset and type of approximation for which both performance and complexity of rule sets were better for one interpretation of missing attributes compared to the other two.


BLAKE BRYANT

Hacking SIEMS to Catch Hackers: Decreasing the Mean Time to Respond to Security Incidents with a Novel Threat Ontology in SIEM Software

When & Where:


2012 BEST

Committee Members:

Hossein Saiedian, Chair
Bo Luo
Gary Minden


Abstract

Information security is plagued with increasingly sophisticated and persistent threats to communication networks. The development of new threat tools or vulnerability exploits often outpaces advancements in network security detection systems. As a result, detection systems often compensate by over reporting partial detections of routine network activity to security analysts for further review. Such alarms seldom contain adequate forensic data for analysts to accurately validate alerts to other stakeholders without lengthy investigations. As a result, security analysts often ignore the vast majority of network security alarms provided by sensors, resulting in security breaches that may have otherwise been prevented. 

Security Information and Event Management (SIEM) software has been introduced recently in an effort to enable data correlation across multiple sensors, with the intent of producing a lower number of security alerts with little forensic value and a higher number of security alerts that accurately reflect malicious actions. However, the normalization frameworks found in current SIEM systems do not accurately depict modern threat activities. As a result, recent network security research has introduced the concept of a "kill chain" model designed to represent threat activities based upon patterns of action, known indicators, and methodical intrusion phases. Such a model was hypothesized by many researchers to result in the realization of the desired goals of SIEM software. 

The focus of this thesis is the implementation of a "kill chain" framework within SIEM software. A novel "Kill chain" model was developed and implemented within a commercial SIEM system through modifications to the existing SIEM database. These modifications resulted in a new log ontology capable of normalizing security sensor data in accordance with modern threat research. New SIEM correlation rules were developed using the novel log ontology compared to existing vendor recommended correlation rules using the default model. The novel log ontology produced promising results indicating improved detection rates, more descriptive security alarms, and a lower number of false positive alarms. These improvements were assessed to provide improved visibility and more efficient investigation processes to security analysts ultimately reducing the mean time required to detect and escalate security incidents. 


SHAUN CHUA

Implementation of a Multichannel Radar Waveform Generator System Controller

When & Where:


317 Nichols Hall

Committee Members:

Carl Leuschen, Chair
Chris Allen
Fernando Rodriguez-Morales


Abstract

Waveform generation is crucial in a radar system operation. There is a recent need for an 8 channel transmitter with high bandwidth chirp signals (100 MHz – 600 MHz). As such, a waveform generator (WFG) hardware module is required for this purpose. The WFG houses 4 Direct Digital Synthesizers (DDS), and an ALTERA Cyclone V FPGA that acts as its controller. The DDS of choice is the AD9915, because its Digital to Analog Converter can be clocked at a maximum rate of 2.5 GHz, allowing plenty of room to produce the high bandwidth and high frequency chirp signals desired, and also because it supports synchronization between multiple AD9915s. 

The brains behind the DDS operations are the FPGA and the radar software developed in NI LabVIEW. These two aspects of the digital systems grants the WFG highly configurable waveform capabilities. The configurable inputs that can be controlled by the user include: number of waveforms in a playlist, start and stop frequency (bandwidth of chirp signal), zero-pi mode, and waveform amplitude and phase control. 

The FPGA acts as a DDS controller that directly configures and control the DDS operations, while also managing and synchronizing the operations of all DDS channels. This project details largely the development of such a controller, named Multichannel Waveform Generator (MWFG) Controller, and the necessary modifications and development in the NI LabVIEW software, so that they complement each other.


DEEPIKA KOTA

Automatic Color Detection of Colored Wires In Electric Cables

When & Where:


2001B Eaton Hall

Committee Members:

Jim Stiles, Chair
Ron Hui
James Rowland


Abstract

An automatic Color detection system checks for the sequence of colored wires in electric cables which are ready to get crimped together. The system inspects for flat connectors with differs in type and number of wires.This is managed in an automatic way with a self learning system without any requirement of manual input from the user to load new data to the machine. The system is coupled to a connector crimping machine and once the system learns the actual sample of cable order , it automatically inspects each cable assembled by the machine. There are three methodologies based on which this automatic detection takes place 1) A self learning system 2) An algorithm for wire segmentation to extract colors from the captured images 3) An algorithm for color recognition to cope up with wires with different illuminations and insulation .The main advantage of this system is when the cables are produced in large batches ,it provides high level of accuracy and prevents false negatives in order to guarantee defect free production.


MOHAMMED ZIAUDDIN

Open Source Python Widget Application to Synchronize Bibliographical References Between Two BibTeX Repositories

When & Where:


246 Nichols Hall

Committee Members:

Andy Gill, Chair
Perry Alexander
Prasad Kulkarni


Abstract

Bibtex is a tool to edit and manage bibliographical references in a document.Researchers face a common problem that they have one copy of their bibliographical reference databases for a specific project and a master bibliographical database file that holds all their bibliographical references. Syncing these two files is an arduous task as one has to search and modify each reference record individually. Most of the bibtex tools available either provide help in maintaining bibtex bibliographies in different file formats or searching for references in web databases but none of them provide a way to synchronize the fields of the same reference record in the two different bibtex database files. 
The intention of this project is to create an application that helps academicians to keep their bibliographical references in two different databases in sync. We have created a python widget application that employs the Tkinter library for GUI and unQLite database for data storage. This application is integrated with Github allowing users to modify bibtex files present on Github.