Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

Md Mashfiq Rizvee

Hierarchical Probabilistic Architectures for Scalable Biometric and Electronic Authentication in Secure Surveillance Ecosystems

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Sumaiya Shomaji, Chair
Tamzidul Hoque
David Johnson
Hongyang Sun
Alexandra Kondyli

Abstract

Secure and scalable authentication has become a primary requirement in modern digital ecosystems, where both human biometrics and electronic identities must be verified under noise, large population growth and resource constraints. Existing approaches often struggle to simultaneously provide storage efficiency, dynamic updates and strong authentication reliability. The proposed work advances a unified probabilistic framework based on Hierarchical Bloom Filter (HBF) architectures to address these limitations across biometric and hardware domains. The first contribution establishes the Dynamic Hierarchical Bloom Filter (DHBF) as a noise-tolerant and dynamically updatable authentication structure for large-scale biometrics. Unlike static Bloom-based systems that require reconstruction upon updates, DHBF supports enrollment, querying, insertion and deletion without structural rebuild. Experimental evaluation on 30,000 facial biometric templates demonstrates 100% enrollment and query accuracy, including robust acceptance of noisy biometric inputs while maintaining correct rejection of non-enrolled identities. These results validate that hierarchical probabilistic encoding can preserve both scalability and authentication reliability in practical deployments. Building on this foundation, Bio-BloomChain integrates DHBF into a blockchain-based smart contract framework to provide tamper-evident, privacy-preserving biometric lifecycle management. The system stores only hashed and non-invertible commitments on-chain while maintaining probabilistic verification logic within the contract layer. Large-scale evaluation again reports 100% enrollment, insertion, query and deletion accuracy across 30,000 templates, therefore, solving the existing problem of blockchains being able to authenticate noisy data. Moreover, the deployment analysis shows that execution on Polygon zkEVM reduces operational costs by several orders of magnitude compared to Ethereum, therefore, bringing enrollment and deletion costs below $0.001 per operation which demonstrate the feasibility of scalable blockchain biometric authentication in practice. Finally, the hierarchical probabilistic paradigm is extended to electronic hardware authentication through the Persistent Hierarchical Bloom Filter (PHBF). Applied to electronic fingerprints derived from physical unclonable functions (PUFs), PHBF demonstrates robust authentication under environmental variations such as temperature-induced noise. Experimental results show zero-error operation at the selected decision threshold and substantial system-level improvements as well as over 10^5 faster query processing and significantly reduced storage requirements compared to large scale tracking.


Fatima Al-Shaikhli

Optical Measurements Leveraging Coherent Fiber Optics Transceivers

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Rongqing Hui, Chair
Shannon Blunt
Shima Fardad
Alessandro Salandrino
Judy Wu

Abstract

Recent advancements in optical technology are invaluable in a variety of fields, extending far beyond high-speed communications. These innovations enable optical sensing, which plays a critical role across diverse applications, from medical diagnostics to infrastructure monitoring and automotive systems. This research focuses on leveraging commercially available coherent optical transceivers to develop novel measurement techniques to extract detailed information about optical fiber characteristics, as well as target information. Through this approach, we aim to enable accurate and fast assessments of fiber performance and integrity, while exploring the potential for utilizing existing optical communication networks to enhance fiber characterization capabilities. This goal is investigated through three distinct projects: (1) fiber type characterization based on intensity-modulated electrostriction response, (2) coherent Light Detection and Ranging (LiDAR) system for target range and velocity detection through different waveform design, including experimental validation of frequency modulation continuous wave (FMCW) implementations and theoretical analysis of orthogonal frequency division multiplexing (OFDM) based approaches and (3) birefringence measurements using a coherent Polarization-sensitive Optical Frequency Domain Reflectometer (P-OFDR) system.

Electrostriction in an optical fiber is introduced by interaction between the forward propagated optical signal and the acoustic standing waves in the radial direction resonating between the center of the core and the cladding circumference of the fiber. The response of electrostriction is dependent on fiber parameters, especially the mode field radius. We demonstrated a novel technique of identifying fiber types through the measurement of intensity modulation induced electrostriction response. As the spectral envelope of electrostriction induced propagation loss is anti-symmetrical, the signal to noise ratio can be significantly increased by subtracting the measured spectrum from its complex conjugate. We show that if the field distribution of the fiber propagation mode is Gaussian, the envelope of the electrostriction-induced loss spectrum closely follows a Maxwellian distribution whose shape can be specified by a single parameter determined by the mode field radius.        

We also present a self-homodyne FMCW LiDAR system based on a coherent receiver. By using the same linearly chirped waveform for both the LiDAR signal and the local oscillator, the self-homodyne coherent receiver performs frequency de-chirping directly in the photodiodes, significantly simplifying signal processing. As a result, the required receiver bandwidth is much lower than the chirping bandwidth of the signal. Simultaneous multi-target of range and velocity detection is demonstrated experimentally. Furthermore, we explore the use of commercially available coherent transceivers for joint communication and sensing using OFDM waveforms.

In addition, we demonstrate a P-OFDR system utilizing a digital coherent optical transceiver to generate a linear frequency chirp via carrier-suppressed single-sideband modulation. This method ensures linearity in chirping and phase continuity of the optical carrier. The coherent homodyne receiver, incorporating both polarization and phase diversity, recovers the state of polarization (SOP) of the backscattered optical signal along the fiber, mixing with an identically chirped local oscillator. With a spatial resolution of approximately 5 mm, a 26 GHz chirping bandwidth, and a 200 us measurement time, this system enables precise birefringence measurements. By employing three mutually orthogonal SOPs of the launched optical signal, we measure relative birefringence vectors along the fiber.


Past Defense Notices

Dates

AKSHATHA RAO

Fountain codes

When & Where:


250 Nichols Hall

Committee Members:

Erik Perrins, Chair
Shannon Blunt
Victor Frost
Jonathan Brumberg

Abstract

Fountain codes are forward-error correcting codes suitable for erasure channels. A binary erasure channel is a memoryless channel, in which the symbols are either transmitted correctly or they are erased. The advantage of fountain codes is that it requires few encoded symbols for decoding. The source symbols can be decoded using any set of encoded symbols. Since fountain codes are rateless, they can adapt to changing channel conditions. They are beneficial for broadcasting and multicasting applications where channels have different erasure probability. 
The project involves the implementation of two different fountain codes: LT code and Raptor code. 
The goal of the project is to measure the performance of the code based on how many encoded symbols are required for successful decoding. The encoders and decoders for the two codes are designed in Matlab. The number of encoded symbols required for decoding of the source symbols for different degree distributions are plotted. 


QI SHI

Application of Split-Step Fourier Method and Gaussian Noise Model in the Calculation of Nonlinear Interference in Uncompensated Optical Coherent WDM System

When & Where:


246 Nichols Hall

Committee Members:

Ron Hui, Chair
Chris Allen
Erik Perrins


Abstract

Wavelength division multiplexing (WDM) is a technology of combining a number of independent information-carrying signals with different wavelengths into the same fiber. This enables us to transmit several channels of high quality, large capacity optical signals in only one fiber simultaneously. WDM is the most popular long distance transmission solution nowadays, which is widely utilized in terrestrial backbone and intercontinental undersea fiber optics transmission system. Extremely effective and efficient analysis method of WDM system is always indispensable due to two reasons. In the first place, the deployment of WDM system is usually a time and money consuming project so that an accurate design is always required before construction. Secondly, optical network routing protocol is based on expeditious and veracious real-time evaluation and prediction of network performance. Two main distinct phenomena affecting the overall WDM system performance are amplified spontaneous emission (ASE) noise accumulation and nonlinear interference (NLI) due to the Kerr effect. The ASE noise has already been well understood but the calculation of NLI is complicated. A popular way called Split-Step Fourier (SSF) method, which directly solves the nonlinear propagation equation numerically is widely used to understand the pulse propagation in nonlinear dispersive media. Though the SSF method can provide an accurate result of NLI, its high computation expense prohibits satisfying the efficiency requirement mentioned above. Fortunately, Gaussian Noise (GN) model, which to a large extent resolves this issue has been proposed and developed in recent years.


RAKSHA GANESH

Structured-Irregular Repeat Accumulate Codes

When & Where:


250 Nichols Hall

Committee Members:

Erik Perrins, Chair
Shannon Blunt
Ron Hui


Abstract

There is a strong need for efficient and reliable communication systems in the present day context. To design an efficient transmission system the errors that occur during transmission should be minimized. This can be achieved by channel encoding. The Irregular repeat accumulate codes are a class of serially concatenated codes that have a linear encoding algorithm, flexibility in code rate and code lengths and good performance. 

Here we implement a design technique for Structured Irregular repeat accumulate codes. S-IRA codes can be decoded reliably using the iterative log likelihood decoding (sum-product) algorithm at low error rates. We perform encoding, decoding and performance analysis of S-IRA codes of different code rates and code word lengths and compare their performances on the AWGN channel. In this project we also design codes with different column weights for the parity check matrices and compare their performances on the AWGN channel with the already designed codes. 


MADHURI MORLA

Effect of SOA Nonlinearities on CO-OFDM System

When & Where:


2001B Eaton Hall

Committee Members:

Ron Hui, Chair
Victor Frost
Erik Perrins


Abstract

The use of Semiconductor Optical Amplifier (SOA) for amplification in Coherent Optical-Orthogonal Frequency Division Multiplexing (CO-OFDM) system has been of interest in recent studies. The gain saturation of SOA induces inter-channel crosstalk. This effect is analyzed by simulation and compared with some recent experimental results. Performance of the optical transmission system is measured using Error Vector Magnitude (EVM) which is the measure of deviation of received symbols from their ideal positions in the constellation diagram. EVM as a function of input power to SOA is investigated. Improvement in EVM has been observed in linear region with the increase of input power due to the increase of Optical Signal to Noise Ratio (OSNR). In the nonlinear region, increase of the input optical power to SOA results in the degradation of EVM due to the nonlinear saturation of SOA The effect of gain saturation on EVM as a function of number of subcarriers is investigated. 
The relation between different evaluation techniques like Bit Error Rate (BER), SNR and EVM is also presented. EVM is analytically estimated from OSNR by considering the ideal case of additive white Gaussian noise (AWGN) without nonlinearities. Bit Error Rate (BER) is estimated from the analytical and simulated EVM. The role of Peak to Average Power Ratio (PAPR) in the degradation of EVM in the nonlinear region is also studied through numerical simulation. 


SAMBHAV SETHIA

Sentiment Analysis on Wikipedia People Pages Using Enhanced Naive Bayes Model

When & Where:


246 Nichols Hall

Committee Members:

Fengjun Li, Chair
Bo Luo
Jerzy Grzymala-Busse
Prasad Kulkarni

Abstract

Sentiment Analysis involves capturing the viewpoint or opinion expressed by the people on various objects. These objects are diverse set of things like a movie, an article, a person of interest, a product, basically anything on which we can opine about. The opinions that are expressed can take different forms, like a review of a movie, feedback on a product, an article in a newspaper expressing the sentiment of the author on the given topic or even a Wikipedia page on a person. The key challenge of sentiment analysis is to classify the underlying text to the correct class i.e., positive, negative or neutral. Sentiment analysis also deals with the computational treatment of opinion, sentiment and the subjectivity in a text. 
Wikipedia provides a large repository of pages of people around the world. This project conducts large scale experiment using one of the popular sentiment analysis tools, which is modeled on an enhanced version of Naïve Bayes. Here a sentence by sentence sentiment analysis is done for each biographical page retrieved from Wikipedia. The overall sentiment of a person is then calculated by taking an average of every sentiment value of all the sentences related to that particular person. There are advantages of doing this type of analysis. First, the results obtained are better calibrated on a decimal scale which provides a clearer distinction about the sentiment value associated with the individual as compared to the standard result provided by the tool which is based on tri-scale i.e., positive, negative and neutral. Second, this will allow us to understand statistically the viewpoint of Wikipedia on those people. Finally, this project enables us to perform large-scale temporal and geographical analysis, e.g., examine the overall sentiment associated with the people of each state, and thus helping us to analyze the opinion trend. 


XIAOMENG SU

A comparison of the Quality of Rule Induction from Inconsistent Data sets and Incomplete Data Sets

When & Where:


246 Nichols Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Prasad Kulkarni
Zongbo Wang


Abstract

In data mining, decision rules inducted from known examples are used to classify unseen cases. There are various rule induction algorithms, such as LEM1 (Learning from Examples Module version 1), LEM2 (Learning from Examples Module version 2) and MLEM2 (Modified Learning from Examples Module version 2). In the real world, many data sets are imperfect, either inconsistent or incomplete. The idea of lower and upper approximations, or more generally, the probabilistic approximation, provides an effective way to induct rules from inconsistent data sets and incomplete data sets. But the accuracy of rule sets inducted from imperfect data sets are expected to be lower. The objective of this project is to investigate which kind of imperfect data sets (inconsistent or incomplete) is worse in terms of the quality of inducted rule set. In this project, experiments were conducted on eight inconsistent data sets and eight incomplete data sets with lost values. We implemented the MLEM2 algorithm to induct certain and possible rules from inconsistent data sets, and implemented the local probabilistic version of MLEM2 algorithm to induct certain and possible rules from incomplete data sets. A program called Rule Checker was also developed to classify unseen cases with inducted rules and measure the classification error rate. Ten-cross fold validation was carried out and the average error rate was used as the criteria for comparison. The Mann-Whitney nonparametric test was performed to compare, separately for certain and possible rules, incompleteness with inconsistency. The results show that there is no significant difference between inconsistent and incomplete data sets in terms of the quality of rule induction.


SIVA PRAMOD BOBBILI

Static Disassembly of Binary using Symbol Table Information

When & Where:


250 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Andy Gill
Jerzy Grzymala-Busse


Abstract

Static binary analysis is an important challenge with many applications in security and performance optimization. One of the main challenges with analyzing an executable file statically is to discover all the instructions in the binary executable. It is often difficult to discover all program instructions due to a well-known limitation in static binary analysis, called the code discovery problem. Some of the main contributors to the code discovery problem are variable length CISC instructions, data interspersed with code, padding bytes for branch target alignment and indirect jumps. All these problems manifest themselves in x86 binary files, which is unfortunate since x86 is the most popular architecture format in desktop and server domains. 
Although much of the research work in the recent times have stated that the symbol table might be of help to overcome the difficulties of code discovery, the extent to which it can actually help in the code discovery problem is still in question. This work focuses on assessing the benefit of using the symbol table information to overcome the limitations of the code discovery problem and identify more or all instructions in x86 binary executable files. We will discuss the details, extent, limitations and impact of instruction discovery with and without symbol table information in this work. 


JONATHAN LUTES

SafeExit: Exit Node Protection for TOR

When & Where:


2001B Eaton Hall

Committee Members:

Bo Luo, Chair
Arvin Agah
Prasad Kulkarni


Abstract

TOR is one of the most important networks for providing anonymity over the internet. However, in some cases its exit node operators open themselves up to various legal challenges, a fact which discourages participation in the network. In this paper, we propose a mechanism for allowing some users to be voluntarily verified by trusted third parties, providing a means by which an exit node can verify that they are not the true source of traffic. This is done by extending TOR’s anonymity model to include 
another class of user, and using a web of trust mechanism to create chains of trust. 


KAVYASHREE PILAR

Digital Down Conversion and Compression of Radar Data

When & Where:


317 Nichols Hall

Committee Members:

Carl Leuschen, Chair
Shannon Blunt
Glenn Prescott


Abstract

Storage and handling of huge amount of received data samples is one of the major challenges associated with Radar system design. Radar data samples potentially have high temporal and spatial correlation depending on the target characteristics and radar settings. This correlation can be utilized to compress them without any loss in sensitivity in post processed products. This project focuses on reducing the storage requirement of a Radar used for remote sensing of ice sheets. At the front-end of Radar receiver, the data sample rate can be reduced at real-time by performing frequency down-conversion and decimation of the incoming data. The decimated signal can be further compressed by applying suitable data compression algorithm. The project implements a digital down-converter, decimator and a data compression module on FPGA. Literature survey suggests that there are quite a few research works being done towards developing customized Radar data compression algorithms. This project analyses the possibility of using general-purpose algorithms like GZIP, JPEG-2000 (lossless) to compress Radar data. It also considers a simple floating point compression technique to convert 16 bit data to 8 bit data, guaranteeing a 50% reduction in data size. The project implements the 16-to-8 bit conversion, JPEG 2000 lossless and GZIP algorithms in Matlab and compares their SNR performance with Radar data. Simulations suggest that all of them have similar SNR performance but JPEG 2000, GZIP algorithms offer a compression ratio of over 90%. However, 16-to-8-bit compression is implemented in this project because of its simplicity. 
A hardware test bed is implemented to integrate the digital radar electronics with the Matlab Simulink Simulation tools in a hardware in the loop (HIL) configuration. The digital down converter, decimator and the data compression module are prototyped on SimuLink. The design is later implemented on FPGA using Verilog code. The functionality is tested at various stages of development using ModelSim simulations, Altera DSPBuilder’s HDL import, HIL co-simulation and using SignalTap. This test bed can also be used for future development efforts. 


SURYA TEJ NIMMAKAYALA

Exploring Causes of Performance Overhead during Dynamic Binary Translation

When & Where:


250 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Fengjun Li
Bo Luo


Abstract

Dynamic Binary Translators (DBT) have applications ranging from program 
portability, instrumentation, optimizations, and improving software security. To achieve these goals and maintain control over the application's execution, DBTs translate and run the original source/guest programs in a sand-boxed environment. DBT systems apply several optimization techniques like code caching, trace creation, etc. to reduce the translation overhead and 
enhance program performance at run-time. However, even with these 
optimizations, DBTs typically impose a significant performance overhead, 
especially for short-running applications. This performance penalty has 
restricted the more wide-spread adoption of DBT technology, in spite of its obvious need. 

The goal of this work is to determine the different factors that contribute to the performance penalty imposed by dynamic binary translators. In this thesis, we describe the experiments that we designed to achieve our goal and report our results and observations. We use a popular and sophisticated DBT, DynamoRio, for our test platform, and employ the industry-standard SPEC CPU2006 benchmarks to capture run-time statistics. Our experiments find that DynamoRio executes a large number of additional instructions when compared to the native application execution. We further measure that this increase in the number of executed instructions is caused by the DBT frequently exiting 
the code cache to perform various management tasks at run-time, including 
code translation, indirect branch resolution and trace formation. We also 
find that the performance loss experienced by the DBT is directly 
proportional to the number of code cache exits. We will discuss the details on the experiments, results, observations, and analysis in this work.