Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

David Felton

Optimization and Evaluation of Physical Complementary Radar Waveforms

When & Where:


Nichols Hall, Room 129 (Apollo Auditorium)

Committee Members:

Shannon Blunt, Chair
Rachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata

Abstract

The RF spectrum is a precious, finite resource with ever-increasing demand. Consequently, the mandate to be a "good spectral neighbor" is in direct conflict with the requirements for high-performance sensing where correlation error is fundamentally limited. As such, matched-filter radar performance is often sidelobe-limited with estimation error being constrained by the time-bandwidth (TB) of the collective emission. The methods developed here seek to bridge this gap between idealized radar performance and practical utility via waveform design.    

Estimation error becomes more complex when employing pulse-agility. In doing so, range-sidelobe modulation (RSM) spreads energy across Doppler, rendering traditional methods ineffective. To address this, the gradient-based complementary-FM framework was developed to produce complementary sidelobe cancellation (CSC) after coherently combining subsets within a pulse-agile emission. In contrast to the majority of complementary signals, explored via phase-coding, these Comp-FM waveform subsets achieve CSC while preserving hardware-compatibility since they are FM (though design distortion is never completely avoided). Although Comp-FM addressed practicality via hardware amenability, CSC was localized to zero-Doppler. This work expands the Comp-FM notion to a Doppler-generalized (DG) framework, extending the cancellation condition to an arbitrary span. The same framework can likewise be employed to jointly optimize an entire coherent processing interval (CPI) to minimize RSM within the radar point-spread-function (PSF), thereby generalizing the notion of complementarity and introducing the potential for cognitive operation if sufficient scattering knowledge is available a-priori.          

Sensing with a single emitter is limited by self-inflicted error alone (e.g., clutter, sidelobes), while MIMO systems must additionally contend with the cross-responses from emitters operating concurrently (e.g., simultaneously, spatially proximate, in a shared spectrum), further degrading radar sensitivity. Now, total correlation error is dictated by the overlapping TB (i.e., how coincident are the signals) and number of operating emitters, compounding difficulty to estimate if left unaddressed. As such, the determination of "orthogonal waveforms" comprises a large portion of MIMO literature, though remains a phenomenological misnomer for pulsed emissions. Here, the notion of complementary-FM is applied to a multi-emitter context in which transmitter-amenable quasi-orthogonal subsets, occupying the same spectral band, are produced via a similar gradient-based approach. To further practicalize these MIMO-Comp-FM waveform subsets, the same "DG" approach described above, addressing the otherwise-default Doppler-induced degradation of complementary signals, is applied. In doing so, Doppler-independent separability and complementarity greatly improves estimation sensitivity for multi-emitter systems. 

This MIMO-Comp-FM framework is developed for standard matched filter processing. Coupling this framework with a "DG" form of the previously explored MIMO-MiCRFt is also investigated, illustrating the added benefit of pairing optimized subsets with similarly calibrated processing. 

Each of these methods is developed to address unique and increasingly complex sources of estimation error. All approaches are initially developed and evaluated via simulated analysis where ground-truth is known. Then, despite hardware-induced distortion being unavoidable, the MIMO-Comp-FM framework is confirmed via loopback measurements to preserve the majority of CSC that was observed in simulation. Finally, open-air demonstration of each approach validates practical utility on a radar system.


Hao Xuan

Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge Discovery

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Cuncong Zhong, Chair
Fengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu

Abstract

Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.

These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.

First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.

Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.


Pramil Paudel

Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless Imaging

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Fengjun Li, Chair
Alex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao

Abstract

Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference. 

We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks. 


Past Defense Notices

Dates

SUSANNA MOSLEH

Intelligent Interference Mitigation for Multi-cell Multi-user MIMO Networks with Limited Feedback

When & Where:


250 Nichols Hall

Committee Members:

Lingjia Liu, Chair
Victor Frost
Ron Hui
Erik Perrins
Jian Li

Abstract

Nowadays, wireless communication are becoming so tightly integrated in our daily lives, especially with the global spread of laptops, tablets and smartphones. This has paved the way to dramatically increasing wireless network dimensions in terms of subscribers and amount of flowing data. With the rapidly growing data traffic, interference has become a major limitation in wireless networks. To deal with this issue and in order to increase the spectral efficiency of wireless networks, various interference mitigation techniques have been suggested among which interference alignment (IA) has been shown to significantly improve network performance. However, how to practically use IA to mitigate inter-cell interference in a downlink multi-cell multi-user MIMO networks still remains an open problem. Besides, more recently, the attention of researchers has been drawn to a new technique for improving the spectral efficiency, namely, massive/full dimension multiple-input multiple-output. Although massive MIMO/FD-MIMO brings a large diversity gain to the network, its practical implementation poses a research challenge. Moreover, new techniques that can mitigate interference impact in such systems remain unexplored. To address these challenges, this proposed research targets to 1) develop an IA technique for downlink multi-cell multi-user MIMO networks; 2) mathematically characterize the performance of IA with limited feedback; and 3) evaluate the performance analysis of IA technique (with/without limited feedback) in massive MIMO/FD-MIMO networks. Preliminary results show that IA with limited feedback significantly increase the spectral efficiency of downlink multi-cell multi-user MIMO networks.


RACHAD ATAT

Enabling Cyber-Physical Communication in 5G Cellular Networks: Challenges, Solutions and Applications

When & Where:


246 Nichols Hall

Committee Members:

Lingjia Liu, Chair; Yang Yi, Co-Chair , Chair
Shannon Blunt
Jim Rowland
James Sterbenz
Jin Feng

Abstract

Cyber-physical systems (CPS) are expected to revolutionize the world through a myriad of applications in health-care, disaster event applications, environmental management, vehicular networks, industrial automation, and so on. The continuous explosive increase in wireless data traffic, driven by the global rise of smartphones, tablets, video streaming, and online social networking applications along with the anticipated wide massive sensors deployments, will create a set of challenges to network providers, especially that future fifth generation (5G) cellular networks will help facilitate the enabling of CPS communications over current network infrastructure. 
In this dissertation, we first provide an overview of CPS taxonomy along with its challenges from energy efficiency, security, and reliability. Then we present different tractable analytical solutions through different 5G technologies, such as device-to-device (D2D) communications, cell shrinking and offloading, in order to enable CPS traffic over cellular networks. These technologies also provide CPS with several benefits such as ubiquitous coverage, global connectivity, reliability and security. By tuning specific network parameters, the proposed solutions allow the achievement of balance and fairness in spectral efficiency and minimum achievable throughout among cellular users and CPS devices. To conclude, we present a CPS mobile-health application as a case study where security of the medical health cyber-physical space is discussed in details. 


HAO CHEN

Mutual Information Accumulation over Wireless Networks:Fundamentals and Applications

When & Where:


250 Nichols Hall

Committee Members:

Lingjia Liu, Chair
Shannon Blunt
Victor Frost
Yang Yi
Zsolt Talata

Abstract

Future wireless networks will face a compound challenge of supporting large traffic volumes, providing ultra-reliable and low latency connections to ultra-dense mobile devices. To meet this challenge, various new technologies have been introduced among which mutual-information accumulation (MIA), an advanced physical (PHY) layer coding technique, has been shown to significantly improve the network performance. Since the PHY layer is the fundamental layer, MIA could potentially impact various network layers of a wireless network. Accordingly, the understanding of improving network design based on MIA is far from being fully developed. The purpose of this dissertation is to study the fundamental performance improvement of MIA over wireless networks and to apply these fundamental results to guide the design of practical systems, such as cognitive radio networks and massive machine type communication networks. 
This dissertation includes three parts. The first part of this dissertation presents the fundamental analysis of the performance of MIA over wireless networks. To begin with, we first analyze the asymptotic performance of MIA in an infinite 2-dimensional(2-D) grid network. Then, we investigate the optimal routing in cognitive radio networks with MIA and derive the closed-form cooperative gain obtained by applying MIA in cognitive radio networks. Finally, we characterize the performance of MIA in random networks using tools from stochastic geometry. 
The second and third part of this dissertation focuses on the application of MIA in cognitive radio networks and massive machine type communication networks. An optimization problem is formulated to identify the cooperative routing and optimal resources allocation to minimize the transmission delay in underlay cognitive radio networks with MIA. Efficient centralized as well as distributed algorithms are developed to solve this cross-layer optimization problem using the fundamental properties obtained in the first part of this dissertation. A new cooperative retransmission strategy is developed for massive MTC networks with MIA. Theoretical analysis of the new developed retransmission strategy is conducted using the same methodology developed in the fundamental part of this dissertation. Monte Carlo simulation results and numerical results are presented to verify our analysis as well as to show the performance improvement of our developed strategy. 

 

 


HAMID MAHMOUDI

Modulated Model Predictive Control for Power Electronic Converters

When & Where:


2001B Eaton Hall

Committee Members:

Reza Ahmadi, Chair
Chris Allen
Glenn Prescott
Alessandro Salandrino
Jim Stiles

Abstract

Advanced switching algorithms and modulation methods for power electronics converters controlled with model predictive control (MPC) strategies have been proposed in this work. The methods under study retain the advantage of conventional MPC methods in programing the nonlinear effects of the converter into the design calculations to improve the overall dynamic and steady state performance of the system and builds upon that by offering new modulation technique for MPC to minimize the voltage and current ripples through using a fixed switching frequency. The proposed method is easy to implement and provides flexibility to prioritize different objectives of the system against each other using the objective weighting factor. To demonstrate the effectiveness of the proposed method, it has been used to overcome the stability problems caused by a constant power load (CPL) in a multi converter system as a case study. 
In addition, to further evaluate the merits of the proposed method, it has been used to control modular multilevel converters (MMCs) in voltage source converter-high voltage DC (VSC-HVDC) systems. The proposed method considers the nonlinear properties of the MMC into the design calculations while minimizing the line total harmonic distortion (THD), circulating current ripple and steady-state error by generating modulated switching signals with a fixed switching frequency. In this work, the predictive modeling of the MMC is provided. Next, the proposed control method is described. Then, the application of the proposed method to a MMC system is detailed. Experimental results from the systems under study illustrate the effectiveness of proposed strategies. 


MD AMIMUL EHSAN

Enabling Technologies for 3D ICs: TSV Modeling and Analysis

When & Where:


246 Nichols Hall

Committee Members:

Yang Yi, Chair
Ron Hui
Lingjia Liu
Alessandro Salandrino
Judy Wu

Abstract

Through silicon via (TSV) based three-dimensional (3D) integrated circuit (IC) aims to stack and interconnect dies or wafers vertically. This forefront technology offers a promising near-term solution for further miniaturization and the performance improvement of electronic systems and follows a more than Moore strategy. 
Along with the need for low-cost and high-yield process technology, the successful application of TSV technology requires further optimization of the TSV electrical modeling and design. In the millimeter wave (mmW) frequency range, the root mean square (rms) height of the TSV sidewall roughness is comparable to the skin depth and hence becomes a critical factor for TSV modeling and analysis. The impact of TSV sidewall roughness on electrical performance, such as the loss and impedance alteration in the mmW frequency range, is examined and analyzed following the second order small perturbation method. Then, an accurate and efficient electrical model for TSVs has been proposed considering the TSV sidewall roughness effect, the skin effect, and the metal oxide semiconductor (MOS) effect. 
However, the emerging application of 3D integration involves an advanced bio-inspired computing system which is currently experiencing an explosion of interest. In neuromorphic computing, the high density membrane capacitor plays a key role in the synaptic signaling process, especially in the spike firing analog implementation of neurons. We proposed a novel 3D neuromorphic design architecture in which the redundant and dummy TSVs are reconfigured as membrane capacitors. This modification has been achieved by taking advantage of the metal insulator semiconductor (MIS) structure along the sidewall, strategically engineering the fixed oxide charges in depletion region surrounding the TSVs, and the addition of oxide layer around the bump without changing any process technology. Without increasing the circuit area, this reconfiguration of TSVs can result in substantial power consumption reduction and a significant boost to chip performance and efficiency. Also, depending on the availability of the TSVs, we proposed a novel CAD framework for TSV assignments based on the force-directed optimization and linear perturbation. 


SANTOSH MALYALA

Estimation of Ice Basal Reflectivity of Byrd Glacier using RES Data

When & Where:


317 Nichols Hall

Committee Members:

Carl Leuschen, Chair
Jilu Li
Chris Allen
John Paden

Abstract

Ice basal reflectivity is much needed for the determination of ice basal conditions and for the accurate modeling of ice sheet to estimate the future global mean sea level rise. Reflectivity values can be determined from the received radio echo sounding data if the power loss caused by different components along the two-way transmission of EM wave are accurately compensated. 

For the large volume of received radio echo sounding data collected over Byrd glacier in 2011-2012 with multichannel radar, the spherical spreading loss caused due to two-way propagation, power reduction due to roughness and relative englacial attenuation are compensated to estimate the relative reflectivity values of the Byrd glacier. 
In order to estimate the scattered incoherent power component due to roughness, the distributions of echo amplitudes returned from air-firn interface and from ice – bed interface are modeled to estimate RMS height variations. The englacial attenuation rate of wave for two-way propagation along the ice depth is modeled using the observed data. The estimated air-firn interface roughness parameters are relatively cross verified using the Neal’s method and with the correlations from the Landsat image mosaic of Antarctica. Estimated relative basal reflectivity values are validated using the cross-over analysis and abruptness index measurements. From the Byrd relative reflectivity map, the corresponding echograms at the locations of potential subglacial water systems are checked for the observable lake features. 
The obtained results are checked for correlations with previously predicted lake locations and subglacial flow paths. While the results doesn’t exactly match with the previously identified locations with elevation changes, high relative reflectivity values are observed close to those locations, aligning exactly or close to previously predicted flow paths providing a new window into the hydrological network of the glacial. Relative reflectivity values are clustered to indicate the different potential basal conditions beneath the Byrd glacier 


RAVALI GINJUPALLI

A Rule Checker and K-Fold Cross Validation for Incomplete Data Sets

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Gary Minden
Suzanne Shontz


Abstract

Rule induction is an important technique of data mining or machine learning. Knowledge is frequently expressed by rules in many areas of AI, including rule based expert systems. The machine learning/ data mining system LERS (Learning from Examples based on Rough Sets) induces a set of rules from examples and classifies new examples using the set of rules induced previously by LERS. LERS induces rules based on supervised learning. The MLEM2 algorithm is a rule induction algorithm in which rule induction, discretization, and handling missing attribute values are all conducted simultaneously. A rule checker is implemented to classify new cases using the rules induced by MLEM2 algorithm. MLEM2 algorithm induces certain and possible rule sets. Bucket Brigade algorithm is implemented to 
classify new examples. K-fold cross-validation technique is implemented to measure the performance of MLEM2 algorithm. The objective of this project is to find out the efficiency of the MLEM2 rule induction method for incomplete data set. 


DHWANI SAXENA

A Modification of the Characteristic Relation for Incomplete Data Sets

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Perry Alexander
Prasad Kulkarni


Abstract

Rough set theory is a popular approach for decision rule induction. However, it requires the objects in the information system to be completely described. Many real life data sets are incomplete, so we cannot directly apply rough set theory for rule induction. A characteristic relation is used to deal with incomplete information systems in which ‘do not care’ data coexist with lost data. There are scenarios in which two objects that do not have the same known value are indiscernible and on the other hand the two objects which have a lot of equivalent known values are very likely to be in different classes. To rectify such situations, a modification of the characteristic relation was introduced. This project implements rule induction from the modification of the characteristic relation for incomplete data sets.


AHMED SYED

Maximal Consistent Block Technique for Rule Acquisition in Incomplete Information Systems

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Perry Alexander
Prasad Kulkarni


Abstract

In this project, an idea of the maximal consistent block is applied to formulate a new approximation to a concept in incomplete data sets. The maximal consistent blocks have smaller cardinality compared to characteristic sets. Because of this, the generated upper approximations will be smaller in size. Two interpretations of missing attribute values are discussed: lost values and “do not care” conditions. Four incomplete data sets are used for experiments with varying levels of missing information. Maximal Consistent Blocks and Characteristics Sets are compared in terms of cardinality of lower and upper approximations. The next objective is to compare the decision rules induced and cases covered by both techniques. The experiments show that both techniques provide the same lower approximations for all the datasets with “do not care” conditions. The best results are achieved by maximal consistent blocks for upper approximations for three datasets.


AMUKTHA CHAKILAM

A Modified ID3 Algorithm for Continuous Numerical Attributes Using Cut Point Approach

When & Where:


2001B Eaton Hall

Committee Members:

Jerzy Grzymala-Busse, Chair
Perry Alexander
Prasad Kulkarni


Abstract

Data classification is a methodology of data mining used to organize data by relevant categories to obtain meaningful information. A model is generated from the input training set which is used to classify the test data into predetermined groups or classes. One of the most widely used models is a decision tree which uses a tree like structure to list all possible outcomes. Decision tree is an important predictive analysis method in Data Mining as it requires minimum effort from the users for data interpretation. 

This project implements ID3, an algorithm for building decision tree using information gain metric. Furthermore, through illustrating the basic ideas of ID3, this project also addresses the inefficiency of ID3 in handling continuous numerical attributes. A cut point approach is presented to discretize the numeric attributes into discrete intervals and enable ID3 functionality for them. Experiments show that such decision trees contain fewer number of nodes and branches in contrast to a tree obtained by basic ID3 algorithm. This modified algorithm can be used to classify real valued domains containing symbolic and numeric attributes with multiple discrete outcomes.