Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

David Felton

Optimization and Evaluation of Physical Complementary Radar Waveforms

When & Where:


Nichols Hall, Room 129 (Apollo Auditorium)

Committee Members:

Shannon Blunt, Chair
Rachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata

Abstract

The RF spectrum is a precious, finite resource with ever-increasing demand. Consequently, the mandate to be a "good spectral neighbor" is in direct conflict with the requirements for high-performance sensing where correlation error is fundamentally limited. As such, matched-filter radar performance is often sidelobe-limited with estimation error being constrained by the time-bandwidth (TB) of the collective emission. The methods developed here seek to bridge this gap between idealized radar performance and practical utility via waveform design.    

Estimation error becomes more complex when employing pulse-agility. In doing so, range-sidelobe modulation (RSM) spreads energy across Doppler, rendering traditional methods ineffective. To address this, the gradient-based complementary-FM framework was developed to produce complementary sidelobe cancellation (CSC) after coherently combining subsets within a pulse-agile emission. In contrast to the majority of complementary signals, explored via phase-coding, these Comp-FM waveform subsets achieve CSC while preserving hardware-compatibility since they are FM (though design distortion is never completely avoided). Although Comp-FM addressed practicality via hardware amenability, CSC was localized to zero-Doppler. This work expands the Comp-FM notion to a Doppler-generalized (DG) framework, extending the cancellation condition to an arbitrary span. The same framework can likewise be employed to jointly optimize an entire coherent processing interval (CPI) to minimize RSM within the radar point-spread-function (PSF), thereby generalizing the notion of complementarity and introducing the potential for cognitive operation if sufficient scattering knowledge is available a-priori.          

Sensing with a single emitter is limited by self-inflicted error alone (e.g., clutter, sidelobes), while MIMO systems must additionally contend with the cross-responses from emitters operating concurrently (e.g., simultaneously, spatially proximate, in a shared spectrum), further degrading radar sensitivity. Now, total correlation error is dictated by the overlapping TB (i.e., how coincident are the signals) and number of operating emitters, compounding difficulty to estimate if left unaddressed. As such, the determination of "orthogonal waveforms" comprises a large portion of MIMO literature, though remains a phenomenological misnomer for pulsed emissions. Here, the notion of complementary-FM is applied to a multi-emitter context in which transmitter-amenable quasi-orthogonal subsets, occupying the same spectral band, are produced via a similar gradient-based approach. To further practicalize these MIMO-Comp-FM waveform subsets, the same "DG" approach described above, addressing the otherwise-default Doppler-induced degradation of complementary signals, is applied. In doing so, Doppler-independent separability and complementarity greatly improves estimation sensitivity for multi-emitter systems. 

This MIMO-Comp-FM framework is developed for standard matched filter processing. Coupling this framework with a "DG" form of the previously explored MIMO-MiCRFt is also investigated, illustrating the added benefit of pairing optimized subsets with similarly calibrated processing. 

Each of these methods is developed to address unique and increasingly complex sources of estimation error. All approaches are initially developed and evaluated via simulated analysis where ground-truth is known. Then, despite hardware-induced distortion being unavoidable, the MIMO-Comp-FM framework is confirmed via loopback measurements to preserve the majority of CSC that was observed in simulation. Finally, open-air demonstration of each approach validates practical utility on a radar system.


Hao Xuan

Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge Discovery

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Cuncong Zhong, Chair
Fengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu

Abstract

Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.

These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.

First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.

Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.


Pramil Paudel

Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless Imaging

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Fengjun Li, Chair
Alex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao

Abstract

Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference. 

We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks. 


Past Defense Notices

Dates

JAMIE ROBINSON

Code Cache Management in Managed Language VMs to Reduce Memory Consumption for Embedded Systems

When & Where:


129 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Bo Luo
Heechul Yun


Abstract

The compiled native code generated by a just-in-time (JIT) compiler in managed language virtual machines (VM) is placed in a region of memory called the code cache. Code cache management (CCM) in a VM is responsible to find and evict methods from the code cache to maintain execution correctness and manage program performance for a given code cache size or memory budget. Effective CCM can also boost program speed by enabling more aggressive JIT compilation, powerful optimizations, and improved hardware instruction cache and I-TLB performance.

Though important, CCM is an overlooked component in VMs. We find that the default CCM policies in Oracle’s production-grade HotSpot VM perform poorlyeven at modest memory pressure. We develop a detailed simulation-based framework to model and evaluate the potential efficiency of many different CCM policies in a controlled and realistic, but VM-independent environment. We make the encouraging discovery that effective CCM policies can sustain high program performance even for very small cache sizes.

Our simulation study provides the rationale and motivation to improve CCM strategies in existing VMs. We implement and study the properties of several CCM policies in HotSpot. We find that in spite of working within the bounds of the HotSpot VM’s current CCM sub-system, our best CCM policy implementation in HotSpot improves program performance over the default CCM algorithm by 39%, 41%, 55%, and 50% with code cache sizes that are 90%, 75%, 50%, and 25% of the desired cache size, on average.


AIME DE BERNER

Application of Machine Learning Techniques to the Diagnosis of Vision Disorders

When & Where:


2001B Eaton Hall

Committee Members:

Arvin Agah, Chair
Nicole Beckage
Jerzy Grzymala-Busse


Abstract

In the age of data collection and as we search for knowledge, over time numerous techniques have been developed and used to capture, manipulate, and to process data to acquire the hidden correlations, relations, patterns, and mappings that one may not be able to see. Computers as machines with the help of improved algorithms have proven to provide Artificial Intelligence (AI) by applying models to predict outcomes within an acceptable margin of error. Through performance metrics applied using Data Mining and Machine Learning models to predict human vision disorders, we are able to see promising models. AI techniques used in this work include an improved version of C.45 called C.48, Neuro-Networks, K-Nearest-Neighbor, Random Forest, Support Vector Machines, AdaBoost, among many. The best predictive models were determined that could be applied to the diagnosis of vision disorders, focusing on Strabismus, the need for patient referral to a specialist.


HAO XUE

Understanding Information Credibility in Social Networks

When & Where:


246 Nichols Hall

Committee Members:

Fengjun Li, Chair
Luke Huan
Prasad Kulkarni
Bo Luo
Hyunjin Seo

Abstract

With the advancement of Internet, increasing portions of people's social and communicative activities now take place in the digital world. The growth and popularity of online social networks have tremendously facilitate the online interaction and information exchange. More people now rely online information for news, opinions, and social networking. As the representative of online social-collaborative platforms, online review systems has enabled people to share information effectively and efficiently. A large volume of user generated content is produced daily, which allows people to make reasonable judgments about the quality of service or product of an unknown provider. However, the freedom and ease of of publishing information online has made these systems no longer the sources of reliable information. Not only does biased and misleading information exist, financial incentives drive individual and professional spammers to insert deceptive reviews to manipulate review rating and content. What's worse, advanced Artificial Intelligence has made it possible to generate realistic-looking reviews automatically. In this proposal, we present our work of measuring the credibility of information in online review systems. We first propose to utilize the social relationships and rating deviations to assist the computation of trustworthiness of users. Secondly, we propose a content-based trust propagation framework by extracting the opinions expressed in review content.  The opinion extraction approach we used was a supervised-learning based methods, which has flexibility limitations. Thus, we propose a enhanced framework that not only automates the opinion mining process, but also integrates social relationships with review content. Finally, we propose our study of the credibility of machine-generated reviews.


MOHAMMADREZA HAJIARBABI

A Face Detection and Recognition System for Color Images using Neural Networks with Boosting and Deep Learning

When & Where:


2001B Eaton Hall

Committee Members:

Arvin Agah, Chair
Prasad Kulkarni
Bo Luo
Richard Wang
Sara Wilson*

Abstract

A face detection and recognition system is a biometric identification mechanism which compared to other methods is shown to be more important both theoretically and practically. In principle, the biometric identification methods use a wide range of techniques such as machine learning, computer vision, image processing, pattern recognition and neural networks. A face recognition system consists of two main components, face detection and recognition. 
In this dissertation a face detection and recognition system using color images with multiple faces is designed, implemented, and evaluated. In color images, the information of skin color is used in order to distinguish between the skin pixels and non-skin pixels, dividing the image into several components. Neural networks and deep learning methods has been used in order to detect skin pixels in the image. In order to improve system performance, bootstrapping and parallel neural networks with voting have been used. Deep learning has been used as another method for skin detection and compared to other methods. Experiments have shown that in the case of skin detection, deep learning and neural networks methods produce better results in terms of precision and recall compared to the other methods in this field. 
The step after skin detection is to decide which of these components belong to human face. A template based method has been modified in order to detect the faces. The designed algorithm also succeeds if there are more than one face in the component. A rule based method has been designed in order to detect the eyes and lips in the detected components. After detecting the location of eyes and lips in the component, the face can be detected.
After face detection, the faces which were detected in the previous step are to be recognized. Appearance based methods used in this work are one of the most important methods in face recognition due to the robustness of the algorithms to head rotation in the images, noise, low quality images, and other challenges. Different appearance based methods have been designed, implemented and tested. Canonical correlation analysis has been used in order to increase the recognition rate.


JASON GEVARGIZIAN

Automatic Measurement Framework: Expected Outcome Generation and Measurer Synthesis for Remote Attestation

When & Where:


246 Nichols Hall

Committee Members:

Prasad Kulkarni, Chair
Arvin Agah
Perry Alexander
Andy Gill
Kevin Leonard

Abstract

A system is said to be trusted if it can be unambiguously identified and observed as behaving in accordance with expectations. Remote attestation is a mechanism to establish trust in a remote system.
Remote attestation requires measurement systems that can sample program state from a wide range of applications, each of which with different program features and expected behavior. Even in cases where applications are similar in purpose, differences in attestation critical structures and program variables render any one measurer incapable of sampling multiple applications. Furthermore, any set of behavioral expectations vague enough to match multiple applications would be too weak to serve as a rubric to establish trust in any one of them. As such, measurement functionality must be tailored to each and every critical application on the target system.
Establishing behavioral expectations and customizing measurement systems to gather meaningful data to evidence said expectations is difficult. The process requires an expert, typically the application developer or a motivated appraiser, to analyze the application's source in order to detail program behavioral expectations critical for establishing trust and to identify critical program structures and variables that can be sampled to evidence said trust. This effort required to customize measurement systems manually prohibits widespread adoption of remote attestation in trusted computing.
We propose automatic generation of expected outcomes and synthesis of measurement policies for a configurable general purpose measurer to enable large scale adoption of remote attestation for trusted computing. As such, we mitigate the cost incurred by existing systems that require manual measurement specification and design by an expert sufficiently skilled and knowledgeable regarding the target application and the methods for evidencing trust in the context of remote attestation.


SALLY SAJADIAN

Model Predictive Control of Impedance Source Inverter for Photovoltaic Applications

When & Where:


2001B Eaton Hall

Committee Members:

Reza Ahmadi, Chair
Glenn Prescott
Alessandro Salandrino
Jim Stiles
Huazhen Fang

Abstract

A model predictive controlled power electronics interface (PEI) based on impedance source inverter for photovoltaic (PV) applications is proposed in this work. The proposed system has the capability of operation in both grid-connected and islanded mode. Firstly, a model predictive based maximum power point tracking (MPPT) method is proposed for PV applications based on single stage grid-connected Z-source inverter (ZSI). This technique predicts the future behavior of the PV side voltage and current using a digital observer that estimates the parameters of the PV module. Therefore, by predicting a priori the behavior of the PV module and its corresponding effects on the system, it improves the control efficacy. The proposed method adaptively updates the perturbation size in the PV voltage using the predicted model of the system to reduce oscillations and increase convergence speed. The operation of the proposed method is verified experimentally. The experimental results demonstrate fast dynamic response to changes in solar irradiance level, small oscillations around maximum power point at steady-state, and high MPPT effectiveness from low to high solar irradiance level. The second part of this work focuses on the dual-mode operation of the proposed PEI based on ZSI with capability to operate in islanded and grid-connected mode. The transition from islanded to grid-connected mode and vice versa can cause significant deviation in voltage and current due to mismatch in phase, frequency, and amplitude of voltages. The proposed controller using MPC offers seamless transition between the two modes of operations. The main predictive controller objectives are direct decoupled power control in grid-connected mode and load voltage regulation in islanded mode. The proposed direct decoupled active and reactive power control in grid connected mode enables the dual-mode ZSI to behave as a power conditioning unit for ancillary services such as reactive power compensation. The proposed controller features simplicity, seamless transition between modes of operations, fast dynamic response, and small tracking error in steady state condition of controller objectives. The operation of the proposed system is verified experimentally.


YI JIA

Online Spectral Clustering on Network Streams

When & Where:


December 10, 2012

Committee Members:

Luke Huan, Chair
Swapan Chakrabarti
Jerzy Grzymala-Busse
Bo Luo
Alfred Tat-Kei Ho

Abstract

Graph is an extremely useful representation of a wide variety of practical systems in data analysis. Recently, with the fast accumulation of stream data from various type of networks, significant research interests have arisen on spectral clustering for network streams (or evolving networks). Compared with the general spectral clustering problem, the data analysis of this new type of problems may have additional requirements, such as short processing time, scalability in distributed computing environments, and temporal variation tracking. 

However, to design a spectral clustering method to satisfy these requirements certainly presents non-trivial efforts. There are three major challenges for the new algorithm design. The first challenge is online clustering computation. Most of the existing spectral methods on evolving networks are off-line methods, using standard eigensystem solvers such as the Lanczos method. It needs to re-compute solutions from scratch at each time point. The second challenge is the parallelization of algorithms. To parallelize such algorithms is non-trivial since standard eigen solvers are iterative algorithms and the number of iterations cannot be predetermined. The third challenge is the very limited existing work. In addition, there exists multiple limitations in the existing method, such as computational inefficiency on large similarity changes, the lack of sound theoretical basis, and the lack of effective way to handle accumulated approximate errors and large data variations over time. 

In this thesis, we proposed a new online spectral graph clustering approach with a family of three novel spectrum approximation algorithms. Our algorithms incrementally update the eigenpairs in an online manner to improve the computational performance. Our approaches outperformed the existing method in computational efficiency and scalability while retaining competitive or even better clustering accuracy. We derived our spectrum approximation techniques GEPT and EEPT through formal theoretical analysis. The well-established matrix perturbation theory forms a solid theoretic foundation for our online clustering method. In addition, we discussed our preliminary work on approximate graph mining with evolutionary process, non-stationary Bayesian Network structure learning from non-stationary time series data, and Bayesian Network structure learning with text priors imposed by non-parametric hierarchical topic modeling.


HAYDER ALMOSA

Downlink Achievable Rate Analysis for FDD Massive MIMO Systems

When & Where:


250 Nichols Hall

Committee Members:

Lingjia Liu, Chair
Shannon Blunt
Ron Hui
Erik Perrins
Hongyi Cai

Abstract

Multiple-Input Multiple-Output (MIMO) systems with large-scale transmit antenna arrays, often called massive MIMO, is a very promising direction for 5G due to its ability to increase capacity and enhance both spectrum and energy efficiency. To get the benefit of massive MIMO system, accurate downlink channel state information at the transmitter (CSIT) is essential for downlink beamforming and resource allocation. Conventional approaches to obtain CSIT for FDD massive MIMO systems require downlink training and CSI feedback. However, such training will cause a large overhead for massive MIMO systems because of the large dimensionality of the channel matrix. In this research proposal, we propose an efficient downlink beamforming method to address the challenging of downlink training overhead. First, we design an efficient downlink beamforming method based on partial CSI. By exploiting the relationship between uplink (UL) DoAs and downlink (DL) DoDs, we derive an expression for estimated downlink DoDs, which will be used for downlink beamforming. Second, we derive an efficient downlink beamforming method based on downlink CSIT estimated at the BS. By exploiting the sparsity structure of downlink channel matrix, we develop an algorithm that select the best features from the measurement matrix to obtain efficient CSIT acquisition that can reduce the downlink training overhead compared with the conventional LS/MMSE estimators. In both cases, we compare the performance of our proposed beamforming method with traditional method in terms of downlink achievable rate and simulation results show that our proposed method outperform the traditional beamforming method.


ANDREW OZOR

Size Up: A Tool for Interactive Comparative Collection Analysis for Very Large Species Collections

When & Where:


2001B Eaton Hall

Committee Members:

Jim Miller, Chair
Man Kong
Brian Potetz


Abstract


BRYAN BANZ

A Framework for Model Development Using Dimension Reduction and Low-Cost Surrogate Functions

When & Where:


2001B Eaton Hall

Committee Members:

James Miller, Chair
Arvin Agah
Jerzy Grzymala-Busse
Nancy Kinnersley
John Doveton*

Abstract