Defense Notices
All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.
Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.
Upcoming Defense Notices
David Felton
Optimization and Evaluation of Physical Complementary Radar WaveformsWhen & Where:
Nichols Hall, Room 129 (Apollo Auditorium)
Committee Members:
Shannon Blunt, ChairRachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata
Abstract
The RF spectrum is a precious, finite resource with ever-increasing demand. Consequently, the mandate to be a "good spectral neighbor" is in direct conflict with the requirements for high-performance sensing where correlation error is fundamentally limited. As such, matched-filter radar performance is often sidelobe-limited with estimation error being constrained by the time-bandwidth (TB) of the collective emission. The methods developed here seek to bridge this gap between idealized radar performance and practical utility via waveform design.
Estimation error becomes more complex when employing pulse-agility. In doing so, range-sidelobe modulation (RSM) spreads energy across Doppler, rendering traditional methods ineffective. To address this, the gradient-based complementary-FM framework was developed to produce complementary sidelobe cancellation (CSC) after coherently combining subsets within a pulse-agile emission. In contrast to the majority of complementary signals, explored via phase-coding, these Comp-FM waveform subsets achieve CSC while preserving hardware-compatibility since they are FM (though design distortion is never completely avoided). Although Comp-FM addressed practicality via hardware amenability, CSC was localized to zero-Doppler. This work expands the Comp-FM notion to a Doppler-generalized (DG) framework, extending the cancellation condition to an arbitrary span. The same framework can likewise be employed to jointly optimize an entire coherent processing interval (CPI) to minimize RSM within the radar point-spread-function (PSF), thereby generalizing the notion of complementarity and introducing the potential for cognitive operation if sufficient scattering knowledge is available a-priori.
Sensing with a single emitter is limited by self-inflicted error alone (e.g., clutter, sidelobes), while MIMO systems must additionally contend with the cross-responses from emitters operating concurrently (e.g., simultaneously, spatially proximate, in a shared spectrum), further degrading radar sensitivity. Now, total correlation error is dictated by the overlapping TB (i.e., how coincident are the signals) and number of operating emitters, compounding difficulty to estimate if left unaddressed. As such, the determination of "orthogonal waveforms" comprises a large portion of MIMO literature, though remains a phenomenological misnomer for pulsed emissions. Here, the notion of complementary-FM is applied to a multi-emitter context in which transmitter-amenable quasi-orthogonal subsets, occupying the same spectral band, are produced via a similar gradient-based approach. To further practicalize these MIMO-Comp-FM waveform subsets, the same "DG" approach described above, addressing the otherwise-default Doppler-induced degradation of complementary signals, is applied. In doing so, Doppler-independent separability and complementarity greatly improves estimation sensitivity for multi-emitter systems.
This MIMO-Comp-FM framework is developed for standard matched filter processing. Coupling this framework with a "DG" form of the previously explored MIMO-MiCRFt is also investigated, illustrating the added benefit of pairing optimized subsets with similarly calibrated processing.
Each of these methods is developed to address unique and increasingly complex sources of estimation error. All approaches are initially developed and evaluated via simulated analysis where ground-truth is known. Then, despite hardware-induced distortion being unavoidable, the MIMO-Comp-FM framework is confirmed via loopback measurements to preserve the majority of CSC that was observed in simulation. Finally, open-air demonstration of each approach validates practical utility on a radar system.
Hao Xuan
Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge DiscoveryWhen & Where:
Nichols Hall, Room 246 (Executive Conference Room)
Committee Members:
Cuncong Zhong, ChairFengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu
Abstract
Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.
These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.
First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.
Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.
Pramil Paudel
Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless ImagingWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Fengjun Li, ChairAlex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao
Abstract
Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference.
We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks.
Past Defense Notices
PATRICK McCORMICK
Design and Optimization of Physical Waveform-Diverse EmissionsWhen & Where:
246 Nichols Hall
Committee Members:
Shannon Blunt, ChairChris Allen
Alessandro Salandrino
Jim Stiles
Emily Arnold*
Abstract
With the advancement of arbitrary waveform generation techniques, new radar transmission modes can be designed via precise control of the waveform's time-domain signal structure. The finer degree of emission control for a waveform (or multiple waveforms via a digital array) presents an opportunity to reduce ambiguities in the estimation of parameters within the radar backscatter. While this freedom opens the door to new emission capabilities, one must still consider the practical attributes for radar waveform design. Constraints such as constant amplitude (to maintain sufficient power efficiency) and continuous phase (for spectral containment) are still considered prerequisites for high-powered radar waveforms. These criteria are also applicable to the design of multiple waveforms emitted from an antenna array in a multiple-input multiple-output (MIMO) mode.
In this work, two spatially-diverse radar emission design methods are introduced that provide constant amplitude, spectrally-contained waveforms. The first design method, denoted as spatial modulation, designs the radar waveforms via a polyphase-coded frequency-modulated (PCFM) framework to steer the coherent mainbeam of the emission within a pulse. The second design method is an iterative scheme to generate waveforms that achieve a desired wideband and/or widebeam radar emission. However, a wideband and widebeam emission can place a portion of the emitted energy into what is known as the `invisible' space of the array, which is related to the storage of reactive power that can damage a radar transmitter. The proposed design method purposefully avoids this space and a quantity denoted as the Fractional Reactive Power (FRP) is defined to assess the quality of the result.
The design of FM waveforms via traditional gradient-based optimization methods is also considered. A waveform model is proposed that is a generalization of the PCFM implementation, denoted as coded-FM (CFM), which defines the phase of the waveform via a summation of weighted, predefined basis functions. Therefore, gradient-based methods can be used to minimize a given cost function with respect to a finite set of optimizable parameters. A generalized integrated sidelobe metric is used as the optimization cost function to minimize the correlation range sidelobes of the radar waveform.
RAKESH YELLA
A Comparison of Two Decision Tree Generating Algorithms CART and Modified ID3When & Where:
2001B Eaton Hall
Committee Members:
Jerzy Grzymala-Busse, ChairMan Kong
Prasad Kulkarni
Abstract
In Data mining, Decision Tree is a type of classification model which uses a tree-like data structure to organize the data to obtain meaningful information. We may use Decision Tree for important predictive analysis in data mining.
In this project, we compare two decision tree generating algorithms CART and the modified ID3 algorithm using different datasets with discrete and continuous numerical values. A new approach to handle the continuous numerical values is implemented in this project since the basic ID3 algorithm is inefficient in handling the continuous numerical values. In the modified ID3 algorithm, we discretize the continuous numerical values by creating cut-points. The decision trees generated by the modified algorithm contain fewer nodes and branches compared to basic ID3.
The results from the experiments indicate that there is statistically insignificant difference between CART and modified ID3 in terms of accuracy on test data. On the other hand, the size of the decision tree generated by CART is smaller than the decision tree generated by modified ID3.
SRUTHI POTLURI
A Web Application for Recommending Movies to UsersWhen & Where:
2001B Eaton hall
Committee Members:
Jerzy Grzymala-Busse, ChairMan Kong
Bo Luo
Abstract
Recommendation systems are becoming more and more important with increasing popularity of e-commerce platforms. An ideal recommendation system recommends preferred items to the user. In this project, an algorithm named item-item collaborative filtering is implemented as premise. The recommendations are smarter by going through movies similar to the movies of different ratings by the user, calculating predictions and recommending those movies which have high predictions. The primary goal of the proposed recommendation algorithm is to include user’s preference and to include lesser known items in recommendations. The proposed recommendation system was evaluated on basis of Mean Absolute Error(MAE) and Root Mean Square Error(RMSE) against 1 Million movie rating involving 6040 users and 3900 movies. The implementation is made as a web-application to simulate the real-time experience for the user.
DEBABRATA MAJHI
IRIM: Interesting Rule Induction Module with Handling Missing Attribute ValuesWhen & Where:
2001B Eaton Hall
Committee Members:
Jerzy Grzymala-Busse, ChairPrasad Kulkarni
Bo Luo
Abstract
In the current era of big data, huge amount of data can be easily collected, but the unprocessed data is not useful on its own. It can be useful only when we are able to find interesting patterns or hidden knowledge. The algorithm to find interesting patterns is known as Rule Induction Algorithm. Rule induction is a special area of data mining and machine learning in which formal rules are extracted from a dataset. The extracted rules may represent some general or local (isolated) patterns related to the data.
In this report, we will focus on the IRIM (Interesting Rule Inducing Module) which induces strong interesting rules that covers most of the concept. Usually, the rules induced by IRIM provides interesting and surprising insight to the expert in the domain area.
The IRIM algorithm was implemented using Python and pySpark library, which is specially customize for data mining. Further, the IRIM algorithm was extended to handle the different types of missing data. Then at the end the performance of the IRIM algorithm with and without missing data feature was analyzed. As an example, interesting rules induced from IRIS dataset are shown.
SUSHIL BHARATI
Vision Based Adaptive Obstacle Detection, Robust Tracking and 3D Reconstruction for Autonomous Unmanned Aerial VehiclesWhen & Where:
246 Nichols Hall
Committee Members:
Richard Wang, ChairBo Luo
Suzanne Shontz
Abstract
Vision-based autonomous navigation of UAVs in real-time is a very challenging problem, which requires obstacle detection, tracking, and depth estimation. Although the problems of obstacle detection and tracking along with 3D reconstruction have been extensively studied in computer vision field, it is still a big challenge for real applications like UAV navigation. The thesis intends to address these issues in terms of robustness and efficiency. First, a vision-based fast and robust obstacle detection and tracking approach is proposed by integrating a salient object detection strategy within a kernelized correlation filter (KCF) framework. To increase its performance, an adaptive obstacle detection technique is proposed to refine the location and boundary of the object when the confidence value of the tracker drops below a predefined threshold. In addition, a reliable post-processing technique is implemented for an accurate obstacle localization. Second, we propose an efficient approach to detect the outliers present in noisy image pairs for the robust fundamental matrix estimation, which is a fundamental step for depth estimation in obstacle avoidance. Given a noisy stereo image pair obtained from the mounted stereo cameras and initial point correspondences between them, we propose to utilize reprojection residual error and 3-sigma principle together with robust statistic based Qn estimator (RES-Q) to efficiently detect the outliers and accurately estimate the fundamental matrix. The proposed approaches have been extensively evaluated through quantitative and qualitative evaluations on a number of challenging datasets. The experiments demonstrate that the proposed detection and tracking technique significantly outperforms the state-of-the-art methods in terms of tracking speed and accuracy, and the proposed RES-Q algorithm is found to be more robust than other classical outlier detection algorithms under both symmetric and asymmetric random noise assumptions.
MOHSEN ALEENEJAD
New Modulation Methods and Control Strategies for Power Electronics InvertersWhen & Where:
1 Eaton Hall
Committee Members:
Reza Ahmadi, ChairGlenn Prescott
Alessandro Salandrino
Jim Stiles
Huazhen Fang*
Abstract
The DC to AC power Converters (so-called Inverters) are widely used in industrial applications. The multilevel inverters are becoming increasingly popular in industrial apparatus aimed at medium to high power conversion applications. In comparison to the conventional inverters, they feature superior characteristics such as lower total harmonic distortion (THD), higher efficiency, and lower switching voltage stress. Nevertheless, the superior characteristics come at the price of a more complex topology with an increased number of power electronic switches. The increased number of power electronics switches results in more complicated control strategies for the inverter. Moreover, as the number of power electronic switches increases, the chances of fault occurrence of the switches increases, and thus the inverter’s reliability decreases. Due to the extreme monetary ramifications of the interruption of operation in commercial and industrial applications, high reliability for power inverters utilized in these sectors is critical. As a result, developing simple control strategies for normal and fault-tolerant operation of multilevel inverters has always been an interesting topic for researchers in related areas. The purpose of this dissertation is to develop new control and fault-tolerant strategies for the multilevel power inverter. For the normal operation of the inverter, a new high switching frequency technique is developed. The proposed method extends the utilization of the dc link voltage while minimizing the dv/dt of the switches. In the event of a fault, the line voltages of the faulty inverters are unbalanced and cannot be applied to the three phase loads. For the faulty condition of the inverter, three novel fault-tolerant techniques are developed. The proposed fault-tolerant strategies generate balanced line voltages without bypassing any healthy and operative inverter element, makes better use of the inverter capacity and generates higher output voltage. These strategies exploit the advantages of the Selective Harmonic Elimination (SHE) and Space Vector Modulation (SVM) methods in conjunction with a slightly modified Fundamental Phase Shift Compensation (FPSC) technique to generate balanced voltages and manipulate voltage harmonics at the same time. The proposed strategies are applicable to several classes of multilevel inverters with three or more voltage levels.
XIAOLI LI
Constructivism LearningWhen & Where:
246 Nichols Hall
Committee Members:
Luke Huan, ChairVictor Frost
Bo Luo
Richard Wang
Alfred Ho*
Abstract
Aiming to achieve the learning capabilities possessed by intelligent beings, especially human, researchers in machine learning field have the long-standing tradition of borrowing ideas from human learning, such as reinforcement learning, active learning, and curriculum learning. Motivated by a philosophical theory called "constructivism", in this work, we propose a new machine learning paradigm, constructivism learning. The constructivism theory has had wide-ranging impact on various human learning theories about how human acquire knowledge. To adapt this human learning theory to the context of machine learning, we first studied how to improve leaning performance by exploring inductive bias or prior knowledge from multiple learning tasks with multiple data sources, that is multi-task multi-view learning, both in offline and lifelong setting. Then we formalized a Bayesian nonparametric approach using sequential Dirichlet Process Mixture Models to support constructivism learning. To further exploit constructivism learning, we also developed a constructivism deep learning method utilizing Uniform Process Mixture Models.
MOHANAD AL-IBADI
Array Processing Techniques for Ice-Sheet Bottom TrackingWhen & Where:
317 Nichols Hall
Committee Members:
Shannon Blunt, ChairJohn Paden
Eric Perrins
Jim Stiles
Huazhen Fang*
Abstract
In airborne multichannel radar sounder signal processing, the collected data are most naturally represented in a cylindrical coordinate system: along-track, range, and elevation angle. The data are generally processed in each of these dimensions sequentially to focus or resolve the data in the corresponding dimension such that a 3D image of the scene can be formulated. Pulse-compression is used to process the data along the range dimension, synthetic aperture radar (SAR) processing is used to process the data in the along-track dimension, and array-processing techniques are used for the elevation angle dimension. After the first two steps, the 3D scene is resolved into toroids with constant along-track and constant range that are centered on the flight path. The targets lying in a particular toroid need to be resolved by estimating their respective elevation angles.
In the proposed work, we focus on the array processing step, where several direction of arrival (DoA) estimation methods will be used to resolve the targets in the elevation-angle dimension, such as MUltiple Signal Classification (MUSIC) and maximum-likelihood estimation (MLE). A tracker is then used on the output of the DoA estimation to track the ice-bottom interface. We propose to use the tree re-weighted message passing algorithm or Kalman filtering, based on the array-processing technique, to track the ice-bottom. The outcome of this is a digital elevation model (DEM) of the ice-bottom. While most published work assumes a narrowband model for the array, we will use a wideband model and focus on issues related to wideband arrays. Along these lines, we propose a theoretical study to evaluate the performance of the radar products based on the array characteristics using different array-processing techniques, such as wideband MLE and focusing-matrices methods. In addition, we will investigate tracking targets using a sparse array composed of three sub-arrays, each separated by a large multiwavelength baseline. Specifically, we propose to develop and investigate the performance of a Kalman tracking solution to this wideband sparse array problem when applied to data collected by the CReSIS radar sounder.
QIAOZHI WANG
Towards the Understanding of Private Content -- Content-based Privacy Assessment and Protection in Social NetworksWhen & Where:
2001B Eaton Hall
Committee Members:
Bo Luo, ChairFengjun Li
Richard Wang
Heechul Yun
Prajna Dhar*
Abstract
In the 2016 presidential election, social networks showed their great power as a “modern form of communication”. With the increasing popularity of social networks, privacy concerns arise. For example, it has been shown that microblogs are revealed to audiences that are significantly larger than users' perceptions. Moreover, when users are emotional, they may post messages with sensitive content and later regret doing so. As a result, users become very vulnerable – private or sensitive information may be accidentally disclosed, even in tweets about trivial daily activities.
Unfortunately, existing research projects on data privacy, such as the k-anonymity and differential privacy mechanisms, mostly focus on protecting individual’s identity from being discovered in large data sets. We argue that the key component of privacy protection in social networks is protecting sensitive content, i.e. privacy as having the ability to control dissemination of information. The overall objectives of the proposed research are: to understand the sensitive content of social network posts, to facilitate content-based protection of private information, and to identify different types of sensitive information. In particular, we propose a user-centered, quantitative measure of privacy based on textual content, and a customized privacy protection mechanism for social networks.
We consider private tweet identification and classification as dual-problems. We propose to develop an algorithm to identify all types of private messages, and, more importantly, automatically score the sensitiveness of private message. We first collect the opinions from a diverse group of users w.r.t. sensitiveness of private information through Amazon Mechanical Turk, and analyze the discrepancies between users' privacy expectations and actual information disclosure. We then develop a computational method to generate the context-free privacy score, which is the “consensus” privacy score for average users. Meanwhile, classification of private tweets is necessary for customized privacy protection. We have made the first attempt to understand different types of private information, and to automatically classify sensitive tweets into 13 pre-defined topic categories. In proposed research, we will further include personal attitudes, topic preferences, and social context into the scoring mechanism, to generate a personalized, context-aware privacy score, which will be utilized in a comprehensive privacy protection mechanism.
STEVE HAENCHEN
A Model to Identify Insider Threats Using Growing Hierarchical Self-Organizing Map of Electronic Media IndicatorsWhen & Where:
1 Eaton Hall
Committee Members:
Hossein Saiedian, ChairArvin Agah
Prasad Kulkarni
Bo Luo
Reza Barati
Abstract
Fraud from insiders costs an estimated $3.7 trillion annually. Current fraud prevention and detection methods that include analyzing network logs, computer events, emails, and behavioral characteristics have not been successful in reducing the losses. The proposed Occupational Fraud Prevention and Detection Model uses existing data from the field of digital forensics along with text clustering algorithms, machine learning, and a growing hierarchical self-organizing map model to predict insider threats based on computer usage behavioral characteristics.
The proposed research leverages research results from information security, software engineering, data science and information retrieval, context searching, search patterns, and machine learning to build and employ a database server and workstations to support 50+ terabytes of data representing entire hard drives from work computers. Forensic software FTK and EnCase are used to generate disk images and test extraction results. Primary research tools are built using modern programming languages. The research data is derived from disk images obtained from actual investigations when fraud was asserted and other disk images when fraud was not asserted.
The research methodology includes building a data extraction tool that is a disk level reader to store the disk, partition, and operating system data in a relational database. An analysis tool is also created to convert the data into information representing usage patterns including summarization, normalization, and redundancy removal. We build a normalizing tool that uses machine learning to adjust the baselines for company, department, and job deviations. A prediction component is developed to derive insider threat scores reflecting the anomalies from the adjusted baseline. The resulting product will allow identification of the computer users most likely to commit fraud so investigators can focus their limited resources on the suspects.
Our primarily plan to evaluate and validate our research results is via empirical study, statistical evaluation and benchmarking with tests of precision and recall from a second set of disk images.