Defense Notices
All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.
Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.
Upcoming Defense Notices
David Felton
Optimization and Evaluation of Physical Complementary Radar WaveformsWhen & Where:
Nichols Hall, Room 129 (Apollo Auditorium)
Committee Members:
Shannon Blunt, ChairRachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata
Abstract
**Currently under security review**
Hao Xuan
Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge DiscoveryWhen & Where:
Nichols Hall, Room 246 (Executive Conference Room)
Committee Members:
Cuncong Zhong, ChairFengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu
Abstract
Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.
These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.
First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.
Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.
Pramil Paudel
Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless ImagingWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Fengjun Li, ChairAlex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao
Abstract
Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference.
We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks.
Sharmila Raisa
Digital Coherent Optical System: Investigation and MonitoringWhen & Where:
Nichols Hall, Room 246 (Executive Conference Room)
Committee Members:
Rongqing Hui, ChairMorteza Hashemi
Erik Perrins
Alessandro Salandrino
Jie Han
Abstract
Coherent wavelength-division multiplexed (WDM) optical fiber systems have become the primary transmission technology for high-capacity data networks, driven by the explosive bandwidth demand of cloud computing, streaming services, and large-scale artificial intelligence training infrastructure. This dissertation investigates two fundamental aspects of digital coherent fiber optic systems under the unifying theme of source and monitoring: the design of multi-wavelength optical sources compatible with high-order coherent detection, and the leveraging of fiber Kerr-effect nonlinearity at the coherent receiver to perform physical-layer link health monitoring and to assess inherent security vulnerabilities — both achieved through digital signal processing of the received complex optical field without dedicated hardware.
We begin by addressing the multi-wavelength transmitter challenge in WDM coherent systems. Existing quantum-dot, quantum-dash, and quantum-well based optical frequency comb (OFC) sources share a common limitation: individual comb line linewidths in the tens of MHz range caused by low output power levels of 1–20 mW, making them incompatible with high-order coherent detection. We demonstrate coherent system application of a single-section InGaAsP QW Fabry-Perot laser diode with greater than 120 mW optical power at the fiber pigtail and 36.14 GHz mode spacing. The high optical power per mode produces Lorentzian equivalent linewidths below 100 kHz — compatible with 16-QAM carrier phase recovery without optical phase locking. Experimental results obtained using a commercial Ciena WaveLogic-Ai coherent transceiver demonstrate 20-channel WDM transmission over 78.3 km of standard single-mode fiber with all channels below the HD-FEC threshold of 3.8 × 10⁻³ at 30 GBaud differential-coded 16-QAM, corresponding to an aggregate capacity of 2.15 Tb/s from a single laser device.
After investigating the QW Fabry-Perot laser as a multi-wavelength source for coherent WDM transmission, we leverage the coherent receiver DSP to exploit fiber Kerr-effect nonlinearity for longitudinal power profile estimation, enabling reconstruction of the signal power distribution P(z) along the full multi-span link without dedicated hardware or traffic interruption. We propose a modified enhanced regular perturbation (ERP) method that corrects two independent physical error sources of the standard RP1 least-squares baseline: the accumulated nonlinear phase rotation, and the dispersion-mediated phase-to-intensity conversion — a second bias source not addressed by prior methods. The RP1 method produces mean absolute error (MAE) that scales quadratically with span count, growing to 1.656 dB at 10 spans and 3 dBm. The modified ERP reduces this to 0.608 dB — an improvement that grows consistently with link length, confirming increasing advantage in the long-haul regime. Extension to WDM through an XPM-aware per-channel formulation achieves MAE of 0.113–0.419 dB across 150–500 km link lengths.
In addition to its role in enabling DSP-based longitudinal power profile estimation, the fiber Kerr-effect nonlinearity is shown to give rise to an inherent physical-layer security vulnerability in coherent WDM systems. We show that an eavesdropper co-tenanting a shared fiber — transmitting a continuous-wave probe at a wavelength adjacent to the legitimate signal — can capture the XPM-induced waveform at the fiber output and apply a bidirectional gated recurrent unit neural network, trained on split-step Fourier method simulation data, to reconstruct the transmitted symbol sequence without physical fiber access and without perturbing the legitimate signal. This eavesdropping mechanism is experimentally validated using a commercial Ciena WaveLogic-Ai coherent transceiver for ASK, BPSK, QPSK, and 16-QAM modulation formats at 4.26 GBaud and 8.56 GBaud over one- and two-span 75 km fiber systems, achieving zero symbol errors under high-OSNR conditions. Noise-aware training over OSNR from 20 to 60 dB maintains symbol error rate below 10⁻² for OSNR above 25–30 dB.
Together, these three contributions demonstrate that the coherent fiber optic system is a versatile physical instrument extending well beyond its role as a data transmission medium. The coherent receiver infrastructure — deployed for high-order modulation and data recovery — simultaneously enables the high-power OFC laser to serve as a practical multi-wavelength transmitter source, and provides the complex field measurement capability through which fiber Kerr-effect nonlinearity can be exploited constructively for distributed link monitoring and, as a direct consequence, reveals an inherent physical-layer security exposure in shared fiber infrastructure. This unified perspective on the coherent system as both a transmission platform and a general-purpose measurement instrument has direct relevance to the design of spectrally efficient, self-monitoring, and physically secure optical interconnects for next-generation AI computing networks.
Arman Ghasemi
Task-Oriented Data Communication and Compression for Timely Forecasting and Control in Smart GridsWhen & Where:
Nichols Hall, Room 246 (Executive Conference Room)
Committee Members:
Morteza Hashemi, ChairAlexandru Bardas
Prasad Kulkarni
Taejoon Kim
Zsolt Talata
Abstract
Advances in sensing, communication, and intelligent control have transformed power systems into data-driven smart grids, where forecasting and intelligent decision-making are essential components. Modern smart grids include distributed energy resources (DERs), renewable generation, battery energy storage systems, and large numbers of grid-edge devices that continuously generate time-series data. At the same time, increasing renewable penetration introduces substantial uncertainty in generation, net load, and market operations, while communication networks impose bandwidth, latency, and reliability constraints on timely data delivery. This dissertation addresses how time-series forecasting, data compression, and task-oriented wireless communication can be jointly designed for smart grid applications.
First, we study weather-aware distributed energy management in prosumer-centric microgrids and show that incorporating day-ahead weather information into decision-making improves battery dispatch and reduces the impact of renewable uncertainty. Second, we introduce forecasting-aware energy management in both wholesale and retail electricity markets, highlighting how renewable generation forecasting affects pricing, scheduling, and uncertainty mitigation. Third, we develop and evaluate deep learning methods for renewable generation forecasting, showing that Transformer-based models outperform recurrent baselines such as RNN and LSTM for wind and solar prediction tasks.
Building on this forecasting foundation, we develop a communication-efficient forecasting framework in which high-dimensional smart grid measurements are compressed into low-dimensional latent representations before transmission. This framework is extended into a task-oriented communication system that jointly optimizes data relevance and information timeliness, so that the receiver obtains compressed updates that remain useful for downstream forecasting tasks. Finally, we extend this framework to a distributed multi-node uplink setting, where multiple grid sensors share a bandwidth-limited channel, and develop scheduling policy that improves both the timeliness and task-relevance of received updates.
Pardaz Banu Mohammad
Towards Early Detection of Alzheimer’s Disease based on Speech using Reinforcement Learning Feature SelectionWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Arvin Agah, ChairDavid Johnson
Sumaiya Shomaji
Dongjie Wang
Sara Wilson
Abstract
Alzheimer’s Disease (AD) is a progressive, irreversible neurodegenerative disorder and the leading cause of dementia worldwide, affecting an estimated 55 million people globally. The window of opportunity for intervention is demonstrably narrow, making reliable early-stage detection a clinical and scientific imperative. While current diagnostic techniques such as neuroimaging and cerebrospinal fluid (CSF) biomarkers carry well-defined limitations in scalability, cost, and access equity, speech has emerged as a compelling non-invasive proxy for cognitive function evaluation.
This work presents a novel approach for using acoustic feature selection as a decision-making technique and implements it using deep reinforcement learning. Specifically, we use a Deep-Q-Network (DQN) agent to navigate a high dimensional feature space of over 6,000 acoustic features extracted using the openSMILE toolkit, dynamically constructing maximally discriminative and non-redundant features subsets. In order to capture the latent structural dependencies among
acoustic features which classifier and wrapper methods have difficulty to model, we introduce the Graph Convolutional Network (GCN) based correlation awareness feature representation layer that operates as an auxiliary input to the DQN state encoder. Post selection interpretability is reinforced through TF-IDF weighting and K-means clustering which together yield both feature level and cluster level explanations that are clinically actionable. The framework is evaluated across five classifiers, namely, support vector machines (SVM), logistic regression, XGBoost, random forest, and feedforward neural network. We use 10-fold stratified cross-validation on established benchmarks of datasets, including DementiaBank Pitt Corpus, Ivanova, and ADReSS challenge data. The proposed approach is benchmarked against state-of-the-art feature selection methods such as LASSO, Recursive feature selection, and mutual information selectors. This research contributes to three primary intellectual advances: (1) a graph augmented state representation that encodes inter-feature relational structure within a reinforcement learning agent, (2) a clinically interpretable pipeline that bridges the gap between algorithmic performance and translational utility, and (3) multilingual data approach for the reinforcement learning agent framework. This study has direct implications for equitable, low-cost and scalable AD screening in both clinical and community settings.
Zhou Ni
Bridging Federated Learning and Wireless Networks: From Adaptive Learning to FLdriven System OptimizationWhen & Where:
Nichols Hall, Room 246 (Executive Conference Room)
Committee Members:
Morteza Hashemi, ChairFengjun Li
Van Ly Nguyen
Han Wang
Shawn Keshmiri
Abstract
Federated learning (FL) has emerged as a promising distributed machine learning
framework that enables multiple devices to collaboratively train models without sharing raw
data, thereby preserving privacy and reducing the need for centralized data collection. However,
deploying FL in practical wireless environments introduces two major challenges. First, the data
generated across distributed devices are often heterogeneous and non-IID, which makes a single
global model insufficient for many users. Second, learning performance in wireless systems is
strongly affected by communication constraints such as interference, unreliable channels, and
dynamic resource availability. This PhD research aims to address these challenges by bridging
FL methods and wireless networks.
In the first thrust, we develop personalized and adaptive FL methods given the underlying
wireless link conditions. To this end, we propose channel-aware neighbor selection and
similarity-aware aggregation in wireless device-to-device (D2D) learning environments. We
further investigate the impacts of partial model update reception on FL performance. The
overarching goal of the first thrust is to enhance FL performance under wireless constraints.
Next, we investigate the opposite direction and raise the question: How can FL-based distributed
optimization be used for the design of next-generation wireless systems? To this end, we
investigate communication-aware participation optimization in vehicular networks, where
wireless resource allocation affects the number of clients that can successfully contribute to FL.
We further extend this direction to integrated sensing and communication (ISAC) systems,
where personalized FL (PFL) is used to support distributed beamforming optimization with joint
sensing and communication objectives.
Overall, this research establishes a unified framework for bridging FL and wireless networks. As
a future direction, this work will be extended to more realistic ISAC settings with dynamic
spectrum access, where communication, sensing, scheduling, and learning performance must be
considered jointly.
Arnab Mukherjee
Attention-Based Solutions for Occlusion Challenges in Person TrackingWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Prasad Kulkarni, ChairSumaiya Shomaji
Hongyang Sun
Jian Li
Abstract
Person re-identification (Re-ID) and multi-object tracking in unconstrained surveillance environments pose significant challenges within the field of computer vision. These complexities stem mainly from occlusion, variability in appearance, and identity switching across various camera views. This research outlines a comprehensive and innovative agenda aimed at tackling these issues, employing a series of increasingly advanced deep learning architectures, culminating in a groundbreaking occlusion-aware Vision Transformer framework.
At the heart of this work is the introduction of Deep SORT with Multiple Inputs (Deep SORT-MI), a cutting-edge real-time Re-ID system featuring a dual-metric association strategy. This strategy adeptly combines Mahalanobis distance for motion-based tracking with cosine similarity for appearance-based re-identification. As a result, this method significantly decreases identity switching compared to the baseline SORT algorithm on the MOT-16 benchmark, thereby establishing a robust foundation for metric learning in subsequent research.
Expanding on this foundation, a novel pose-estimation framework integrates 2D skeletal keypoint features extracted via OpenPose directly into the association pipeline. By capturing the spatial relationships among body joints along with appearance features, this system enhances robustness against posture variations and partial occlusion. Consequently, it achieves substantial reductions in false positives and identity switches compared to earlier methods, showcasing its practical viability.
Furthermore, a Diverse Detector Integration (DDI) study meticulously assessed the influence of detector choices—including YOLO v4, Faster R-CNN, MobileNet SSD v2, and Deep SORT—on the efficacy of metric learning-based tracking. The results reveal that YOLO v4 consistently delivers exceptional tracking accuracy on both the MOT-16 and MOT-17 datasets, establishing its superiority in this competitive landscape.
In conclusion, this body of research notably advances occlusion-aware person Re-ID by illustrating a clear progression from metric learning to pose-guided feature extraction and ultimately to transformer-based global attention modeling. The findings underscore that lightweight, meticulously parameterized Vision Transformers can achieve impressive generalization for occlusion detection, even under constrained data scenarios. This opens up exciting prospects for integrated detection, localization, and re-identification in real-world surveillance systems, promising to enhance their effectiveness and reliability.
Sai Katari
Android Malware Detection SystemWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
David Johnson, ChairArvin Agah
Prasad Kulkarni
Abstract
Android malware remains a significant threat to mobile security, requiring efficient and scalable detection methods. This project presents an Android Malware Detection System that uses machine learning to classify applications as benign or malicious based on static permission-based analysis. The system is trained on the TUANDROMD dataset of 4,464 applications using four models-Logistic Regression, XGBoost, Random Forest, and Naive Bayes-with a 75/25 train/test split and 5-fold cross-validation on the training set for evaluation. To improve reliability, the system incorporates a hybrid decision approach that combines machine learning confidence scores with a rule-based static analysis engine, using a three-zone confidence routing mechanism to capture threats that ML alone may miss. The solution is deployed as a Flask web application with both a manual detection interface and an APK file scanner, providing predictions, confidence scores, and risk insights, ultimately supporting more informed and secure decision-making.
Past Defense Notices
Serigne Seck
Packet Loss Prevention in Queues using SDNWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Taejoon Kim, ChairMorteza Hashemi, Co-Chair
David Johnson
Abstract
Packets are transferred between nodes within a network. However, a packet can be dropped while trying to join the queue of a node it was routed to. In networking, this is referred to as packet loss. It can be caused by buffer scarcity in a congested network. Such phenomenon results in a reduced data rate and a delay increase due to packet retransmissions.
In this work, we propose an algorithm to perform load balancing on a network of queues via SDN to prevent packet loss. It implements a parameter K, based on the queues occupancy and traffic flow, to control an iterative packet redistribution process. In different experiments conducted on network models in which the queues varied in number, size and occupancy, our algorithm outperformed a load balancer using the Round-Robin technique.
Brian Quiroz
Mobile Edge Computing for Unmanned VehiclesWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Morteza Hashemi, ChairTaejoon Kim
Prasad Kulkarni
Abstract
Unmanned aerial vehicles (UAVs) and autonomous vehicles are becoming more ubiquitous than ever before. From medical to delivery drones, to space exploration rovers and self-driving taxi services, these vehicles are starting to play a prominent role in society as well as in our day to day lives.
Efficient computation and communication strategies are paramount to the effective functioning of these vehicles. Mobile Edge Computing (MEC) is an innovative network technology that enables resource-constrained devices - such as UAVs and autonomous vehicles - to offload computationally intensive tasks to a nearby MEC server. Moreover, vehicles such as self-driving cars must reliably and securely relay and receive latency-sensitive information to improve traffic safety. Extensive research performed on vehicle to vehicle (V2V) and vehicle to everything (V2X) communication indicates that they will both be further enhanced by the widespread usage of 5G technology.
We consider two relevant problems in mobile edge computing for unmanned vehicles. The first problem was to satisfy resource-constrained UAV's need for a resource-efficient offloading policy. To that end, we implemented both a computation and an energy consumption model and trained a DQN agent that seeks to maximize task completion and minimize energy consumption. The second problem was establishing communication between two autonomous vehicles and between an autonomous vehicle and an MEC server. To accomplish this goal, we experimented by leveraging an autonomous vehicle's server to send and receive custom messages in real time. These experiments will serve as a stepping stone towards enabling mobile edge computing and device-to-device communication and computation.
Ruturaj Vaidya
Explore Effectiveness and Performance of Security Checks on Software BinariesWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Prasad Kulkarni, ChairAlex Bardas
Drew Davidson
Esam El-Araby
Michael Vitevitch
Abstract
Binary analysis is difficult, as most of semantic and syntactic information available at source-level gets lost during the compilation process. If the binary is stripped and/ or optimized, then it negatively affects the efficacy of binary analysis frameworks. Moreover, handwritten assembly, obfuscation, excessive indirect calls or jumps, etc. further degrade their accuracy. Thus, it is important to investigate and assess the challenges to improve the binary analysis. One way of doing that is by studying security techniques implemented at binary-level.
In this dissertation we propose to implement existing compiler-level techniques for binary executables and thereby evaluate how does the loss of information at binary-level affect the performance of existing compiler-level techniques in terms of both efficiency and effectiveness.
Michael Bechtel
Shared Resource Denial-of-Service Attacks on Multicore PlatformsWhen & Where:
Eaton Hall, Room 2001B
Committee Members:
Heechul Yun, ChairMohammad Alian
Drew Davidson
Prasad Kulkarni
Shawn Keshmiri
Abstract
With the increased adoption of machine learning algorithms across many different fields, powerful computing platforms have become necessary to meet their computational needs. Multicore platforms are a popular choice due to their ability to provide greater computing capabilities and still meet the different size, weight, and power (SWaP) constraints. As a result, multicore systems are also being employed at an increasing rate. However, contention for hardware resources between the multiple cores is a significant challenge as it can lead to interference and unpredictable timing behaviors. Furthermore, this contention can be intentionally induced by malicious actors with the specific goals of inhibiting system performance and increasing the execution time of safety-critical tasks. This is done by performing Denial-of-Service (DoS) attacks that target shared resources in order to prevent other cores from accessing them. When done properly, these DoS attacks can have significant impacts to performance and can threaten system safety. For example, we find that DoS attacks can cause >300X slowdown on the popular Raspberry Pi 3 embedded platform. Due to the inherent risks, it is vital that we discover and understand the mechanisms through which shared resource contention can occur and develop solutions that mitigate or prevent the potential impacts.
In this work, we investigate and evaluate shared resource contention on multicore platforms and the impacts it can have on the performance of real-time tasks. Leveraging this contention, we propose various Denial-of-Service attacks that each target different shared resources in the memory hierarchy with the goal of causing as much slowdown as possible. We show that each attack can inflict significant temporal slowdowns to victim tasks on target platforms by exploiting different hardware and software mechanisms. We then develop and analyze techniques for providing shared resource isolation and temporal performance guarantees for safety-critical tasks running on multicore platforms. In particular, we find that bandwidth throttling mechanisms are effective solutions against many DoS attacks and can protect the performance of real-time victim tasks.
Anushka Bhattacharya
Predicting In-Season Soil Mineral Nitrogen in Corn Production Using Deep Learning ModelWhen & Where:
Nichols Hall, Room 246
Committee Members:
Taejoon Kim, ChairMorteza Hashemi
Dorivar Ruiz Diaz
Abstract
One of the biggest challenges in nutrient management in corn (Zea mays) production is determining the amount of plant-available nitrogen (N) that will be supplied to the crop by the soil. Measuring a soil’s N-supplying power is quite difficult and approximations are often used in-lieu of intensive soil testing. This can lead to under/over-fertilization of crops, and in turn increased risk of crop N-deficiencies or environmental degradation. In this paper, we propose a deep learning algorithm to predict the inorganic-N content of the soil on a given day of the growing season. Since the historic data for inorganic nitrogen (IN) is scarce, deep learning has not yet been implemented in predicting fertilizer content. To overcome this hurdle, Generative Adversarial Network (GAN) is used to produce synthetic IN data and is trained using offline simulation data from the Decision Support System for Agrotechnology Transfer (DSSAT). Additionally, the time-series prediction problem is solved using long-short term memory (LSTM) neural networks. This model proves to be economical as it gives an estimate without the need for comprehensive soil testing, overcomes the issue of limited available data, and the accuracy makes it reliable for use.
Krushi Patel
Image Classification & Segmentation based on Enhanced CNN and Transformer NetworksWhen & Where:
Nichols Hall, Room 250 - Gemini Room
Committee Members:
Fengjun Li, ChairPrasad Kulkarni
Bo Luo
Cuncong Zhong
Guanghui Wang
Abstract
Convolutional Neural Networks (CNNs) have significantly improved the performance on various computer vision tasks such as image recognition and segmentation based on their rich representation power. To enhance the performance of CNN, a self-attention module is embedded after each layer in the network. Recently proposed Transformer-based models achieve outstanding performance by employing a multi-head self-attention module as the main building block. However, several challenges still need to be addressed, such as (1) focusing only on class-specified limited channels in CNN; (2) limited respective field in the local transformer; and (3) addition of redundant features and lack of multi-scale features in U-Net type segmentation architecture.
In our work, we propose new strategies to address these issues. First, we propose a novel channel-based self-attention module to diversify the focus more on the discriminative and significant channels, and the module can be embedded at the end of any backbone network for image classification. Second, to limit the noise added by the shallow layers of an encoder in U-Net type architecture, we replaced the skip connections with the Adaptive Global Context Module (AGCM). In addition, we introduced the Semantic Feature Enhancement Module (SFEM) for multi-scale feature enhancement in polyp segmentation. Third, we propose a Multi-scaled Overlapped Attention (MOA) mechanism in the local transformer-based network for image classification to establish the long-range dependencies and initiate the neighborhood window communication.
Justinas Lialys
Parametrically resonant surface plasmon polaritonsWhen & Where:
2001B Eaton Hall
Committee Members:
Alessandro Salandrino, ChairKenneth Demarest
Shima Fardad
Rongqing Hui
Xinmai Yang
Abstract
The surface electromagnetic waves that propagate along a metal-dielectric or a metal-air interface are called surface plasmon polaritons (SPPs). These SPPs are advantageous in a broad range of applications, including in optical waveguides to increase the transmission rates of carrier waves, in near field optics to enhance the resolution beyond the diffraction limit, and in Raman spectroscopy to amplify the Raman signal. However, they have an inherent limitation: as the tangential wavevector component of propagation is larger than what is permitted for the homogenous plane wave in the dielectric medium, this poses a phase-matching issue. In other words, the available spatial vector in the dielectric at a given frequency is smaller than what is required by SPP to be excited. The most commonly known technique to bypass this problem is by using the Otto and Kretschmann configurations. A glass prism is used to increase the available spatial vector in dielectric/air. Other methods are the evanescent field directional coupling, optical grating, localized scatterers, and coupling via highly focused beams. However, even with all these methods at our disposal, it is still challenging to couple SPPs that have a large propagation constant.
As SPPs apply to a wide range of purposes, it is vitally important to overcome the SPP excitation dilemma. Presented here is a novel way to efficiently inject power into SPPs via temporal modulation of the dielectric adhered to the metal. In this configuration, the dielectric constant is modulated in time using an incident pump field. As a result of the induced changes in the dielectric constant, we show that efficient phase-matched coupling can be achieved even by a perpendicularly incident uniform plane wave. This novel method of exciting SPPs paves the way for further understanding and implementation of SPPs in a plethora of applications. For example, optical waveguides can be investigated under such excitation. Hence, this technique opens new possibilities in conventional plasmonics, as well as in the emerging field of nonlinear plasmonics.
Andrei Elliott
Promise Land: Proving Correctness with Strongly Typed Javascript-Style PromisesWhen & Where:
Nichols Hall, Room 250, Gemini Room
Committee Members:
Matt Moore, ChairPerry Alexander
Drew Davidson
Abstract
Code that can run asynchronously is important in a wide variety of situations, from user interfaces to communication over networks, to the use of concurrency for performance gains. One widely used method of specifying asynchronous control flow is the Promise model as used in Javascript. Promises are powerful, but can be confusing and hard-to-debug. This problem is exacerbated by Javascript’s permissive type system, where erroneous code is likely to fail silently, with values being implicitly coerced into unexpected types at runtime.
The present work implements Javascript-style Promises in Haskell, translating the model to a strongly typed framework where we can use the type system to rule out some classes of bugs.
Common errors – such as failure to call one of the callbacks of an executor, which would, in Javascript, leave the Promise in an eternally-pending deadlock state – can be detected for free by the type system at compile time and corrected without even needing to run the code.
We also demonstrate that Promises form a monad, providing a monad instance that allows code using Promises to be written using Haskell’s do notation.
Hoang Trong Mai
Design and Development of Multi-band and Ultra-wideband Antennas and Circuits for Ice and Snow Radar MeasurementsWhen & Where:
Nichols Hall, Room 317
Committee Members:
Carl Leuschen, ChairFernando Rodriguez-Morales, Co-Chair
Christopher Allen
Abstract
Remote sensing based on radar technology has been successfully used for several decades as an effective tool of scientific discovery. A particular application of radar remote sensing instruments is the systematic monitoring of ice and snow masses in both hemispheres of the Earth. The operating requirements of these instruments are driven by factors such as science requirements and platform constraints, often necessitating the development of custom electronic components to enable the desired radar functionality.
This work focuses on component development and trade studies for two multichannel radar systems. First, this thesis presents the design and implementation of two dual-polarized ultra-wideband antennas for a ground-based dual-band ice penetrating radar. The first antenna operates at UHF (600–900 MHz) while the second antenna operates at VHF (140–215 MHz). Each antenna element is composed of two orthogonal octagon-shaped dipoles, two inter-locked printed circuit baluns and an impedance matching network for each polarization. Prototype of each band shows a VSWR of less than 2:1 at both polarizations over a fractional bandwidth exceeding 40%. The antennas developed offer cross-polarization isolation larger than 30 dB, an E-plane 3-dB beamwidth of 69 degrees, and a gain of at least 4 dBi with a variation of ± 1 dB across the bandwidth. This design with high power handling in mind also allows for straightforward adjustment of the antenna dimensions to meet other bandwidth constrains. It is being used as the basis for an airborne system.
Next, this work documents design details and measured performance of an improved and integrated x16 frequency multiplier system for an airborne snow-probing radar. This sub-system produces a 40 – 56 GHz linear frequency sweep from a 2.5 – 3.5 GHz chirp and mixes it down to the 2 – 18 GHz range. The resulting chirp is used for transmission and analog de-chirping of the receive signal. The initial prototype developed through this work provided a higher level of integration and wider fractional bandwidth (>135%) compared to earlier versions implemented with the same frequency plan and a path to guide future realizations.
Lastly, this work documents a series of trade studies on antenna array configurations for both radar systems using electromagnetic simulation tools and measurements.
Xi Mo
Convolutional Neural Network in Pattern RecognitionWhen & Where:
Zoom Meeting, please contact jgrisafe@ku.edu for link.
Committee Members:
Cuncong Zhong, ChairTaejoon Kim
Fengjun Li
Bo Luo
Hauzhen Fang
Abstract
Since convolutional neural network (CNN) was first implemented by Yann LeCun et al. in 1989, CNN and its variants have been widely implemented to numerous topics of pattern recognition, and have been considered as the most crucial techniques in the field of artificial intelligence and computer vision. This dissertation not only demonstrates the implementation aspect of CNN, but also lays emphasis on the methodology of neural network (NN) based classifier.
As known to many, one general pipeline of NN-based classifier can be recognized as three stages: pre-processing, inference by models, and post-processing. To demonstrate the importance of pre-processing techniques, this dissertation presents how to model actual problems in medical pattern recognition and image processing by introducing conceptual abstraction and fuzzification. In particular, a transformer on the basis of self-attention mechanism, namely beat-rhythm transformer, greatly benefits from correct R-peak detection results and conceptual fuzzification.
Recently proposed self-attention mechanism has been proven to be the top performer in the fields of computer vision and natural language processing. In spite of the pleasant accuracy and precision it has gained, it usually consumes huge computational resources to perform self-attention. Therefore, realtime global attention network is proposed to make a better trade-off between efficiency and performance for the task of image segmentation. To illustrate more on the stage of inference, we also propose models to detect polyps via Faster R-CNN - one of the most popular CNN-based 2D detectors, as well as a 3D object detection pipeline for regressing 3D bounding boxes from LiDAR points and stereo image pairs powered by CNN.
The goal for post-processing stage is to refine artifacts inferred by models. For the semantic segmentation task, the dilated continuous random field is proposed to be better fitted to CNN-based models than the widely implemented fully-connected continuous random field. Proposed approaches can be further integrated into a reinforcement learning architecture for robotics.