Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

David Felton

Optimization and Evaluation of Physical Complementary Radar Waveforms

When & Where:


Nichols Hall, Room 129 (Apollo Auditorium)

Committee Members:

Shannon Blunt, Chair
Rachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata

Abstract

**Currently under security review**


Hao Xuan

Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge Discovery

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Cuncong Zhong, Chair
Fengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu

Abstract

Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.

These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.

First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.

Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.


Pramil Paudel

Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless Imaging

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Fengjun Li, Chair
Alex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao

Abstract

Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference. 

We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks. 


Sharmila Raisa

Digital Coherent Optical System: Investigation and Monitoring

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Rongqing Hui, Chair
Morteza Hashemi
Erik Perrins
Alessandro Salandrino
Jie Han

Abstract

Coherent wavelength-division multiplexed (WDM) optical fiber systems have become the primary transmission technology for high-capacity data networks, driven by the explosive bandwidth demand of cloud computing, streaming services, and large-scale artificial intelligence training infrastructure. This dissertation investigates two fundamental aspects of digital coherent fiber optic systems under the unifying theme of source and monitoring: the design of multi-wavelength optical sources compatible with high-order coherent detection, and the leveraging of fiber Kerr-effect nonlinearity at the coherent receiver to perform physical-layer link health monitoring and to assess inherent security vulnerabilities — both achieved through digital signal processing of the received complex optical field without dedicated hardware.

We begin by addressing the multi-wavelength transmitter challenge in WDM coherent systems. Existing quantum-dot, quantum-dash, and quantum-well based optical frequency comb (OFC) sources share a common limitation: individual comb line linewidths in the tens of MHz range caused by low output power levels of 1–20 mW, making them incompatible with high-order coherent detection. We demonstrate coherent system application of a single-section InGaAsP QW Fabry-Perot laser diode with greater than 120 mW optical power at the fiber pigtail and 36.14 GHz mode spacing. The high optical power per mode produces Lorentzian equivalent linewidths below 100 kHz — compatible with 16-QAM carrier phase recovery without optical phase locking. Experimental results obtained using a commercial Ciena WaveLogic-Ai coherent transceiver demonstrate 20-channel WDM transmission over 78.3 km of standard single-mode fiber with all channels below the HD-FEC threshold of 3.8 × 10⁻³ at 30 GBaud differential-coded 16-QAM, corresponding to an aggregate capacity of 2.15 Tb/s from a single laser device.

After investigating the QW Fabry-Perot laser as a multi-wavelength source for coherent WDM transmission, we leverage the coherent receiver DSP to exploit fiber Kerr-effect nonlinearity for longitudinal power profile estimation, enabling reconstruction of the signal power distribution P(z) along the full multi-span link without dedicated hardware or traffic interruption. We propose a modified enhanced regular perturbation (ERP) method that corrects two independent physical error sources of the standard RP1 least-squares baseline: the accumulated nonlinear phase rotation, and the dispersion-mediated phase-to-intensity conversion — a second bias source not addressed by prior methods. The RP1 method produces mean absolute error (MAE) that scales quadratically with span count, growing to 1.656 dB at 10 spans and 3 dBm. The modified ERP reduces this to 0.608 dB — an improvement that grows consistently with link length, confirming increasing advantage in the long-haul regime. Extension to WDM through an XPM-aware per-channel formulation achieves MAE of 0.113–0.419 dB across 150–500 km link lengths.

In addition to its role in enabling DSP-based longitudinal power profile estimation, the fiber Kerr-effect nonlinearity is shown to give rise to an inherent physical-layer security vulnerability in coherent WDM systems. We show that an eavesdropper co-tenanting a shared fiber — transmitting a continuous-wave probe at a wavelength adjacent to the legitimate signal — can capture the XPM-induced waveform at the fiber output and apply a bidirectional gated recurrent unit neural network, trained on split-step Fourier method simulation data, to reconstruct the transmitted symbol sequence without physical fiber access and without perturbing the legitimate signal. This eavesdropping mechanism is experimentally validated using a commercial Ciena WaveLogic-Ai coherent transceiver for ASK, BPSK, QPSK, and 16-QAM modulation formats at 4.26 GBaud and 8.56 GBaud over one- and two-span 75 km fiber systems, achieving zero symbol errors under high-OSNR conditions. Noise-aware training over OSNR from 20 to 60 dB maintains symbol error rate below 10⁻² for OSNR above 25–30 dB.

Together, these three contributions demonstrate that the coherent fiber optic system is a versatile physical instrument extending well beyond its role as a data transmission medium. The coherent receiver infrastructure — deployed for high-order modulation and data recovery — simultaneously enables the high-power OFC laser to serve as a practical multi-wavelength transmitter source, and provides the complex field measurement capability through which fiber Kerr-effect nonlinearity can be exploited constructively for distributed link monitoring and, as a direct consequence, reveals an inherent physical-layer security exposure in shared fiber infrastructure. This unified perspective on the coherent system as both a transmission platform and a general-purpose measurement instrument has direct relevance to the design of spectrally efficient, self-monitoring, and physically secure optical interconnects for next-generation AI computing networks.


Arman Ghasemi

Task-Oriented Data Communication and Compression for Timely Forecasting and Control in Smart Grids

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Morteza Hashemi, Chair
Alexandru Bardas
Prasad Kulkarni
Taejoon Kim
Zsolt Talata

Abstract

Advances in sensing, communication, and intelligent control have transformed power systems into data-driven smart grids, where forecasting and intelligent decision-making are essential components. Modern smart grids include distributed energy resources (DERs), renewable generation, battery energy storage systems, and large numbers of grid-edge devices that continuously generate time-series data. At the same time, increasing renewable penetration introduces substantial uncertainty in generation, net load, and market operations, while communication networks impose bandwidth, latency, and reliability constraints on timely data delivery. This dissertation addresses how time-series forecasting, data compression, and task-oriented wireless communication can be jointly designed for smart grid applications.

First, we study weather-aware distributed energy management in prosumer-centric microgrids and show that incorporating day-ahead weather information into decision-making improves battery dispatch and reduces the impact of renewable uncertainty. Second, we introduce forecasting-aware energy management in both wholesale and retail electricity markets, highlighting how renewable generation forecasting affects pricing, scheduling, and uncertainty mitigation. Third, we develop and evaluate deep learning methods for renewable generation forecasting, showing that Transformer-based models outperform recurrent baselines such as RNN and LSTM for wind and solar prediction tasks.

Building on this forecasting foundation, we develop a communication-efficient forecasting framework in which high-dimensional smart grid measurements are compressed into low-dimensional latent representations before transmission. This framework is extended into a task-oriented communication system that jointly optimizes data relevance and information timeliness, so that the receiver obtains compressed updates that remain useful for downstream forecasting tasks. Finally, we extend this framework to a distributed multi-node uplink setting, where multiple grid sensors share a bandwidth-limited channel, and develop scheduling policy that improves both the timeliness and task-relevance of received updates.


Pardaz Banu Mohammad

Towards Early Detection of Alzheimer’s Disease based on Speech using Reinforcement Learning Feature Selection

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Arvin Agah, Chair
David Johnson
Sumaiya Shomaji
Dongjie Wang
Sara Wilson

Abstract

Alzheimer’s Disease (AD) is a progressive, irreversible neurodegenerative disorder and the leading cause of dementia worldwide, affecting an estimated 55 million people globally. The window of opportunity for intervention is demonstrably narrow, making reliable early-stage detection a clinical and scientific imperative. While current diagnostic techniques such as neuroimaging and cerebrospinal fluid (CSF) biomarkers carry well-defined limitations in scalability, cost, and access equity, speech has emerged as a compelling non-invasive proxy for cognitive function evaluation.

This work presents a novel approach for using acoustic feature selection as a decision-making technique and implements it using deep reinforcement learning. Specifically, we use a Deep-Q-Network (DQN) agent to navigate a high dimensional feature space of over 6,000 acoustic features extracted using the openSMILE toolkit, dynamically constructing maximally discriminative and non-redundant features subsets. In order to capture the latent structural dependencies among

acoustic features which classifier and wrapper methods have difficulty to model, we introduce the Graph Convolutional Network (GCN) based correlation awareness feature representation layer that operates as an auxiliary input to the DQN state encoder. Post selection interpretability is reinforced through TF-IDF weighting and K-means clustering which together yield both feature level and cluster level explanations that are clinically actionable. The framework is evaluated across five classifiers, namely, support vector machines (SVM), logistic regression, XGBoost, random forest, and feedforward neural network. We use 10-fold stratified cross-validation on established benchmarks of datasets, including DementiaBank Pitt Corpus, Ivanova, and ADReSS challenge data. The proposed approach is benchmarked against state-of-the-art feature selection methods such as LASSO, Recursive feature selection, and mutual information selectors. This research contributes to three primary intellectual advances: (1) a graph augmented state representation that encodes inter-feature relational structure within a reinforcement learning agent, (2) a clinically interpretable pipeline that bridges the gap between algorithmic performance and translational utility, and (3) multilingual data approach for the reinforcement learning agent framework. This study has direct implications for equitable, low-cost and scalable AD screening in both clinical and community settings.


Zhou Ni

Bridging Federated Learning and Wireless Networks: From Adaptive Learning to FLdriven System Optimization

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Morteza Hashemi, Chair
Fengjun Li
Van Ly Nguyen
Han Wang
Shawn Keshmiri

Abstract

Federated learning (FL) has emerged as a promising distributed machine learning
framework that enables multiple devices to collaboratively train models without sharing raw
data, thereby preserving privacy and reducing the need for centralized data collection. However,
deploying FL in practical wireless environments introduces two major challenges. First, the data
generated across distributed devices are often heterogeneous and non-IID, which makes a single
global model insufficient for many users. Second, learning performance in wireless systems is
strongly affected by communication constraints such as interference, unreliable channels, and
dynamic resource availability. This PhD research aims to address these challenges by bridging
FL methods and wireless networks.
In the first thrust, we develop personalized and adaptive FL methods given the underlying
wireless link conditions. To this end, we propose channel-aware neighbor selection and
similarity-aware aggregation in wireless device-to-device (D2D) learning environments. We
further investigate the impacts of partial model update reception on FL performance. The
overarching goal of the first thrust is to enhance FL performance under wireless constraints.
Next, we investigate the opposite direction and raise the question: How can FL-based distributed
optimization be used for the design of next-generation wireless systems? To this end, we
investigate communication-aware participation optimization in vehicular networks, where
wireless resource allocation affects the number of clients that can successfully contribute to FL.
We further extend this direction to integrated sensing and communication (ISAC) systems,
where personalized FL (PFL) is used to support distributed beamforming optimization with joint
sensing and communication objectives.
Overall, this research establishes a unified framework for bridging FL and wireless networks. As
a future direction, this work will be extended to more realistic ISAC settings with dynamic
spectrum access, where communication, sensing, scheduling, and learning performance must be
considered jointly.


Arnab Mukherjee

Attention-Based Solutions for Occlusion Challenges in Person Tracking

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Prasad Kulkarni, Chair
Sumaiya Shomaji
Hongyang Sun
Jian Li

Abstract

Person re-identification (Re-ID) and multi-object tracking in unconstrained surveillance environments pose significant challenges within the field of computer vision. These complexities stem mainly from occlusion, variability in appearance, and identity switching across various camera views. This research outlines a comprehensive and innovative agenda aimed at tackling these issues, employing a series of increasingly advanced deep learning architectures, culminating in a groundbreaking occlusion-aware Vision Transformer framework.

At the heart of this work is the introduction of Deep SORT with Multiple Inputs (Deep SORT-MI), a cutting-edge real-time Re-ID system featuring a dual-metric association strategy. This strategy adeptly combines Mahalanobis distance for motion-based tracking with cosine similarity for appearance-based re-identification. As a result, this method significantly decreases identity switching compared to the baseline SORT algorithm on the MOT-16 benchmark, thereby establishing a robust foundation for metric learning in subsequent research.

Expanding on this foundation, a novel pose-estimation framework integrates 2D skeletal keypoint features extracted via OpenPose directly into the association pipeline. By capturing the spatial relationships among body joints along with appearance features, this system enhances robustness against posture variations and partial occlusion. Consequently, it achieves substantial reductions in false positives and identity switches compared to earlier methods, showcasing its practical viability.

Furthermore, a Diverse Detector Integration (DDI) study meticulously assessed the influence of detector choices—including YOLO v4, Faster R-CNN, MobileNet SSD v2, and Deep SORT—on the efficacy of metric learning-based tracking. The results reveal that YOLO v4 consistently delivers exceptional tracking accuracy on both the MOT-16 and MOT-17 datasets, establishing its superiority in this competitive landscape.

In conclusion, this body of research notably advances occlusion-aware person Re-ID by illustrating a clear progression from metric learning to pose-guided feature extraction and ultimately to transformer-based global attention modeling. The findings underscore that lightweight, meticulously parameterized Vision Transformers can achieve impressive generalization for occlusion detection, even under constrained data scenarios. This opens up exciting prospects for integrated detection, localization, and re-identification in real-world surveillance systems, promising to enhance their effectiveness and reliability.


Past Defense Notices

Dates

Theresa Moore

Array Manifold Calibration for Multichannel SAR Sounders

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

James Stiles, Chair
Shannon Blunt
Carl Leuschen
John Paden
Leigh Stearns

Abstract

Multichannel synthetic aperture radar (SAR) ice sounders rely on parametric angle estimators in tomography to resolve elevation angle beyond the Rayleigh resolution limit of their cross-track arrays. The potential super resolution capability of these techniques is predicated on perfect knowledge of the array’s response to directional sources, referred to as the array manifold. Array manifold calibration improves angle estimator performance by reducing the mismatch between the model of the array’s transfer function and truth; its study straddles the fields of both signal processing and antenna theory, yet associated literature reveals dichotomous methodologies that perpetuate fragmented interpretations of the manifold calibration problem. This dissertation addresses calibration for SAR ice sounders that three dimensionally image ice sheet and glacier beds with tomographic techniques. The approach is rooted in array signal processing first but seeks a more unifying perspective of the manifold calibration problem by leveraging commercial computational electromagnetics software to understand error mechanisms and algorithm performance with a deterministic model of an electromagnetic manifold. The research outlined here proposes creation of large snapshot databases that aid in identifying calibration targets in SAR pixels with known arrival angles. The signal processing methodology taxonomizes manifold calibration into parametric and nonparametric forms and advances both in the context of SAR sounders. A parametric estimator of nonlinear manifold parameters that are common across disjoint sets is derived. The algorithm framework capitalizes on a snapshot database to aggregate many angularly diverse observations in estimating unknown model parameters. The technique, which handles multitarget calibration, is desirable in the SAR sounder problem but requires a parametric model of the angle-dependent manifold. Nonparametric calibration techniques characterize the array response over the field of view but require many observations of single sources over dense calibration grids. A subspace clustering technique is proposed to identify snapshots with a single dominant source, thereby enabling a principal components-based characterization of the sounder manifold. The measured manifold leads to significant performance improvements over the traditional array response model in tomography. These results indicate that manifold calibration will reduce uncertainty in sounder-derived maps of the subsurface, leading to more accurate estimates of total fresh ice volume.


Shravan Kaundinya

Investigative Development of an UWB radar for UAS-borne applications

When & Where:


Nichols Hall, Room 317

Committee Members:

Carl Leuschen, Chair
Christopher Allen
Fernando Rodriguez-Morales
Emily Arnold

Abstract

Over the last few years, one of the primary focuses in engineering development has been system packaging and miniaturization. This is apparent in various areas such as the rise of Internet of Things (IoT), CubeSats, and Unmanned Aerial Systems (UAS). The simultaneous miniaturization in multiple industries has enabled advancements in remote sensing instrument development. Sensors such as radars, lidars, and cameras are used on UAS to characterize various aspects of the Earth System like ice, soil, and vegetation, thereby improving our understanding. In this work, an Ultra-wideband (UWB) radar system design for the Vapor 55 UAS rotorcraft is investigated. A compact, lightweight 2 – 18 GHz Frequency Modulated Continuous Wave (FMCW) radar with two channels on transmit and receive is designed to characterize extended targets like soil and snow. This thesis reports initial proof-of-concept field measurements performed with soil as the target to identify backscatter signatures that are indicative of moisture content. The thesis also describes the exploratory design, development, and laboratory test results of the miniaturized radar electronics and compact antenna front-end.


Zeus Gannon

Designing a SODAR testbed for RADAR applications

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Christopher Allen, Chair
Shannon Blunt
James Stiles


Abstract

In research there exists a need to constantly test and develop systems. Testing a radar system requires costly resources in terms of equipment and spectrum. These challenges relegate most testing to simulations, which are a poor approximation of reality. An alternative to over-the-air radar testing is presented here in the form of an over-the-air ultrasonic detection and ranging (SODAR) system. This system takes advantage of the similar wave-like propagation properties of acoustic and electromagnetic waves. With a SODAR testbed, radar waveform design can quickly move out of simulation and into the real world with minimal overhead. In this thesis, basic and advanced radar sensing techniques are demonstrated with a SODAR setup. Range detection, Doppler sensing, and pulse compression are shown as examples of basic radar concepts. For advanced sensing applications, array-based direction finding and synthetic aperture radar (SAR) are shown.


Usman Sajid

Effective Uni-modal to Multi-modal Crowd Estimation based on Deep Neural Networks

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Taejoon Kim, Chair
Fengjun Li
Bo Luo
Cuncong Zhong
Guanghui Wang

Abstract

Crowd estimation is a vital component of crowd analysis. It finds many applications in real-world scenarios, e.g., huge gatherings management like Hajj, sporting and musical events, or political rallies. Automated crowd counting facilitates better and effective management of such events and consequently prevents any undesired situation. This is a very challenging problem in practice since there exists a significant difference in the crowd number in and across different images, varying image resolution, large perspective, severe occlusions, and dense crowd-like cluttered background regions. Current approaches do not handle huge crowd diversity well and thus perform poorly in cases ranging from extreme low to high crowd-density, thus, yielding huge crowd underestimation or overestimation. Also, manual crowd counting proves to be infeasible due to very slow and inaccurate results. To address these major crowd counting issues and challenges, we investigate two different types of input data: uni-modal (image) and multi-modal (image and audio). 

In the uni-modal setting, we propose and analyze four novel end-to-end crowd counting networks, ranging from multi-scale fusion-based models to uni-scale one-pass and two-pass multi-task networks. The multi-scale networks employ the attention mechanism to enhance the model efficacy. On the other hand, the uni-scale models are well-equipped with novel and simple-yet-effective patch re-scaling module (PRM) that functions identical but is more lightweight than multi-scale approaches. Experimental evaluation demonstrates that the proposed networks outperform the state-of-the-art in majority cases on four different benchmark datasets with up to 12.6% improvement for the RMSE evaluation metric. Better cross-dataset performance also validates the better generalization ability of our schemes. For the multi-modal input, effective feature-extraction (FE) and strong information fusion between two modalities remain a big challenge. Thus, the multi-modal novel network design focuses on investigating different features fusion techniques amid improving the FE. Based on the comprehensive experimental evaluation, the proposed multi-modal network increases the performance under all standard evaluation criteria with up to 33.8% improvement in comparison to the state-of-the-art. The application of multi-scale uni-modal attention networks also proves more effective in other deep learning domains, as demonstrated successfully on seven different scene-text recognition task datasets with better performance.


Giordanno Castro Garcia

pyCatalstReader: Extracting Text and Tokenization of Technical

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Michael Branicky , Chair
Fengjun Li
Bo Luo
Kevin Leonard

Abstract

Catalysts are an essential and ubiquitous component of our modern life, from empowering our agriculture to reducing toxic emissions. There is a constant need for more and better catalysts.  The catalysis research literature is immense, growing, and scattered.   Natural Language Processing (NLP), a sub-field of Machine Learning (ML), offers a potential solution to automatically make full use of all this valuable information and speed innovation. Even though NLP has made much progress in the analysis of everyday text, its application in more technical text has not been as successful.  Specifically, there are even a dearth of tools that can appropriately extract text from the PDF files of research articles, which are the most common format used in the catalyst field. Therefore, this project aims to define a tool that can extract text out PDF files of catalysis science articles, which is prerequisite to applying NLP and ML tools.  We also explore the first stage of the NLP pipeline, tokenization, by objectively comparing different tokenizers for catalysis science articles.


Sai Manudeep Gadde

Landmark Classification and Tagging using Convolutional Neural Networks

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Prasad Kulkarni, Chair
Michael Branicky
Esam Al-Araby


Abstract

Photo sharing and photo storage services like to have location data for each photo that is uploaded. With the location data, these services can build advanced features, such as automatic suggestion of relevant tags or automatic photo organization, which help provide a compelling user experience. Although a photo's location can often be obtained by looking at the photo's metadata, many photos uploaded to these services will not have location metadata available. This can happen when, for example, the camera capturing the picture does not have GPS or if a photo's metadata is scrubbed due to privacy concerns.

If no location metadata for an image is available, one way to infer the location is to detect and classify a discernible landmark in the image.  Given the large number of landmarks across the world and the immense volume of images that are uploaded to photo sharing services, using human judgement to classify these landmarks would not be feasible. In this project, we aim to address this problem by building models to automatically predict the location of the image based on any landmarks depicted in the image. We will go through the machine learning design process end-to-end: performing data preprocessing, designing and training CNNs, comparing the accuracy of different CNNs, and using some own images to heuristically evaluate the best CNN.


Dalton Brucker-Hahn

Anvil: Flexible and Dynamic Service Mesh Security Design for Microservice Architectures and Future Network Security Research

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Alexandru Bardas, Chair
Drew Davidson
Fengjun Li
Bo Luo
Huazhen Fang

Abstract

Modern cloud computing environments are evolving with a focus upon speed of deployments, frequency of changes, and a greater adoption of microservice architectures.  To handle these high-level business goals, an emerging series of tools and methodologies referred to as DevOps have been adopted to handle the dynamic and flexible environments being employed in enterprise software.  A popular class of tools within the DevOps toolset are service meshes which aim to manage and connect swarms of microservices.  Service meshes are also responsible for providing service discovery and security for the requests and responses occurring between microservices in a deployment.

Previous work has demonstrated several shortcomings and design limitations in existing, state-of-art service meshes.  Due to this, studies focusing upon improving the security and providing dynamic solutions to these challenges have been proposed but fall short of addressing the issue.  This work will propose a novel design to better address the existing challenges and security needs within this domain.  Anvil, a novel, proof-of-concept service mesh will be designed, implemented, and evaluated

with the trade-off of security and performance in mind.  The goal of Anvil is to provide a security-focused service mesh that can be extended and modified as needed for future research efforts involving service meshes and service mesh design.  With flexibility and extensibility as primary design considerations, future research efforts within the domain of zero-trust networking and distributed system security will be explored and evaluated leveraging Anvil as the underlying service mesh infrastructure.  The potential design and security benefits to the domain of microservice architectures by utilizing Anvil as a testbed and platform for security research is immense.


Laurynas Lialys

Near-Infrared Coherent Raman Spectroscopy and Microscopy

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Shima Fardad, Chair
Rongqing Hui
Alessandro Salandrino


Abstract

Coherent Raman Scattering (CRS) spectroscopy and microscopy is a widely used technique in biology, chemistry, and physics to determine the chemical structure as well as provide a label-free image of the sample. The system uses two coherent laser beams one of which is constantly tuned in wavelength. Thus, a tunable laser source or optical parametric oscillator (OPO) is commonly used to achieve this requirement. However, the aforementioned devices are extremely expensive and work only for a specific wavelength range. In this study, we replace an OPO system with a photonic crystal fiber (PCF) in order to significantly reduce the cost and increase the flexibility of our microscopy system. Here, by exploiting the nonlinear phenomenon in the fiber called the soliton self-frequency shift (SSFS), we are able to shift the pulse central frequency while preserving its shape. Also, by switching to a near-infrared (NIR) source, the undesired fluorescence is reduced while the penetration depth increases. Moreover, the NIR laser source is more biologically friendly as each photon carries less energy than the visible laser counterpart. This reduces the probability of the photodamage effect. Based on this system, we designed and implemented CRS microscopy and spectroscopy, using Coherent anti-Stokes Raman Scattering (CARS) and Stimulated Raman Scattering (SRS) spectroscopy techniques. 


Lazarus Sandhagala Francis

Sentiment Analysis for detecting depression through Social Media Posts

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Prasad Kulkarni, Chair
David Johnson, Co-Chair
Michael Branicky


Abstract

Depression is a common and serious medical condition that negatively affects how one thinks, feels, and acts. Emotional symptoms of depression include loss of interest and/or sad mood. Lack of hope, a sense of guilt or worthlessness, and recurring thoughts of death or suicide are also reported in some cases. After the recent pandemic, depression rates have increased dramatically. Although depression is a major burden for the healthcare system worldwide, it is treatable. Only 47.3% of mental health cases are detected accurately by professionals as Patient Health Questionnaire is used as a screening tool that is heavily dependent on what the patient can remember from the past few weeks. Considering the challenges Healthcare professionals are facing, we can supply helpful resources to those users who have been detected with any depressive symptoms from their social media posts. As social media platforms have altered our world, most people are now connected than ever and are showing a digital persona. We can use all the user-generated content to help them. Sentiment Analysis, also called opinion mining, is a process of detecting the emotional tone behind any piece of text. It is majorly used to analyze news articles, User-generated content, and the text of research papers. This project aims to create a dataset by scrapping tweets and detecting a probably depressed twitter user based on their tweets using Natural Language Processing techniques. Currently, Social media platforms like Twitter have A.I. systems that flag tweets about misinformation, misleading tweets, or those tweets that violate the site’s terms and conditions. Like that, we can also have a depression detection system that will supply users who are probably exhibiting depressive emotions with helpful articles, images, or videos.


Ashwin Rathore

Wireless Communications for Unmanned Vehicles in the Sky and on the Ground

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Morteza Hashemi, Chair
David Johnson
Prasad Kulkarni


Abstract

Given the ever-increasing use of unmanned aerial vehicles (UAV), there are great potentials as well strict requirements for their safe operation in beyond-visual-line-of-sight (BVLOS) environments. Commercial package delivery, emergency services, tracking, inspection, arejust some of those applications. To support these applications under the BVLOS scenarios, a reliable command and control (C2) communication channel with an extended range is needed. To investigate performance of different communication technologies, we use an open-source simulator that integrates the flight simulator ArduPilot with the network simulator NS-3. We implement several flight missions and investigate the performance of 4G cellular network compared with Wi-Fi for establishing the connection between the UAV and groundcontrol station (GCS). Our simulation results demonstrate the benefits of using 4G to satisfy the C2 requirements. Our simulated flight mission consists of multiple UAVs on the same network and also using external interference to observe network performance in terms of average delay, communication range, and received signal strength. In the second part of this project, we explore wireless connectivity between unmanned (autonomous) vehicles on the ground. To this end, we use Amazon’s Deepracer autonomous car that is primarily used for developing and testing machine learning algorithms for multi-vehicle racing, track completion, and obstacle avoidance. We leverage Deepraccer cars to establish peer-to-peer wireless connection between multiple vehicles operating in the same environment. This will enable autonomous vehicles to share crucial information such as positions, velocity, obstacle,and accidents on the way to enhance roads safety.