Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

David Felton

Optimization and Evaluation of Physical Complementary Radar Waveforms

When & Where:


Nichols Hall, Room 129 (Apollo Auditorium)

Committee Members:

Shannon Blunt, Chair
Rachel Jarvis
Patrick McCormick
James Stiles
Zsolt Talata

Abstract

The RF spectrum is a precious, finite resource with ever-increasing demand. Consequently, the mandate to be a "good spectral neighbor" is in direct conflict with the requirements for high-performance sensing where correlation error is fundamentally limited. As such, matched-filter radar performance is often sidelobe-limited with estimation error being constrained by the time-bandwidth (TB) of the collective emission. The methods developed here seek to bridge this gap between idealized radar performance and practical utility via waveform design.    

Estimation error becomes more complex when employing pulse-agility. In doing so, range-sidelobe modulation (RSM) spreads energy across Doppler, rendering traditional methods ineffective. To address this, the gradient-based complementary-FM framework was developed to produce complementary sidelobe cancellation (CSC) after coherently combining subsets within a pulse-agile emission. In contrast to the majority of complementary signals, explored via phase-coding, these Comp-FM waveform subsets achieve CSC while preserving hardware-compatibility since they are FM (though design distortion is never completely avoided). Although Comp-FM addressed practicality via hardware amenability, CSC was localized to zero-Doppler. This work expands the Comp-FM notion to a Doppler-generalized (DG) framework, extending the cancellation condition to an arbitrary span. The same framework can likewise be employed to jointly optimize an entire coherent processing interval (CPI) to minimize RSM within the radar point-spread-function (PSF), thereby generalizing the notion of complementarity and introducing the potential for cognitive operation if sufficient scattering knowledge is available a-priori.          

Sensing with a single emitter is limited by self-inflicted error alone (e.g., clutter, sidelobes), while MIMO systems must additionally contend with the cross-responses from emitters operating concurrently (e.g., simultaneously, spatially proximate, in a shared spectrum), further degrading radar sensitivity. Now, total correlation error is dictated by the overlapping TB (i.e., how coincident are the signals) and number of operating emitters, compounding difficulty to estimate if left unaddressed. As such, the determination of "orthogonal waveforms" comprises a large portion of MIMO literature, though remains a phenomenological misnomer for pulsed emissions. Here, the notion of complementary-FM is applied to a multi-emitter context in which transmitter-amenable quasi-orthogonal subsets, occupying the same spectral band, are produced via a similar gradient-based approach. To further practicalize these MIMO-Comp-FM waveform subsets, the same "DG" approach described above, addressing the otherwise-default Doppler-induced degradation of complementary signals, is applied. In doing so, Doppler-independent separability and complementarity greatly improves estimation sensitivity for multi-emitter systems. 

This MIMO-Comp-FM framework is developed for standard matched filter processing. Coupling this framework with a "DG" form of the previously explored MIMO-MiCRFt is also investigated, illustrating the added benefit of pairing optimized subsets with similarly calibrated processing. 

Each of these methods is developed to address unique and increasingly complex sources of estimation error. All approaches are initially developed and evaluated via simulated analysis where ground-truth is known. Then, despite hardware-induced distortion being unavoidable, the MIMO-Comp-FM framework is confirmed via loopback measurements to preserve the majority of CSC that was observed in simulation. Finally, open-air demonstration of each approach validates practical utility on a radar system.


Hao Xuan

Toward an Integrated Computational Framework for Metagenomics: From Sequence Alignment to Automated Knowledge Discovery

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Cuncong Zhong, Chair
Fengjun Li
Suzanne Shontz
Hongyang Sun
Liang Xu

Abstract

Metagenomic sequencing has become a central paradigm for studying complex microbial communities and their interactions with the host, with emerging applications in clinical prediction and disease modeling. In this work, we first investigate two representative application scenarios: predicting immune checkpoint inhibitor response in non-small cell lung cancer using gut microbial signatures, and characterizing host–microbiome interactions in neonatal systems. The proposed reference-free neural network captures both compositional and functional signals without reliance on reference genomes, while the neonatal study demonstrates how environmental and genetic factors reshape microbial communities and how probiotic intervention can mitigate pathogen-induced immune activation.

These studies highlight both the promise and the inherent difficulty of metagenomic analysis: transforming raw sequencing data into clinically actionable insights remains an algorithmically fragmented and computationally intensive process. This challenge arises from two key limitations: the lack of a unified algorithmic foundation for sequence alignment and the absence of systematic approaches for selecting and organizing analytical tools. Motivated by these challenges, we present a unified computational framework for metagenomic analysis that integrates complementary algorithmic and systems-level solutions.

First, to resolve fragmentation at the alignment level, we develop the Versatile Alignment Toolkit (VAT), a unified algorithmic system for biological sequence alignment across diverse applications. VAT introduces an asymmetric multi-view k-mer indexing scheme that integrates multiple seeding strategies within a single architecture and enables dynamic seed-length adjustment via longest common prefix (LCP)–based inference without re-indexing. A flexible seed-chaining mechanism further supports diverse alignment scenarios, including collinear, rearranged, and split alignments. Combined with a hardware-efficient in-register bitonic sorting algorithm and dynamic index-loading strategy, VAT achieves high efficiency and broad applicability across read mapping, homology search, and whole-genome alignment. Second, to address the challenge of tool selection and pipeline construction, we develop SNAIL, a natural language processing system for automated recognition of bioinformatics tools from large-scale and rapidly growing scientific literature. By integrating XGBoost and Transformer-based models such as SciBERT, SNAIL enables structured extraction of analytical tools and supports automated, reproducible pipeline construction.

Together, this work establishes a unified framework that is grounded in real-world applications and addresses key bottlenecks in metagenomic analysis, enabling more efficient, scalable, and clinically actionable workflows.


Pramil Paudel

Learning Without Seeing: Privacy-Preserving and Adversarial Perspectives in Lensless Imaging

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Fengjun Li, Chair
Alex Bardas
Bo Luo
Cuncong Zhong
Haiyang Chao

Abstract

Conventional computer vision relies on spatially resolved, human-interpretable images, which inherently expose sensitive information and raise privacy concerns. In this study, we explore an alternative paradigm based on lensless imaging, where scenes are captured as diffraction patterns governed by the point spread function (PSF). Although unintelligible to humans, these measurements encode structured, distributed information that remains useful for computational inference. 

We propose a unified framework for privacy-preserving vision that operates directly on lensless sensor measurements by leveraging their frequency-domain and phase-encoded properties. The framework is developed along two complementary directions. First, we enable reconstruction-free inference by exploiting the intrinsic obfuscation of lensless data. We show that semantic tasks such as classification can be performed directly on diffraction patterns using models tailored to non-local, phase-scrambled representations. We further design lensless-aware architectures and integrate them into practical pipelines, including a Swin Transformer-based steganographic framework (DiffHide) for secure and imperceptible information embedding. To assess robustness, we formalize adversarial threat models and develop defenses against learning-based reconstruction attacks, particularly GAN-driven inversion. Second, we investigate the limits of privacy by studying the reconstructability of lensless measurements without explicit knowledge of the forward model. We develop learning-based reconstruction methods that approximate the inverse mapping and analyze conditions under which sensitive information can be recovered. Our results demonstrate that lensless measurements enable effective vision tasks without reconstruction, while providing a principled framework to evaluate and mitigate privacy risks. 


Sharmila Raisa

Digital Coherent Optical System: Investigation and Monitoring

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Rongqing Hui, Chair
Morteza Hashemi
Erik Perrins
Alessandro Salandrino
Jie Han

Abstract

Coherent wavelength-division multiplexed (WDM) optical fiber systems have become the primary transmission technology for high-capacity data networks, driven by the explosive bandwidth demand of cloud computing, streaming services, and large-scale artificial intelligence training infrastructure. This dissertation investigates two fundamental aspects of digital coherent fiber optic systems under the unifying theme of source and monitoring: the design of multi-wavelength optical sources compatible with high-order coherent detection, and the leveraging of fiber Kerr-effect nonlinearity at the coherent receiver to perform physical-layer link health monitoring and to assess inherent security vulnerabilities — both achieved through digital signal processing of the received complex optical field without dedicated hardware.

We begin by addressing the multi-wavelength transmitter challenge in WDM coherent systems. Existing quantum-dot, quantum-dash, and quantum-well based optical frequency comb (OFC) sources share a common limitation: individual comb line linewidths in the tens of MHz range caused by low output power levels of 1–20 mW, making them incompatible with high-order coherent detection. We demonstrate coherent system application of a single-section InGaAsP QW Fabry-Perot laser diode with greater than 120 mW optical power at the fiber pigtail and 36.14 GHz mode spacing. The high optical power per mode produces Lorentzian equivalent linewidths below 100 kHz — compatible with 16-QAM carrier phase recovery without optical phase locking. Experimental results obtained using a commercial Ciena WaveLogic-Ai coherent transceiver demonstrate 20-channel WDM transmission over 78.3 km of standard single-mode fiber with all channels below the HD-FEC threshold of 3.8 × 10⁻³ at 30 GBaud differential-coded 16-QAM, corresponding to an aggregate capacity of 2.15 Tb/s from a single laser device.

After investigating the QW Fabry-Perot laser as a multi-wavelength source for coherent WDM transmission, we leverage the coherent receiver DSP to exploit fiber Kerr-effect nonlinearity for longitudinal power profile estimation, enabling reconstruction of the signal power distribution P(z) along the full multi-span link without dedicated hardware or traffic interruption. We propose a modified enhanced regular perturbation (ERP) method that corrects two independent physical error sources of the standard RP1 least-squares baseline: the accumulated nonlinear phase rotation, and the dispersion-mediated phase-to-intensity conversion — a second bias source not addressed by prior methods. The RP1 method produces mean absolute error (MAE) that scales quadratically with span count, growing to 1.656 dB at 10 spans and 3 dBm. The modified ERP reduces this to 0.608 dB — an improvement that grows consistently with link length, confirming increasing advantage in the long-haul regime. Extension to WDM through an XPM-aware per-channel formulation achieves MAE of 0.113–0.419 dB across 150–500 km link lengths.

In addition to its role in enabling DSP-based longitudinal power profile estimation, the fiber Kerr-effect nonlinearity is shown to give rise to an inherent physical-layer security vulnerability in coherent WDM systems. We show that an eavesdropper co-tenanting a shared fiber — transmitting a continuous-wave probe at a wavelength adjacent to the legitimate signal — can capture the XPM-induced waveform at the fiber output and apply a bidirectional gated recurrent unit neural network, trained on split-step Fourier method simulation data, to reconstruct the transmitted symbol sequence without physical fiber access and without perturbing the legitimate signal. This eavesdropping mechanism is experimentally validated using a commercial Ciena WaveLogic-Ai coherent transceiver for ASK, BPSK, QPSK, and 16-QAM modulation formats at 4.26 GBaud and 8.56 GBaud over one- and two-span 75 km fiber systems, achieving zero symbol errors under high-OSNR conditions. Noise-aware training over OSNR from 20 to 60 dB maintains symbol error rate below 10⁻² for OSNR above 25–30 dB.

Together, these three contributions demonstrate that the coherent fiber optic system is a versatile physical instrument extending well beyond its role as a data transmission medium. The coherent receiver infrastructure — deployed for high-order modulation and data recovery — simultaneously enables the high-power OFC laser to serve as a practical multi-wavelength transmitter source, and provides the complex field measurement capability through which fiber Kerr-effect nonlinearity can be exploited constructively for distributed link monitoring and, as a direct consequence, reveals an inherent physical-layer security exposure in shared fiber infrastructure. This unified perspective on the coherent system as both a transmission platform and a general-purpose measurement instrument has direct relevance to the design of spectrally efficient, self-monitoring, and physically secure optical interconnects for next-generation AI computing networks.


Arman Ghasemi

Task-Oriented Data Communication and Compression for Timely Forecasting and Control in Smart Grids

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Morteza Hashemi, Chair
Alexandru Bardas
Prasad Kulkarni
Taejoon Kim
Zsolt Talata

Abstract

Advances in sensing, communication, and intelligent control have transformed power systems into data-driven smart grids, where forecasting and intelligent decision-making are essential components. Modern smart grids include distributed energy resources (DERs), renewable generation, battery energy storage systems, and large numbers of grid-edge devices that continuously generate time-series data. At the same time, increasing renewable penetration introduces substantial uncertainty in generation, net load, and market operations, while communication networks impose bandwidth, latency, and reliability constraints on timely data delivery. This dissertation addresses how time-series forecasting, data compression, and task-oriented wireless communication can be jointly designed for smart grid applications.

First, we study weather-aware distributed energy management in prosumer-centric microgrids and show that incorporating day-ahead weather information into decision-making improves battery dispatch and reduces the impact of renewable uncertainty. Second, we introduce forecasting-aware energy management in both wholesale and retail electricity markets, highlighting how renewable generation forecasting affects pricing, scheduling, and uncertainty mitigation. Third, we develop and evaluate deep learning methods for renewable generation forecasting, showing that Transformer-based models outperform recurrent baselines such as RNN and LSTM for wind and solar prediction tasks.

Building on this forecasting foundation, we develop a communication-efficient forecasting framework in which high-dimensional smart grid measurements are compressed into low-dimensional latent representations before transmission. This framework is extended into a task-oriented communication system that jointly optimizes data relevance and information timeliness, so that the receiver obtains compressed updates that remain useful for downstream forecasting tasks. Finally, we extend this framework to a distributed multi-node uplink setting, where multiple grid sensors share a bandwidth-limited channel, and develop scheduling policy that improves both the timeliness and task-relevance of received updates.


Pardaz Banu Mohammad

Towards Early Detection of Alzheimer’s Disease based on Speech using Reinforcement Learning Feature Selection

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Arvin Agah, Chair
David Johnson
Sumaiya Shomaji
Dongjie Wang
Sara Wilson

Abstract

Alzheimer’s Disease (AD) is a progressive, irreversible neurodegenerative disorder and the leading cause of dementia worldwide, affecting an estimated 55 million people globally. The window of opportunity for intervention is demonstrably narrow, making reliable early-stage detection a clinical and scientific imperative. While current diagnostic techniques such as neuroimaging and cerebrospinal fluid (CSF) biomarkers carry well-defined limitations in scalability, cost, and access equity, speech has emerged as a compelling non-invasive proxy for cognitive function evaluation.

This work presents a novel approach for using acoustic feature selection as a decision-making technique and implements it using deep reinforcement learning. Specifically, we use a Deep-Q-Network (DQN) agent to navigate a high dimensional feature space of over 6,000 acoustic features extracted using the openSMILE toolkit, dynamically constructing maximally discriminative and non-redundant features subsets. In order to capture the latent structural dependencies among

acoustic features which classifier and wrapper methods have difficulty to model, we introduce the Graph Convolutional Network (GCN) based correlation awareness feature representation layer that operates as an auxiliary input to the DQN state encoder. Post selection interpretability is reinforced through TF-IDF weighting and K-means clustering which together yield both feature level and cluster level explanations that are clinically actionable. The framework is evaluated across five classifiers, namely, support vector machines (SVM), logistic regression, XGBoost, random forest, and feedforward neural network. We use 10-fold stratified cross-validation on established benchmarks of datasets, including DementiaBank Pitt Corpus, Ivanova, and ADReSS challenge data. The proposed approach is benchmarked against state-of-the-art feature selection methods such as LASSO, Recursive feature selection, and mutual information selectors. This research contributes to three primary intellectual advances: (1) a graph augmented state representation that encodes inter-feature relational structure within a reinforcement learning agent, (2) a clinically interpretable pipeline that bridges the gap between algorithmic performance and translational utility, and (3) multilingual data approach for the reinforcement learning agent framework. This study has direct implications for equitable, low-cost and scalable AD screening in both clinical and community settings.


Zhou Ni

Bridging Federated Learning and Wireless Networks: From Adaptive Learning to FLdriven System Optimization

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Morteza Hashemi, Chair
Fengjun Li
Van Ly Nguyen
Han Wang
Shawn Keshmiri

Abstract

Federated learning (FL) has emerged as a promising distributed machine learning
framework that enables multiple devices to collaboratively train models without sharing raw
data, thereby preserving privacy and reducing the need for centralized data collection. However,
deploying FL in practical wireless environments introduces two major challenges. First, the data
generated across distributed devices are often heterogeneous and non-IID, which makes a single
global model insufficient for many users. Second, learning performance in wireless systems is
strongly affected by communication constraints such as interference, unreliable channels, and
dynamic resource availability. This PhD research aims to address these challenges by bridging
FL methods and wireless networks.
In the first thrust, we develop personalized and adaptive FL methods given the underlying
wireless link conditions. To this end, we propose channel-aware neighbor selection and
similarity-aware aggregation in wireless device-to-device (D2D) learning environments. We
further investigate the impacts of partial model update reception on FL performance. The
overarching goal of the first thrust is to enhance FL performance under wireless constraints.
Next, we investigate the opposite direction and raise the question: How can FL-based distributed
optimization be used for the design of next-generation wireless systems? To this end, we
investigate communication-aware participation optimization in vehicular networks, where
wireless resource allocation affects the number of clients that can successfully contribute to FL.
We further extend this direction to integrated sensing and communication (ISAC) systems,
where personalized FL (PFL) is used to support distributed beamforming optimization with joint
sensing and communication objectives.
Overall, this research establishes a unified framework for bridging FL and wireless networks. As
a future direction, this work will be extended to more realistic ISAC settings with dynamic
spectrum access, where communication, sensing, scheduling, and learning performance must be
considered jointly.


Past Defense Notices

Dates

Ashwin Rathore

Wireless Communications for Unmanned Vehicles in the Sky and on the Ground

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Morteza Hashemi, Chair
David Johnson
Prasad Kulkarni


Abstract

Given the ever-increasing use of unmanned aerial vehicles (UAV), there are great potentials as well strict requirements for their safe operation in beyond-visual-line-of-sight (BVLOS) environments. Commercial package delivery, emergency services, tracking, inspection, arejust some of those applications. To support these applications under the BVLOS scenarios, a reliable command and control (C2) communication channel with an extended range is needed. To investigate performance of different communication technologies, we use an open-source simulator that integrates the flight simulator ArduPilot with the network simulator NS-3. We implement several flight missions and investigate the performance of 4G cellular network compared with Wi-Fi for establishing the connection between the UAV and groundcontrol station (GCS). Our simulation results demonstrate the benefits of using 4G to satisfy the C2 requirements. Our simulated flight mission consists of multiple UAVs on the same network and also using external interference to observe network performance in terms of average delay, communication range, and received signal strength. In the second part of this project, we explore wireless connectivity between unmanned (autonomous) vehicles on the ground. To this end, we use Amazon’s Deepracer autonomous car that is primarily used for developing and testing machine learning algorithms for multi-vehicle racing, track completion, and obstacle avoidance. We leverage Deepraccer cars to establish peer-to-peer wireless connection between multiple vehicles operating in the same environment. This will enable autonomous vehicles to share crucial information such as positions, velocity, obstacle,and accidents on the way to enhance roads safety.


Gordon Ariho

MULTIPASS SAR PROCESSING FOR ICE SHEET VERTICAL VELOCITY AND TOMOGRAPHY MEASUREMENTS

When & Where:


Nichols Hall, Room 317

Committee Members:

James Stiles, Chair
John Paden
Christopher Allen
Shannon Blunt
Carl Leuschen

Abstract

Ice dynamics are a major factor in ice sheet mass balance and play a huge role in sea level rise (and future sea-level rise projections). Ice velocity measures the direction and rate at which ice is redistributed from the accumulation to the ablation regions of glaciers and ice sheets. We propose to apply multipass differential interferometric synthetic aperture radar (DInSAR) techniques to data from the Multichannel Coherent Radar Depth Sounder (MCoRDS) to measure the vertical displacement of englacial layers within an ice sheet. DInSAR’s accuracy is usually on the order of a small fraction of the wavelength (e.g. millimeter to centimeter precision is common) in monitoring ground displacement along the radar line of sight (LOS).  Unlike ground-based Autonomous phase-sensitive Radio-Echo Sounder (ApRES) units that can be precisely positioned and used to produce vertical velocity fields, airborne systems suffer from unknown baseline errors. In the case of ice sheet internal layers, vertical displacement is estimated by compensating for the spatial baseline using precise trajectory information and estimates of the cross-track layer slope from direction of arrival analysis. The current DInSAR algorithm is applied to radar depth sounder data to produce results for Summit camp in central Greenland and a high accumulation region near Camp Century in northwest Greenland using the CReSIS toolbox. This approach has a drawback arising from the baseline error due to the GPS being estimated after Direction of Arrival (DOA) estimation yet DOA estimation is dependent on the baseline being accurate. We propose to extend this work by implementing a maximum likelihood estimator that jointly estimates the vertical velocity, the cross-track internal layer slope, and the unknown baseline error due to GPS and INS (Inertial Navigation System) errors. The multipass algorithm will be applied to additional flights from the decade long NASA Operation IceBridge airborne mission that flew MCoRDS on many repeated flight tracks. We also propose to improve the accuracy of tomographic swaths produced from multipass measurements and investigate the possibility to use focusing matrices to improve wideband tomographic processing.


Madhu Peduri

Training a Smart cab agent Using a Reinforcement Q – Learning

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Prasad Kulkarni, Chair
David Johnson
Bo Luo


Abstract

Reinforcement learning is a method to map situations to actions to maximize a numerical reward signal. In most forms of machine learning, the model must discover which actions to take unlike reinforcement learning in which the model must discover which actions yield the most reward by trying them. These actions may affect not only the immediate reward, but also the next situation and, through that, all subsequent rewards. This type of learning is different from supervised learning, where domain knowledge comes from an external supervisor. This is an important kind of learning, but alone it is not adequate for learning from interaction. In interactive problems it is often impractical to obtain examples of desired behavior that are both correct and representative of all the situations in which the agent must act and must be able to learn from its own experience. As a part of this project, we attempt to train a Smart-cab agent that will navigate through its environment towards a goal. With following elements as our Reinforcement model, Agent – We use a Car as the agent to interact with the environment. The goal for the agent is to reach the destination with the maximum value; Environment – Our environment is a grid like structure with pathways that represent the roads with cars (5 without the agent) moving along them stochastically; Policy – We have a set of actions and constraints within which states and actions would be mapped. The agent has to perform the appropriate action that results into maximum Q-value. We use the Pygame tool to build our environment to visualize the interaction of the agent with our environment and Q-Learning to find the optimal policy that determines the optimal action that can be taken keeping all the constraints under purview.


Sushmitha Boddi Reddy

Conversational AI Chatbots

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Prasad Kulkarni, Chair
David Johnson, Co-Chair
Zijun Yao


Abstract

We know that AI and Machine learning are being used in every industry and in every experience. Chatbots are among most visible applications of AI technology and since the time chatbots have entered the digital world, every industry wants to use chatbots to help their customers with the common issues. Most of the chatbots are kind of a messaging interface where instead of humans answering, message bots will be responding.

Chatbots are an artificial intelligence software that can initiate a human like conversation with a real human based on the training.AI uses a natural language to communicate with Artificial Intelligence features embedded in them. The conversation humans have with bots is powered by Machine Learning algorithms which breaks down the messages received into human understandable languages using Natural Language Processing techniques and responds to your queries like what you can expect from any human on the other side. In this project I've trained a chatbot using a dataset which responds to our messages.


Kailani Jones

Metrics Identifying Gaps in the SOC's Alert Handling Process

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Alexandru Bardas, Chair
Drew Davidson
Fengjun Li
Bo Luo
John Symons

Abstract

In the recent years, organization's attack surfaces continue to increase with the rise of data storage, application diversity, and ransomware attacks.

The response typically falls to the enterprise Security Operation Center (SOC). However, even with an expanding attack surface, organizations continue to decommission or completely remove their SOCs due to the uncertainty around their respective value. This work traces and analyzes (1) the SOC's effort to reimpose endpoint monitoring and content handling as a result of the Internet's new and different sociotechnical environment resulting from COVID-19's “work from home” and (2) propose a metrics framework that captures the gaps within the SOC analysts' core function: the alert handling process. By intersecting historical analysis (starting in the 1970s) and ethnography (analyzed 256 field notes and performed two rounds of semi-structured interviews across 770+ hours in a SOC over 26 months) whilst complementing with quantitative interviews (covering 7 other SOCs), we find additional causal forces that, for decades, have pushed network management toward endpoints and content. With a similar ethnographic approach (participation observation paired with semi-structured interviews), we further locate expert judgement in the alert handling process and utilize those limitations as key performance indicators to identify gaps and capture the needs within the SOC.


Likitha Vemulapalli

Identification of Foliar Diseases in Plants using Deep Learning Techniques

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Prasad Kulkarni, Chair
David Johnson
Suzanne Shontz


Abstract

Artificial Intelligence has been gathering tremendous support lately by bridging the gap between humans and machines. Amazing discoveries in numerous fields are paving way for state-of-the-art technologies. Deep Learning has shown immense progress in the field of computer vision in recent years, with neural networks repeatedly pushing the frontier of visual recognition technology. Recent years have witnessed an exponential increase in the use of mobile and embedded devices. With great success of deep learning, there is an emerging trend to deploy deep learning models on mobile and embedded devices. However, it is not a simple task, the limited resources of mobile and embedded devices make it challenging to fulfill the intensive computation and storage demand of deep learning models and state-of-the-art Convolutional Neural Networks (CNN) require computation at billions of floating-point operations per second (FLOP) which inhibit them from being utilized in mobile and embedded devices. Mobile convolutional neural networks use depth wise and group convolutions rather than standard “fully-connected” convolutions.  In this project we will be applying mobile convolutional models to identify the diseases in plants. Plant diseases are responsible for serious economic losses every year. Due to various reasons, the crops are affected based on climate conditions, various kinds of diseases, heavy usage of pesticides and many other factors. Due to the rise in use of pesticides, the farmers are experiencing irreplaceable losses. Less use of pesticides can help in better crop production. Using these mobile CNNs we can identify the diseases in plants with leaf images and based on the type of disease pesticides can be used respectively. The main goal is to use an efficient model which can assist farmers in recognizing leaf symptoms and providing targeted information for rational use of pesticides. 


Truc Anh Ngoc Nguyen

ResTP: A Configurable and Adaptable Multipath Transport Protocol for Future Internet Resilience

When & Where:


2001B Eaton Hall

Committee Members:

Victor Frost, Chair
Morteza Hashemi
Taejoon Kim
Bo Luo
Hyunjin Seo

Abstract

Motivated by the shortcomings of common transport protocols, e.g., TCP, UDP, and MPTCP, in modern networking and the belief that a general-purpose transport-layer protocol, which can operate efficiently over diverse network environments while being able to provide desired services for various application types, we design a new transport protocol, ResTP. The rapid advancement of networking technology and use paradigms is continually supporting new applications. The configurable and adaptable multipath-capable ResTP is not only distinct from the standard protocols by its flexibility in satisfying the requirements of different traffic classes considering the characteristics of the underlying networks, but by its emphasis on providing resilience. Resilience is an essential property that is unfortunately missing in the current Internet. In this dissertation, we present the design of ResTP, including the services that it supports and the set of algorithms that implement each service. We also discuss our modular implementation of ResTP in the open-source network simulator ns-3. Finally, the protocol is simulated under various network scenarios, and the results are analyzed in comparison with conventional protocols such as TCP, UDP, and MPTCP to demonstrate that ResTP is a promising new transport-layer protocol providing resilience in the Future Internet (FI).


Dinesh Mukharji Dandamudi

Analyzing the short squeeze caused by Reddit community by Using Machine learning

When & Where:


Zoom defense, please email jgrisafe@ku.edu for the meeting information

Committee Members:

Matthew Moore, Chair
Drew Davidson
Cuncong Zhong


Abstract

Algorithmic trading (sometimes termed automated trading, black-box trading, or algo-trading) is a computerized trading system where a computer program follows a set of specified instructions to make a transaction. Theoretically, the transaction should allow traders to make profits at a rate and frequency that a human trader cannot attain. Algorithmic trading is an automated trading method that is carried out using a computer algorithm. Trade theory theoretically posits that humans cannot earn profits at a pace and frequency comparable to those generated by computers.  

 

Traders have a tough time keeping track of the many handles that originate data. NLP (Natural Language Processing) can be used to rapidly scan various news sources, identifying opportunities to gain an advantage before other traders do. 

 

Based on this background, this project aims to select and implement an NLP and Machine Learning process that produces an algorithm, which can detect OR predict the future value from scraped data using Natural language processing and Machine Learning. This algorithm builds the basic structure for an approach to evaluate these documents. 


Lyndon Meadow

Remote Lensing

When & Where:


2001B Eaton Hall

Committee Members:

Matthew Moore, Chair
Perry Alexander
Prasad Kulkarni


Abstract

The problem of the manipulation of remote data is typically solved used complex methods to guarantee consistency. This is an instance of the remote bidirectional transformation problem. From the inspiration that several versions of this problem have been addressed using lenses, we now extend this technique of lenses to the Remote Procedure Calls setting, and provide a few simple example implementations.

    Taking the host side to be the strongly-typed language with lensing properties, and the client side to be a weakly-typed language with minimal lensing properties, this work contributes to the existing body of research that has brought lenses from the realm of math to the space of computer science. This shall give a formal look on remote editing of data in type safety with Remote Monads and their local variants.


Chanaka Samarathungage

NextG Wireless Networks: Applications of the Millimeter Wave Networking and Integration of UAVs with Cellular Systems

When & Where:


Zoom Meeting, please contact jgrisafe@ku.edu for link.

Committee Members:

Morteza Hashemi, Chair
Taejoon Kim
Erik Perrins


Abstract

Considering the growth of wireless and cellular devices and applications, the spectrum-rich millimeter wave (mmWave) frequencies have the potential to alleviate the spectrum crunch that wireless and cellular operators are already experiencing. However, there are several challenges to overcome when using mmWave bands. Since mmWave frequencies have small wavelengths compared to sub-6 GHz bands, most objects such as human body, cause significant additional path losses, which can entirely break the link. Highly directional mmWave links are susceptible to frequent link failures in such environments. Limited range of communication is another challenge in mmWave communications. In this research, we demonstrate the benefits of multi-hop routing in mitigating the blockage and extending communication range in the mmWave band. We develop a hop-by-hop multi-path routing protocol that finds one primary and one backup next-hop per destination to guarantee reliable and robust communication under extreme stress conditions. We also extend our solution by proposing a proactive route refinement scheme for AODV and Backpressure routing protocols under dynamic scenarios.
In the second part, the integration of Unmanned Aerial Vehicles (UAVs) to the NextG cellular systems is considered for various applications such as commercial package delivery, public health and safety, surveying, and inspection, to name a few. We present network simulation results based on 4G and 5G technologies using raytracing software. Based on the results, we propose several network adjustments to optimize 5G network operation for the ground users as well as the UAV users.