Defense Notices


All students and faculty are welcome to attend the final defense of EECS graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

Andrew Mertz

Multiple Input Single Output (MISO) Receive Processing Techniques for Linear Frequency Modulated Continuous Wave Frequency Diverse Array (LFMCW-FDA) Transmit Structures

When & Where:


Nichols Hall, Room 317 (Richard K. Moore Conference Room)

Committee Members:

Patrick McCormick, Chair
Chris Allen
Shannon Blunt
James Stiles

Abstract

This thesis focuses on the multiple processing techniques that can be applied to a single receive element co-located with a Frequency Diverse Array (FDA) transmission structure that illuminates a large volume to estimate the scattering characteristics of objects within the illuminated space in the range, Doppler, and spatial dimensions. FDA transmissions consist of a number of evenly spaced transmitting elements all of which are radiating a linear frequency modulated (LFM) waveform. The elements are configured into a Uniform Linear Array (ULA) and the waveform of each element is separated by a frequency spacing across the elements where the time duration of the chirp is inversely proportional to an integer multiple of the frequency spacing between elements. The complex transmission structure created by this arrangement of multiple transmitting elements can be received and processed by a single receive element. Furthermore, multiple receive processing techniques, each with their own advantages and disadvantages, can be applied to the data received from the single receive element to estimate the range, velocity, and spatial direction of targets in the illuminated volume relative to the co-located transmit array and receive element. Three different receive processing techniques that can be applied to FDA transmissions are explored. Two of these techniques are novel to this thesis, including the spatial matched filter processing technique for FDA transmission structures, and stretch processing using virtual array processing for FDA transmissions. Additionally, this thesis introduces a new type of FDA transmission structure referred to as ”slow-time” FDA.


Amin Shojaei

Exploring Cooperative and Robust Multi-Agent Reinforcement Learning in Networked Cyber-Physical Systems: Applications in Smart Grids

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Morteza Hashemi, Chair
Alex Bardas
Taejoon Kim
Prasad Kulkarni
Shawn Keshmiri

Abstract

Significant advances in information and networking technologies have transformed Cyber-Physical Systems (CPS) into networked cyber-physical systems (NCPS). A noteworthy example of such systems is smart grid networks, which include distributed energy resources (DERs), renewable generation, and the widespread adoption of Electric Vehicle (EV). Such complex NCPS require intelligent and autonomous control solutions. For example, the increasing number of EVs introduces significant sources of demand and user behavior uncertainty that can jeopardize the grid stability during peak hours. Traditional model-based demand-supply controls fail to accurately model and capture the complex nature of smart grid systems in the presence of different uncertainties and as the system size grows. To address these challenges, data-driven approaches have emerged as an effective solution for informed decision-making, predictive modeling, and adaptive control to enhance the resiliency of NCPS in uncertain environments.

As a powerful data-driven approach, Multi-Agent Reinforcement Learning (MARL) enables agents to learn and adapt in dynamic and uncertain environments. However, MARL techniques introduce complexities related to communication, coordination, and synchronization among agents. In this PhD research, we investigate autonomous control for smart grid decision networks using MARL. Within this context, first, we examine the issue of imperfect state information, which frequently arises due to the inherent uncertainties and limitations in observing the system state. Secondly, we investigate the challenges associated with distributed MARL techniques, with a special focus on the central training distributed execution (CTDE) methods. Throughout this research, we highlight the significance of cooperation in MARL for achieving autonomous control in smart grid systems and other cyber-physical domains. Thirdly, we propose a novel robust MARL framework using a hierarchical structure. We perform an extensive analysis and evaluation of our proposed hierarchical MARL model for large-scale EV networks, thereby addressing the scalability and robustness challenges as the number of agents within a NCPS increases.


Sameera Katamaneni

Revolutionizing Forensic Identification: A Dual-Method Facial Recognition Paradigm for Enhanced Criminal Identification

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Prasad Kulkarni, Chair
Hongyang Sun



Abstract

In response to the challenges posed by increasingly sophisticated criminal behaviour that strategically evades conventional identification methods, this research advocates for a paradigm shift in forensic practices. Departing from reliance on traditional biometric techniques such as DNA matching, eyewitness accounts, and fingerprint analysis, the study introduces a pioneering biometric approach centered on facial recognition systems. Addressing the limitations of established methods, the proposed methodology integrates two key components. Firstly, facial features are meticulously extracted using the Histogram of Oriented Gradients (HOG) methodology, providing a robust representation of individualized facial characteristics. Subsequently, a face recognition system is implemented, harnessing the power of the K-Nearest Neighbours machine learning classifier. This innovative dual-method approach aims to significantly enhance the accuracy and reliability of criminal identification, particularly in scenarios where conventional methods prove inadequate. By capitalizing on the inherent uniqueness of facial features, this research strives to introduce a formidable tool for forensic practitioners, offering a more effective means of addressing the evolving landscape of criminal tactics and safeguarding the integrity of justice systems. 


Durga Venkata Suraj Tedla

Bioinformatics pipeline for MicroRNA analysis data from oxford min ION instruments on anti aging research

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Cuncong Zhong, Chair
Zijun Yao



Abstract

In the pursuit of unraveling the intricate mechanisms underlying the aging process, this study presents a sophisticated bioinformatics pipeline tailored for the analysis of MicroRNA (miRNA) data obtained from Oxford MinION instruments. MiRNAs, small non-coding RNA molecules, play pivotal roles in the regulation of gene expression and have emerged as key players in various biological processes, including aging. The utilization of Oxford MinION, a nanopore sequencing platform, allows for high-throughput and real-time sequencing of miRNAs, providing a comprehensive view of the miRNA landscape.

 

Our integrative bioinformatics pipeline encompasses raw data preprocessing, quality control, alignment, and quantification of miRNAs. Leveraging state-of-the-art algorithms, the pipeline enables the identification of differentially expressed miRNAs associated with the aging phenotype. Functional enrichment analysis is employed to elucidate the biological pathways influenced by these miRNAs, shedding light on the molecular intricacies of the aging process. Furthermore, network analysis is employed to unravel potential regulatory interactions among identified miRNAs and their target genes.

 

This comprehensive approach amalgamates cutting-edge sequencing technology with advanced bioinformatics methodologies, offering a robust framework for deciphering the complex regulatory networks governing aging. The findings derived from this pipeline not only enhance our understanding of the molecular basis of aging but also provide potential targets for therapeutic interventions aimed at modulating the aging process.


Christian Jones

Robust and Efficient Structure-Based Radar Receive Processing

When & Where:


Nichols Hall, Room 129 (Apollo Auditorium)

Committee Members:

Shannon Blunt, Chair
Chris Allen
Suzanne Shontz
James Stiles
Zsolt Talata

Abstract

Legacy radar systems largely rely on repeated emission of a linear frequency modulated (LFM) or chirp waveform to ascertain scattering information from an environment. The prevalence of these chirp waveforms largely stems from their simplicity to generate, process, and the general robustness they provide towards hardware effects. However, this traditional design philosophy often lacks the flexibility and dimensionality needed to address the dynamic “complexification” of the modern radio frequency (RF) environment or achieve current operational requirements where unprecedented degrees of sensitivity, maneuverability, and adaptability are necessary.

Over the last couple of decades analog-to-digital and digital-to-analog technologies have advanced exponentially, resulting in tremendous design degrees of freedom and arbitrary waveform generation (AWG) capabilities that enable sophisticated design of emissions to better suit operational requirements. However, radar systems typically require high powered amplifiers (HPA) to contend with the two-way propagation. Thus, transmitter-amenable waveforms are effectively constrained to be both spectrally contained and constant amplitude, resulting in a non-convex NP-hard design problem.

While determining the global optimal waveform can be intractable for even modest time-bandwidth products (TB), locally optimal transmitter-amenable solutions that are “good enough” are often readily available. However, traditional matched filtering may not satisfy operational requirements for these sub-optimal emissions. Using knowledge of the transmitter-receiver chain, a discrete linear model can be formed to express the relationship between observed measurements and the complex scattering of the environment. This structured representation then enables more sophisticated least-square and adaptive estimation techniques to better satisfy operational needs, improve estimate fidelity, and extend dynamic range.

However, radar dimensionality can be enormous and brute force implementations of these techniques may have unwieldy computational burden on even cutting-edge hardware. Additionally, a discrete linear representation is fundamentally an approximation of the dynamic continuous physical reality and model errors may induce bias, create false detections, and limit dynamic range. As such, these structure-based approaches must be both computationally efficient and robust to reality.

Here several generalized discrete radar receive models and structure-based estimation schemes are introduced. Modifications and alternative solutions are then proposed to improve estimate fidelity, reduce computational complexity, and provide further robustness to model uncertainty.


Shawn Robertson

A secure framework for at risk populations in austere environments utilizing Bluetooth Mesh communications

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Alex Bardas, Chair
Drew Davidson
Fengjun Li
Bo Luo
Huazhen Fang

Abstract

Austere environments are defined by the US Military as those regularly experiencing significant environmental hazards, have limited access to reliable electricity, or require prolonged use of body armor or chemical protection equipment.  We propose that in modern society, this definition can extend also to telecommunications infrastructure, areas where an active adversary controls the telecommunications infrastructure and works against the people such as protest areas in Iran, Russia, and China or areas experiencing conflict and war such as Eastern Ukraine.  People in these austere environments need basic text communications and the ability to share simple media like low resolution pictures.  This communication is complicated by the adversaries’ capabilities as a potential nation-state actor. To address this, Low Earth Orbit satellite clusters, like Starlink, can be used to exfiltrate communications completely independent of local infrastructure.  This, however, creates another issue as these satellite ground terminals are not inherently designed to support many users over a large area.  Traditional means of extending this connectivity create both power and security concerns.  We propose that Bluetooth Mesh can be used to extend connectivity and provide communications. 

Bluetooth Mesh provides a low signal footprint to reduce the risk of detection, blends into existent signals within the 2.4ghz spectrum, has security aspects in the specification, and devices can utilize small batteries maintaining a covert form factor.  To realize this security enhancements must be made to both the provisioning process of the Bluetooth Mesh network and a key management scheme that ensures the regular and secure changing of keys either in response to an adversary’s action or as a prevention of an adversary’s action must be implemented.  We propose a provisioning process using whitelists on both provisioner and device and uses attestation for passwords allowing devices to be provisioned on deployment to protect the at-risk population and prevent BlueMirror attacks.  We also propose, implement, and measure the impact of an automated key exchange that meets the Bluetooth Mesh 3 phase specification.  Our experimentation, in a field environment, shows that Bluetooth Mesh has the throughput, reliability and security to meet the requirements of at-risk populations in austere environments. 


Sohan Chandra

Predicting inorganic nitrogen content in the soil using Machine Learning

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Taejoon Kim, Chair
Prasad Kulkarni
Cuncong Zhong


Abstract

This ground-breaking project addresses a critical issue in crop production: precisely determining plant-available inorganic nitrogen (IN) in soil to optimize fertilization strategies. Current methodologies frequently struggle with the complexities of determining a soil's nitrogen content, resorting to approximations and labor-intensive soil testing procedures that can lead to the pitfalls of under or over-fertilization, endangering agricultural productivity. Recognizing the scarcity of historical inorganic nitrogen (IN) data, this solution employs a novel approach that employs Generative Adversarial Networks (GANs) to generate statistically similar inorganic nitrogen (IN) data. 

 

This synthetic data set works in tandem with data from the Decision Support System for Agrotechnology Transfer (DSSAT). To address the data's inherent time-series nature, we use the power of Long Short-Term Memory (LSTM) neural networks in our predictive model. The resulting model is a sophisticated and accurate tool that can provide reliable estimates without extensive soil testing. This not only ensures precision in nutrient management but is also a cost-effective and dependable solution for crop production optimization. 


Thomas Woodruff

Model Predictive Control of Nonlinear Latent Force Models

When & Where:


M2SEC, Room G535

Committee Members:

Jim Stiles, Chair
Michael Branicky
Heechul Yun


Abstract

Model Predictive Control (MPC) has emerged as a potent approach for controlling nonlinear systems in the robotics field and various other engineering domains. Its efficacy lies in its capacity to predictively optimize system behavior while accommodating state and input constraints. Although MPC typically relies on precise dynamic models to be effective, real-world dynamic systems often harbor uncertainties. Ignoring these uncertainties can lead to performance degradation or even failure in MPC.

Nonlinear latent force models, integrating latent uncertainties characterized as Gaussian processes, hold promise for effectively representing nonlinear uncertain systems. Specifically, these models incorporate the state-space representation of a Gaussian process into known nonlinear dynamics, providing the ability to simultaneously predict future states and uncertainties.

This thesis delves into the application of MPC to nonlinear latent force models, aiming to control nonlinear uncertain systems. We formulate a stochastic MPC problem and, to address the ensuing receding-horizon stochastic optimization problem, introduce a scenario-based approach for a deterministic approximation. The resulting scenario-based approach is assessed through simulation studies centered on the motion planning of an autonomous vehicle. The simulations demonstrate the controller's adeptness in managing constraints and consistently mitigating the effects of disturbances. This proposed approach holds promise for various robotics applications and beyond.


Sai Soujanya Ambati

BERT-NEXT: Exploring Contextual Sentence Understanding

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Prasad Kulkarni, Chair
Hongyang Sun



Abstract

The advent of advanced natural language processing (NLP) techniques has revolutionized the way we handle textual data. This project presents the implementation of exploring contextual sentence understanding on the Quora Insincere Questions dataset using the pretrained BERT architecture. In this study, we explore the application of BERT, a bidirectional transformer model, for text classification tasks. The goal is to classify if a question contains hateful, disrespectful or toxic content. BERT represents the state-of-the-art in language representation models and has shown strong performance on various natural language processing tasks. In this project, the pretrained BERT base model is fine-tuned on a sample of the Quora dataset for next sentence prediction. Results show that with just 1% of the data (around 13,000 examples), the fine-tuned model achieves over 90% validation accuracy in identifying insincere questions after 4 epochs of training. This demonstrates the effectiveness of leveraging BERT for text classification tasks with minimal labeled data requirements. Being able to automatically detect toxic, hateful or disrespectful content is important to maintain healthy online discussions. However, the nuances of human language make this a challenging natural language processing problem. Insincere questions may contain offensive language, hate speech, or misinformation, making their identification crucial for maintaining a positive and safe online environment. In this project, we explore using the pretrained Bidirectional Encoder Representations from Transformers (BERT) model for next sentence prediction on the task of identifying insincere questions.


Swathi Koyada

Feature balancing of demographic data using SMOTE

When & Where:


Zoom Meeting, please email jgrisafe@ku.edu for defense link.

Committee Members:

Prasad Kulkarni, Chair
Cuncong Zhong



Abstract

The research investigates the utilization of Synthetic Minority Oversampling Techniques (SMOTE) in the context of machine learning models applied to biomedical datasets, particularly focusing on mitigating demographic data disparities. The study is most relevant to underrepresented demographic data. The primary objective is to enhance the SMOTE methodology, traditionally designed for addressing class imbalances, to specifically tackle ethnic imbalances within feature representation. In contrast to conventional approaches that merely exclude race as a fundamental or additive factor without rectifying misrepresentation, this work advocates an innovative modification of the original SMOTE framework, emphasizing dataset augmentation based on participants' demographic backgrounds. The predominant aim of the project is to enhance and reshape the distribution to optimize model performance for unspecified demographic subgroups during training. However, the outcomes indicate that despite the application of feature balancing in this adapted SMOTE method, no statistically significant enhancement in accuracy was discerned. This observation implies that while rectifying imbalances is crucial, it may not independently suffice to overcome challenges associated with heterogeneity in species representation within machine learning models applied to biomedical databases. Consequently, further research endeavors are necessary to identify novel methodologies aimed at enhancing sampling accuracy and fairness within diverse populations.


Jessica Jeng

Exploiting Data Locality for Improving Multidimensional Variational Quantum Classification

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Esam El-Araby, Chair
Drew Davidson
Prasad Kulkarni


Abstract

Quantum computing presents an opportunity to accelerate machine learning (ML) tasks on quantum processors in a similar vein to existing classical accelerators, such as graphical processing units (GPUs). In the classical domain, convolutional neural networks (CNNs) effectively exploit data locality using the convolution operation to reduce the number of fully-connected operations in multi-layer perceptrons (MLPs). Preserving data locality enables the pruning of training parameters, which results in reduced memory requirements and shorter training time without compromising classification accuracy. However, contemporary quantum machine learning (QML) algorithms do not leverage the data locality of input features in classification workloads, particularly for multidimensional data. This work presents a multidimensional quantum convolutional classifier (MQCC) that adapts the CNN structure to a variational quantum algorithm (VQA). The proposed MQCC uses quantum implementations of multidimensional convolution, pooling based on the quantum Haar transform (QHT) and partial measurement, and fully-connected operations. Time-complexity analysis will be presented to demonstrate the speedup of the proposed techniques in comparison to classical convolution and pooling operations on modern CPUs and/or GPUs. Experimental work is conducted on state-of-the-art quantum simulators from IBM Quantum and Xanadu modeling noise-free and noisy quantum devices. High-resolution multidimensional images are used to demonstrate the correctness and scalability of the convolution and pooling operations. Furthermore, the proposed MQCC model is tested on a variety of common datasets against multiple configurations of related ML and QML techniques. Based on standard metrics such as log loss, classification accuracy, number of training parameters, circuit depth, and gate count, it will be shown that MQCC can deliver a faithful implementation of CNNs on quantum machines. Additionally, it will be shown that by exploiting data locality MQCC can achieve improved classification over contemporary QML methods. 


Ashish Adhikari

Towards Assessing the Security of Program Binaries

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Prasad Kulkarni, Chair
Fengjun Li
Sumaiya Shomaji


Abstract

Software vulnerabilities, stemming from coding weaknesses and poor development practices, have become increasingly prevalent. These vulnerabilities could be exploited by attackers to pose risks to the confidentiality, integrity, and availability of software. To protect themselves, end-users of software may have an interest in knowing if the software they buy and use is secure from such attacks. Our work is motivated by this need to automatically assess and rate the security properties of binary software.

To increase user trust in third-party software, researchers have devised several techniques and tools to identify and mitigate coding weaknesses in binary software. Therefore, our first task in this work is to assess the current landscape and comprehend the capabilities and challenges faced by binary-level techniques aimed at detecting critical coding weaknesses in software binaries. We categorize the most important coding weaknesses in compiled programming languages, and conduct a comprehensive survey, exploration, and comparison of static techniques designed to locate these weaknesses in software binaries. Furthermore, we perform an independent assessments of the efficacy of open-source tools using standard benchmarks.

Next, we develop techniques to assess if secure coding principles were adopted during the generation of the software binary. Towards this goal, we first develop techniques to determine the high-level source language used to produce the binary. Then, we check the feasibility of detecting the use of secure coding best practices during code development. Finally, we check the feasibility of detecting the vulnerable regions of code in any binary executable. Our ultimate future goal is to employ all of our developed techniques to rate the security-quality of the given binary software.


Hunter Glass

MeshMapper: Creating a Bluetooth Mesh Communication Network

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Alex Bardas, Chair
Drew Davidson
Fengjun Li


Abstract

With threat actors ever evolving, the need for secure communications continues to grow. By using non-traditional means as a way of a communication network, it is possible to securely communicate within a region using the bluetooth mesh protocol. The goal is to automatically place these mesh devices in a defined region in order to ensure the integrity and reliability of the network, while also ensuring the least number of devices are placed. By placing a provisioner node, the rest of the specified region populates with mesh nodes that act as relays, creating a network allowing users to communicate within. By utilizing Dijkstra’s algorithm, it is possible to calculate the Time to Live (TTL) between two given nodes in the network, which is an important metric as it directly affects how far apart two users can be within the region. When placing the nodes, a range for the nodes being used is specified and accounted for, which impacts the number of nodes needed within the region. Results show that when nodes are placed at coordinate points given by the generated map, users are able to communicate effectively across the specified region. In this project, a web interface is created in order to allow a user to specify the TTL, range, and the number of nodes to use, and proceeds to place each device within the region drawn by the user.


Abdul Baseer Mohammed

Enhancing Parameter-Efficient Fine-Tuning of Large Language Models with Alignment Adapters and LoRA

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Hongyang Sun, Chair
David Johnson
Prasad Kulkarni


Abstract

Large Language Models (LLMs) have become integral to natural language processing, involving initial broad pretraining on generic data followed by fine-tuning for specific tasks or domains. While advancements in Parameter Efficient Fine-Tuning (PEFT) techniques have made strides in reducing resource demands for LLM fine-tuning, they possess individual constraints. This project addresses the challenges posed by PEFT in the context of transformers architecture for sequence-to-sequence tasks, by integrating two pivotal techniques: Low-Rank Adaptation (LoRA) for computational efficiency and adaptive layers for task-specific customization. To overcome the limitations of LoRA, we introduce a simple yet effective hyper alignment adapter, that leverages a hypernetwork to generate decoder inputs based on encoder outputs, thereby serving as a crucial bridge to improve alignment between the encoder and the decoder. This fusion strikes a balance between the fine-tuning complexity and task performance, mitigating the individual drawbacks while improving the encoder-decoder alignment. As a result, we achieve more precise and contextually relevant sequence generation. The proposed solution improves the overall efficiency and effectiveness of LLMs in sequence-to-sequence tasks, leading to better alignment and more accurate output generation.


Laurynas Lialys

Engineering Laser Beams for Particle Trapping, Lattice Formation and Microscopy

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Shima Fardad, Chair
Morteza Hashemi
Rongqing Hui
Alessandro Salandrino
Xinmai Yang

Abstract

Having control over nano- and micro-sized objects' position inside a suspension is crucial in many applications such as: trapping and manipulating microscopic objects, sorting particles and living microorganisms, and building microscopic size 3D crystal structures and lattices. This control can be achieved by judiciously engineering optical forces and light-matter interactions inside colloidal suspensions that result in optical trapping. However, in the current techniques, to confine and transport particles in 3D, the use of high NA (Numerical Aperture) optics is a must. This in turn leads to several disadvantages such as alignment complications, narrow field of view, low stability values, and undesirable thermal effects. Hence, here we study a novel optical trapping method that we named asymmetric counter-propagating beams where optical forces are engineered to overcome the aforementioned limitations of existing methods. This novel system is significantly easier to align due to its utilization of much lower NA optics in combination with engineered beams which create a very flexible manipulating system. This new approach allows the trapping and manipulation of different shape objects, sizing from tens of nanometers to hundreds of micrometers by exploiting asymmetrical optical fields with high stability. In addition, this technique also allows for significantly larger particle trapping volumes. As a result, we can apply this method to trapping much larger particles and microorganisms that have never been trapped optically before as well as building 3D lattices and crystal structures of microscopic-sized particles. Finally, this novel approach allows for the integration of a variety of spectroscopy and microscopy techniques, such as light-sheet fluorescence microscopy, to extract time-sensitive information and acquire images with detailed features from trapped entities.


Past Defense Notices

Dates

Yoganand Pitta

Insightful Visualization: An Interactive Dashboard Uncovering Disease Patterns in Patient Healthcare Data

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Zijun Yao, Chair
Prasad Kulkarni
Hongyang Sun


Abstract

As Electronic Health Records (EHRs) become more available, there is increasing interest in discovering hidden disease patterns by leveraging cutting-edge data visualization techniques, such as graph-based knowledge representation and interactive graphical user interfaces (GUIs). In this project, we have developed a web-based interactive EHR analytics and visualization tool to provide healthcare professionals with valuable insights that can ultimately improve the quality and cost-efficiency of patient care. Specifically, we have developed two visualization panels: one for the intelligence of individual patients and the other for the relevance among diseases. For individual patients, we capture the similarity between them by linking them based on their relatedness in diagnosis. By constructing a graph representation of patients based on this similarity, we can identify patterns and trends in patient data that may not be apparent through traditional methods. For disease relationships, we provide an ontology graph for the specific diagnosis (ICD10 code), which helps to identify ancestors and predecessors of a particular diagnosis. Through the demonstration of this dashboard, we show that this approach can provide valuable insights to better understand patient outcomes with an informative and user-friendly web interface.

 


Brandon Ravenscroft

Spectral Cohabitation and Interference Mitigation via Physical Radar Emissions

When & Where:


Nichols Hall, Room 129, Ron Evans Apollo Auditorium

Committee Members:

Shannon Blunt, Chair
Chris Allen
Erik Perrins
James Stiles
Chris Depcik

Abstract

Auctioning of frequency bands to support growing demand for high bandwidth 5G communications is driving research into spectral cohabitation strategies for next generation radar systems. The loss of radio frequency (RF) spectrum once designated for radar operation is forcing radar systems to either learn how to coexist in these frequency spectrum bands, without causing mutual interference, or move to other bands of the spectrum, the latter being the more undesirable choice. Two methods of spectral cohabitation are presented in this work, each taking advantage of recent developments in non-repeating, random FM (RFM) waveforms. RFM waveforms are designed via one of many different optimization procedures to have favorable radar waveform properties while also readily incorporating agile spectrum notches. The first method of spectral cohabitation uses these spectral notches to avoid narrow-band RF interference (RFI) in the form of other spectrum users residing in the same band as the radar system, allowing both to operate while minimizing mutual interference. The second method of spectral cohabitation uses spectral notches, along with an optimization procedure, to embed a communications signal into a dual-function radar/communications (DFRC) emission, allowing one waveform to serve both functions simultaneously. Results of simulation and open-air experimentation with physically realized, spectrally notched and DFRC emissions are shown which demonstrate the efficacy of these two methods of spectral cohabitation.


Divya Harshitha Challa

Crop Prediction Based on Soil Classification using Machine Learning with Classifier Ensembling

When & Where:


Zoom Meeting, please email jgrisafe@ku.edu for defense link.

Committee Members:

Prasad Kulkarni, Chair
David Johnson
Hongyang Sun


Abstract

Globally, agriculture is the most significant source, which is the backbone of any country, and is an emerging field of research these days. There are many different types of soil, and each type has different characteristics for crops. Different methods and models are used daily in this region to increase yields. The macronutrient and micronutrient content of the soil, which is also a parametric representation of various climatic conditions like rain, humidity, temperature, and the soil's pH, is largely responsible for the crop's growth. Consequently, farmers are unable to select the appropriate crops depending on environmental and soil factors. The method of manually predicting the selection of the appropriate crops on land has frequently failed. We use machine learning techniques in this system to recommend crops based on soil classification or soil series. A comparative analysis of several popular classification algorithms, including K-Nearest Neighbors (KNN), Random Forest (RF), Decision Tree (DT), Support Vector Machines (SVM), Gaussian Naive Bayes (GNB), Gradient Boosting (GB), Extreme Gradient Boosting (XGBoost), and Voting Ensemble classifiers, is carried out in this work to assist in recommending the cultivable crop(s) that are most suitable for a particular piece of land depending on the characteristics of the soil and environment. To achieve our goal, we collected and preprocessed a large dataset of crop yield and environmental data from multiple sources. Our results show that the voting ensemble classifier outperforms the other classifiers in terms of prediction accuracy, achieving an accuracy of 94.67%. Feature importance analysis reveals that weather conditions such as temperature and rainfall, and fertilizer usage are the most critical factors in predicting crop yield. 


Oluwanisola Ibikunle

DEEP LEARNING ALGORITHMS FOR RADAR ECHOGRAM LAYER TRACKING

When & Where:


Nichols Hall, Room 317 (Richard K. Moore Conference Room)

Committee Members:

Shannon Blunt, Chair
John Paden (Co-Chair)
Carl Leuschen
Jilu Li
James Stiles

Abstract

The accelerated melting of ice sheets in the polar regions of the world, specifically in Greenland and Antarctica, due to contemporary climate warming is contributing to global sea level rise. To understand and quantify this phenomenon, airborne radars have been deployed to create echogram images that map snow accumulation patterns in these regions. Using advanced radar systems developed by the Center for Remote Sensing and Integrated Systems (CReSIS), a significant amount (1.5 petabytes) of climate data has been collected. However, the process of extracting ice phenomenology information, such as accumulation rate, from the data is limited. This is because the radar echograms require tracking of the internal layers, a task that is still largely manual and time-consuming. Therefore, there is a need for automated tracking.

Machine learning and deep learning algorithms are well-suited for this problem given their near-human performance on optical images. Moreover, the significant overlap between classical radar signal processing and machine learning techniques suggests that fusion of concepts from both fields can lead to optimized solutions for the problem. However, supervised deep learning algorithms suffer the circular problem of first requiring large amounts of labeled data to train the models which do not exist currently.

In this work, we propose custom algorithms, including supervised, semi-supervised, and self-supervised approaches, to deal with the limited annotated data problem to achieve accurate tracking of radiostratigraphic layers in echograms. Firstly, we propose an iterative multi-class classification algorithm, called “Row Block,” which sequentially tracks internal layers from the top to the bottom of an echogram given the surface location. We aim to use the trained iterative model in an active learning paradigm to progressively increase the labeled dataset. We also investigate various deep learning semantic segmentation algorithms by casting the echogram layer tracking problem as a binary and multiclass classification problem. These require post-processing to create the desired vector-layer annotations, hence, we propose a custom connected-component algorithm as a post-processing routine. Additionally, we propose end-to-end algorithms that avoid the post-processing to directly create annotations as vectors. Furthermore, we propose semi-supervised algorithms using weakly-labeled annotations and unsupervised algorithms that can learn the latent distribution of echogram snow layers while reconstructing echogram images from a sparse embedding representation.

A concurrent objective of this work is to provide the deep learning and science community with a large fully-annotated dataset. To achieve this, we propose synchronizing radar data with outputs from a regional climate model to provide a dataset with overlapping measurements that can enhance the performance of the trained models.


Jonathan Rogers

Faster than Thought Error Detection Using Machine Learning to Detect Errors in Brain Computer Interfaces

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Suzanne Shontz, Chair
Adam Rouse
Cuncong Zhong


Abstract

This research thesis seeks to use machine learning on data from invasive brain-computer interfaces (BCIs) in rhesus macaques to predict their state of movement during center-out tasks. Our research team breaks down movements into discrete states and analyzes the data using Linear Discriminant Analysis (LDA). We find that a simplified model that ignores the biological systems unpinning it can still detect the discrete state changes with a high degree of accuracy. Furthermore, when we account for underlying systems, our model achieved high levels of accuracy at speeds that ought to be imperceptible to the primate brain.


Abigail Davidow

Exploring the Gap Between Privacy and Utility in Automated Decision-Making

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Drew Davidson, Chair
Fengjun Li
Alexandra Kondyli


Abstract

The rapid rise of automated decision-making systems has left a gap in researchers’ understanding of how developers and consumers balance concerns about the privacy and accuracy of such systems against their utility.  With our goal to cover a broad spectrum of concerns from various angles, we initiated two experiments on the perceived benefit and detriment of interacting with automated decision-making systems. We refer to these two experiments as the Patch Wave study and Automated Driving study. This work approaches the study of automated decision making at different perspectives to help address the gap in empirical data on consumer and developer concerns. In our Patch Wave study, we focus on developers’ interactions with automated pull requests that patch widespread vulnerabilities on GitHub. The Automated Driving study explores older adults’ perceptions of data privacy in highly automated vehicles. We find quantitative and qualitative differences in the way that our target populations view automated decision-making systems compared to human decision-making. In this work, we detail our methodology for these studies, experimental results, and recommendations for addressing consumer and developer concerns.


Bhuneshwari Sharma Joshi

Applying ML Models for the Analysis of Bankruptcy Prediction

When & Where:


Zoom Meeting, please email jgrisafe@ku.edu for defense link.

Committee Members:

Prasad Kulkarni, Chair
Drew Davidson
David Johnson


Abstract

Bankruptcy prediction helps to evaluate the financial condition of a company and it helps not only the policymakers but the investors and all concerned people so they can take all required steps to avoid or to reduce the after-effects of bankruptcy. Bankruptcy prediction will not only help in making the best decision but also provides insight to reduce losses. The major reasons for the business organization’s failure are due to economic conditions such as proper allocation of resources, Input to policymakers, appropriate steps for business managers, identification of sector-wide problems, too much debt, insufficient capital, signal to Investors, etc. These factors can lead to making business unsustainable. The failure rate of businesses has tended to fluctuate with the state of the economy. The area of corporate bankruptcy prediction attains high economic importance, as it affects many stakeholders. The prediction of corporate bankruptcy has been extensively studied in economics, accounting, banking, and decision sciences over the past two decades. Many traditional approaches were suggested based on hypothesis testing and statistical analysis. Therefore, our focus and research are to come up with an approach that can estimate the probability of corporate bankruptcy and by evaluating its occurrence of failure using different machine learning models such as random forest, Random forest, XGboost, logistic method and choosing the one which gives highest accuracy. The dataset used was not well prepared and contained missing values, various data mining and data pre-processing techniques were utilized for data preparation. We use models such asRandom forest, Logistic method, random forest, XGBoost to predict corporate bankruptcy earlier to the occurrence. The accuracy results for accurate predictions of whether an organization will go bankrupt within the next 30, 90, or 180 days, using financial ratios as input features. The XGBoost-based model performs exceptionally well, with 98-99% accuracy.


Laurynas Lialys

Engineering laser beams for particle trapping, lattice formation and microscopy

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Committee Members:

Shima Fardad, Chair
Morteza Hashemi
Rongqing Hui
Alessandro Salandrino
Xinmai Yang

Abstract

Having control over nano- and micro-sized objects' position inside a suspension is crucial in many applications such as: sorting and delivery of particles, studying cells and microorganisms, spectroscopy imaging techniques, and building microscopic size lattices and artificial structures. This control can be achieved by judiciously engineering optical forces and light-matter interactions inside colloidal suspensions that result in optical trapping. However, in the current techniques, to confine and transport particles in 3D, the use of high-NA (Numerical Aperture) optics is a must. This in turn leads to several disadvantages such as alignment complications, lower trap stability, and undesirable thermal effects. Hence, here we study novel optical trapping methods such as asymmetric counter-propagating beams where we have engineered the optical forces to overcome the aforementioned limitations. This system is significantly easier to align as it uses much lower NA optics which creates a very flexible manipulating system. This new approach allows the trapping and transportation of different shape objects, sizing from hundreds of nanometers to hundreds of micrometers by exploiting asymmetrical optical fields with higher stability. In addition, this technique also allows for significantly longer particle trapping lengths of up to a few millimeters. As a result, we can apply this method to trapping much larger particles and microorganisms that have never been trapped optically before. Another application that the larger trapping lengths of the proposed system allow for is the creation of 3D lattices of microscopic size particles and other artificial structures, which is one important application of optical trapping.  

This system can be used to create a fully reconfigurable medium by optically controlling the position of selected nano- and micro-sized dielectric and metallic particles to mimic a certain medium. This “table-top” emulation can significantly simplify our studies of wave-propagation phenomena on transmitted signals in the real world. 

Furthermore, an important application of an optical tweezer system is that it can be combined with a variety of spectroscopy and microscopy techniques to extract valuable, time-sensitive information from trapped entities. In this research, I plan to integrate several spectroscopy techniques into the proposed trapping method in order to achieve higher-resolution images, especially for biomaterials such as microorganisms.  


Michael Cooley

Machine Learning for Navel Discharge Review

When & Where:


Eaton Hall, Room 1

Committee Members:

Prasad Kulkarni, Chair
David Johnson (Co-Chair)
Jerzy Grzymala-Busse


Abstract

This research project aims to predict the outcome of the Naval Discharge Review Board decision for an applicant based on factors in the application, using Machine Learning techniques. The study explores three popular machine learning algorithms: MLP, Adaboost, and KNN, with KNN providing the best results. The training is verified through hyperparameter optimization and cross fold validation.

Additionally, the study investigates the ability of ChatGPT's API to classify the data that couldn't be classified manually. A total of over 8000 samples were classified by ChatGPT's API, and an MLP model was trained using the same hyperparameters that were found to be optimal for the 3000 size manual sample.The model was then tested on the manual sample. The results show that the model trained on data labeled by ChatGPT performed equivalently, suggesting that ChatGPT's API is a promising tool for labeling in this domain.


Vasudha Yenuganti

RNA Structure Annotation Based on Base Pairs Using ML Based Classifiers

When & Where:


Eaton Hall, Room 2001B

Committee Members:

Cuncong Zhong, Chair
David Johnson
Prasad Kulkarni


Abstract

RNA molecules play a crucial role in the regulation of gene expression and other cellular processes. Understanding the three-dimensional structure of RNA is essential for predicting its function and interactions with other molecules. One key feature of RNA structure is the presence of base pairs, where nucleotides i.e., adenine(A), guanine(G), cytosine(C), and uracil(U), form hydrogen bonds with each other. The limited availability of high-quality RNA structural data combined with associated atomic coordinate errors in low resolution structures, presents significant challenges for extracting important geometrical characteristics from RNA's complex three-dimensional structure, particularly in terms of base interactions.

In this study, we propose an approach for annotating base-pairing interactions in low-resolution RNA structures using machine learning (ML) based classifiers and leveraging the more precise structural information available in high-resolution homologs to annotate base-pairing interactions in low-resolution structures. We first use DSSR tool to extract annotations of high-resolution RNA structures and extract distances of atoms of interacting base pairs. The distances serve as features, and 12 standard annotations are used as labels for our ML model. We then apply different ML classifiers, including support vector machines, neural networks, and random forests, to predict RNA annotations. We evaluate the performance of these classifiers using a benchmark dataset and report their precision, recall, and F1-score. Low-resolution RNA structures are then annotated based on the sequence-similarity with high-resolution structures and the corresponding predicted annotations.

For future aspects, the presented approach can also help to explore the plausible base pair interactions to identify conserved motifs in low-resolution structures. The detected interactions along with annotations can aid in the study of RNA tertiary structures, which can lead to a better understanding of their functions in the cell.