The mDOT Center

Transforming health and wellness via temporally-precise mHealth interventions
mDOT@MD2K.org
901.678.1526
 

TR&D3: Translation

mDOT Center > Research Projects > TR&D3: Translation

Translation of Temporally Precise mHealth via Efficient and Embeddable Privacy-aware Biomarker Implementations

Tinyodom: Hardware-Aware Efficient Neural Inertial Navigation
Authors:
Publication Venue:

ACM on Interactive, Mobile, Wearable, and Ubiquitous Technologies

Publication Date:

July 7, 2022

Keywords:
robotics, embedded systems, inertial odometry, dead-reckoning, sequence-learning, resource-constrained devices, neural architecture search, hardware-in-the-loop, machine-learning, deep-learning, tracking
Abstract:
Deep inertial sequence learning has shown promising odometric resolution over model-based approaches for trajectory estimation in GPS-denied environments. However, existing neural inertial dead-reckoning frameworks are not suitable for real-time deployment on ultra-resource-constrained (URC) devices due to substantial memory, power, and compute bounds. Current deep inertial odometry techniques also suffer from gravity pollution, high-frequency inertial disturbances, varying sensor orientation, heading rate singularity, and failure in altitude estimation. In this paper, we introduce TinyOdom, a framework for training and deploying neural inertial models on URC hardware. TinyOdom exploits hardware and quantization-aware Bayesian neural architecture search (NAS) and a temporal convolutional network (TCN) backbone to train lightweight models targetted towards URC devices. In addition, we propose a magnetometer, physics, and velocity-centric sequence learning formulation robust to preceding inertial perturbations. We also expand 2D sequence learning to 3D using a model-free barometric g-h filter robust to inertial and environmental variations. We evaluate TinyOdom for a wide spectrum of inertial odometry applications and target hardware against competing methods. Specifically, we consider four applications: pedestrian, animal, aerial, and underwater vehicle dead-reckoning. Across different applications, TinyOdom reduces the size of neural inertial models by 31× to 134× with 2.5m to 12m error in 60 seconds, enabling the direct deployment of models on URC devices while still maintaining or exceeding the localization resolution over the state-of-the-art. The proposed barometric filter tracks altitude within ±0.1m and is robust to inertial disturbances and ambient dynamics. Finally, our ablation study shows that the introduced magnetometer, physics, and velocity-centric sequence learning formulation significantly improve localization performance even with notably lightweight models.
TL;DR:

In this paper, we introduce TinyOdom, a framework for training and deploying neural inertial models on URC hardware.

Auritus: An Open-Source Optimization Toolkit for Training & Development of Human Movement Models & Filters Using Earables
Authors:
Publication Venue:

ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies

Publication Date:

July 7, 2022

Keywords:

earable, network architecture search, neural networks, machine learning, datasets, filters, human activity, head-pose, TinyML, optimization, hardware-in-the-loop

Abstract:
Smart ear-worn devices (called earables) are being equipped with various onboard sensors and algorithms, transforming earphones from simple audio transducers to multi-modal interfaces making rich inferences about human motion and vital signals. However, developing sensory applications using earables is currently quite cumbersome with several barriers in the way. First, time-series data from earable sensors incorporate information about physical phenomena in complex settings, requiring machine-learning (ML) models learned from large-scale labeled data. This is challenging in the context of earables because large-scale open-source datasets are missing. Secondly, the small size and compute constraints of earable devices make on-device integration of many existing algorithms for tasks such as human activity and head-pose estimation difficult. To address these challenges, we introduce Auritus, an extendable and open-source optimization toolkit designed to enhance and replicate earable applications. Auritus serves two primary functions. Firstly, Auritus handles data collection, pre-processing, and labeling tasks for creating customized earable datasets using graphical tools. The system includes an open-source dataset with 2.43 million inertial samples related to head and full-body movements, consisting of 34 head poses and 9 activities from 45 volunteers. Secondly, Auritus provides a tightly-integrated hardware-in-the-loop (HIL) optimizer and TinyML interface to develop lightweight and real-time machine-learning (ML) more »
TL;DR:
We introduce Auritus, an extendable and open-source optimization toolkit designed to enhance and replicate earable applications.
THIN-Bayes: Platform-Aware Machine Learning for Low-End IoT Devices
Authors:
Publication Venue:

tinyML Summit

Publication Date:

March 2022

Keywords:

neural networks, edge computing, IoT platforms, AI-based inference, TinyML, activity detection models

Abstract:
Neural networks have been shown to provide rich and complicated inferences from time-series data over first-principle approaches. However, with inference moving to the edge, and IoT platforms shrinking, realizing AI-based inference on-board is challenging. While communication bandwidth, energy budget, and form factor of these platforms have gone down, the workload and complexity of neural networks have skyrocketed, requiring systematic software and tools to guide on-board TinyML implementation. Keeping the challenges in mind, we developed THIN-Bayes, a completely open-source black-box optimization framework for training and deploying ultra-lightweight models on extremely resource-constrained platforms using TensorFlow Lite Micro. THIN-Bayes is designed based on three insights: 1. hardware proxies deviate significantly from true values for low-end microcontrollers, thereby requiring exact hardware metrics; 2. the linear programming formulation stemming from hardware-in-the-loop neural architecture search is not gradient-friendly. 3. the type of model should also be a hyperparameter for low-end microcontrollers. Based on ARM Mango, THIN-Bayes features a tightly integrated ultra-lightweight model zoo and a gradient-free, hardware-in-the-loop, and parallelizable Bayesian neural architecture search framework, receiving hardware metrics directly from the target hardware during optimization. We demonstrate the efficacy of our framework by optimizing neural-inertial navigation models and earable human activity detection models for microcontrollers, two of the most challenging applications of inertial sensors. The inertial-odometry models found by THIN-Bayes are 31-134x smaller than state-of-the-art neural-inertial navigation models, while the activity detection models are 98x smaller and 6% more accurate over the state-of-the-art. THIN-Bayes is an important step towards bringing in challenging AI applications to TinyML platforms.
TL;DR:
We developed THIN-Bayes, a completely open-source black-box optimization framework for training and deploying ultra-lightweight models on extremely resource-constrained platforms using TensorFlow Lite Micro.
CardiacGen: A Hierarchical Deep Generative Model for Cardiac Signals
Authors:
Publication Venue:

Machine Learning for Health (ML4H)

Publication Date:

November 15, 2022

Keywords:

generative model, electrocardiogram, data augmentation

Abstract:

We present CardiacGen, a Deep Learning framework for generating synthetic but physiologically plausible cardiac signals like ECG. Based on the physiology of cardiovascular system function, we propose a modular hierarchical generative model and impose explicit regularizing constraints for training each module using multi-objective loss functions. The model comprises 2 modules, an HRV module focused on producing realistic Heart-Rate-Variability characteristics and a Morphology module focused on generating realistic signal morphologies for different modalities. We empirically show that in addition to having realistic physiological features, the synthetic data from CardiacGen can be used for data augmentation to improve the performance of Deep Learning based classifiers.

TL;DR:

We present CardiacGen, a Deep Learning framework for generating synthetic but physiologically plausible cardiac signals like ECG.

Robust Finger Interactions with COTS Smartwatches via Unsupervised Siamese Adaptation
Authors:
Publication Venue:

ACM Symposium on User
Interface Software and Technology (2023)

Publication Date:

October 29, 2023

Keywords:

gesture recognition, finger interaction, vibration sensing, unsupervised adversarial training

Abstract:

Wearable devices like smartwatches and smart wristbands have gained substantial popularity in recent years. However, their small interfaces create inconvenience and limit computing functionality. To fill this gap, we propose ViWatch, which enables robust finger interactions under deployment variations, and relies on a single IMU sensor that is ubiquitous in COTS smartwatches. To this end, we design an unsupervised Siamese adversarial learning method. We built a real-time system on commodity smartwatches and tested it with over one hundred volunteers. Results show that the system accuracy is about 97% over a week. In addition, it is resistant to deployment variations such as different hand shapes, finger activity strengths, and smartwatch positions on the wrist. We also developed a number of mobile applications using our interactive system and conducted a user study where all participants preferred our unsupervised approach to supervised calibration. The demonstration of ViWatch is shown at https://youtu.be/N5-ggvy2qfI.

TL;DR:
We propose ViWatch, which enables robust finger interactions under deployment variations, and relies on a single IMU sensor that is ubiquitous in COTS smartwatches. To this end, we design an unsupervised Siamese adversarial learning method.
TinyNS: Platform-Aware Neurosymbolic Auto Tiny Machine Learning
Authors:
Publication Venue:

ACM Transactions on Embedded Computing Systems (2023)

Publication Date:

May 31, 2023

Keywords:

neurosymbolic, neural architecture search, TinyML, AutoML, bayesian, platform-aware

Abstract:

Machine learning at the extreme edge has enabled a plethora of intelligent, time-critical, and remote applications. However, deploying interpretable artificial intelligence systems that can perform high-level symbolic reasoning and satisfy the underlying system rules and physics within the tight platform resource constraints is challenging. In this paper, we introduce TinyNS, the first platform-aware neurosymbolic architecture search framework for joint optimization of symbolic and neural operators. TinyNS provides recipes and parsers to automatically write microcontroller code for five types of neurosymbolic models, combining the context awareness and integrity of symbolic techniques with the robustness and performance of machine learning models. TinyNS uses a fast, gradient-free, black-box Bayesian optimizer over discontinuous, conditional, numeric, and categorical search spaces to find the best synergy of symbolic code and neural networks within the hardware resource budget. To guarantee deployability, TinyNS talks to the target hardware during the optimization process. We showcase the utility of TinyNS by deploying microcontroller-class neurosymbolic models through several case studies. In all use cases, TinyNS outperforms purely neural or purely symbolic approaches while guaranteeing execution on real hardware.

TL;DR:

We introduce TinyNS, the first platform-aware neurosymbolic architecture search framework for joint optimization of symbolic and neural operators.

Automated Detection of Stressful Conversations Using Wearable Physiological & Inertial Sensors
Authors:
Publication Venue:

Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies

Keywords:

stressful conversations, stress detection, wearables, physiological sensors, intertial sensors, 

Publication Date:

December 2020

Abstract:

Stressful conversation is a frequently occurring stressor in our daily life. Stressors not only adversely affect our physical and mental health but also our relationships with family, friends, and coworkers. In this paper, we present a model to automatically detect stressful conversations using wearable physiological and inertial sensors. We conducted a lab and a field study with cohabiting couples to collect ecologically valid sensor data with temporally-precise labels of stressors. We introduce the concept of stress cycles, i.e., the physiological arousal and recovery, within a stress event. We identify several novel features from stress cycles and show that they exhibit distinguishing patterns during stressful conversations when compared to physiological response due to other stressors. We observe that hand gestures also show a distinct pattern when stress occurs due to stressful conversations. We train and test our model using field data collected from 38 participants. Our model can determine whether a detected stress event is due to a stressful conversation with an F1-score of 0.83, using features obtained from only one stress cycle, facilitating intervention delivery within 3.9 minutes since the start of a stressful conversation.

TL;DR:

In this paper, we present a model to automatically detect stressful conversations using wearable physiological and intertial sensors.

mTeeth: Identifying Brushing Teeth Surfaces Using Wrist-Worn Inertial Sensors
Authors:
Publication Venue:

Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies

Keywords:

mHealth, brushing detection, flossing detection, hand-to-mouth gestures

Publication Date:

June 2021

Related Project:
Abstract:

Ensuring that all the teeth surfaces are adequately covered during daily brushing can reduce the risk of several oral diseases. In this paper, we propose the mTeeth model to detect teeth surfaces being brushed with a manual toothbrush in the natural free-living environment using wrist-worn inertial sensors. To unambiguously label sensor data corresponding to different surfaces and capture all transitions that last only milliseconds, we present a lightweight method to detect the micro-event of brushing strokes that cleanly demarcates transitions among brushing surfaces. Using features extracted from brushing strokes, we propose a Bayesian Ensemble method that leverages the natural hierarchy among teeth surfaces and patterns of transition among them. For training and testing, we enrich a publicly-available wrist-worn inertial sensor dataset collected from the natural environment with time-synchronized precise labels of brushing surface timings and moments of transition. We annotate 10,230 instances of brushing on different surfaces from 114 episodes and evaluate the impact of wide between-person and within-person between-episode variability on machine learning model’s performance for brushing surface detection.

TL;DR:

In this paper, we propose the mTeeth model to detect teeth surfaces being brushed with a manual toothbrush in the natural free-living environment using wrist-worn inertial sensors.

MOODS: Mobile Open Observation of Daily Stressors
Authors:
Publication Venue:

Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24)

Publication Date:

May 11, 2024

Keywords:

stress-tracking, stressor-logging, visualizations, behavioral changes, stress intervention, emotion/affective computing, wearable sensors, field studies, empirical studies in HCI, longitudinal study, self-reflection, personal informatics, stress reduction, self-awareness, smartwatch apps, smartphone apps, physiological events, and daily stressors.

Related Projects:

CP7

Abstract:

Commercial wearables from Fitbit, Garmin, and Whoop have recently introduced real-time notifications based on detecting changes in physiological responses indicating potential stress. In this paper, we investigate how these new capabilities can be leveraged to improve stress management. We developed a smartwatch app, a smartphone app, and a cloud service, and conducted a 100-day field study with 122 participants who received prompts triggered by physiological responses several times a day. They were asked whether they were stressed, and if so, to log the most likely stressor. Each week, participants received new visualizations of their data to self-reflect on patterns and trends. Participants reported better awareness of their stressors, and self-initiating fourteen kinds of behavioral changes to reduce stress in their daily lives. Repeated self-reports over 14 weeks showed reductions in both stress intensity (in 26,521 momentary ratings) and stress frequency (in 1,057 weekly surveys).

TL;DR:

A 100-day field study (MOODS) explored how wearables can enhance stress management by combining real-time physiological stress detection with momentary stressor logging and weekly self-reflective visualizations. The study found significant reductions in self-reported stress intensity and frequency, fostering greater self-awareness of stressors and prompting participants to make 14 types of self-initiated behavioral changes for improved emotional regulation, productivity, and self-care. This work highlights the potential of personal informatics systems in driving lasting stress reduction through user engagement and data insights.

Artificial Intelligence for End Tidal Capnography Guided Resuscitation: A Conceptual Framework
Authors:

Nassal, M. Sugavanam, N., Aramendi, E., Jaureguibeitia, X., Elola, A., Panchal, A., Ulintz, A., Wang, H., Ertin, E.

Publication Venue:

Proceedings of the 57th Annual Hawaii International Conference on System Sciences, HICSS 2024

Publication Date:

January 3, 2024

Keywords:

Artificial Intelligence (AI), cardiac arrest, resuscitation, end tidal capnography, reinforcement learning

Related Project:
Abstract:

Artificial Intelligence (AI) and machine learning have advanced healthcare by defining relationships in complex conditions. Out-of-hospital cardiac arrest (OHCA) is a medically complex condition with several etiologies. Survival for OHCA has remained static at 10% for decades in the United States. Treatment of OHCA requires the coordination of numerous interventions, including the delivery of multiple medications. Current resuscitation algorithms follow a single strict pathway, regardless of fluctuating cardiac physiology. OHCA resuscitation requires a real-time biomarker that can guide interventions to improve outcomes. End tidal capnography (ETCO2) is commonly implemented by emergency medical services professionals in resuscitation and can serve as an ideal biomarker for resuscitation. However, there are no effective conceptual frameworks utilizing the continuous ETCO2 data. In this manuscript, we detail a conceptual framework using AI and machine learning techniques to leverage ETCO2 in guided resuscitation.

TL;DR:

This publication proposes a conceptual framework for utilizing Artificial Intelligence (AI) and machine learning to create End Tidal Capnography (ETCO2) guided resuscitation for Out-of-Hospital Cardiac Arrest (OHCA). The aim is to move beyond rigid, fixed-interval resuscitation algorithms by leveraging continuous ETCO2 data as a real-time biomarker, alongside other physiological measurements, to develop personalized, dynamic interventions that are responsive to a patient’s evolving cardiac physiology. This approach seeks to improve the currently static survival rates for OHCA by enabling a deeper analysis of ETCO2 trends in relation to patient characteristics and interventions, potentially revealing “hidden” patterns and allowing for reward-based algorithms to guide optimal treatment strategies.

Detecting Context Shifts in the Human Experience Using Multimodal Foundation Models
Authors:

Iris Nguyen, Liying Han, Burke Dambly, Alireza Kazemi, Marina Kogan, Cory Inman, Mani Srivastava, Luis Garcia

Publication Venue:

Proceedings of the 23rd ACM Conference on Embedded Networked Sensor Systems

Publication Date:

May 6, 2025

Keywords:

multimodal foundation models, embodied artificial intelligence

Related Projects:
Abstract:
Detecting context shifts in human experience is critical for applications in cognitive modeling, human-AI interaction, and adaptive neurotechnology. However, formalizing and identifying these shifts in real-world settings remains challenging due to annotation inconsistencies, data sparsity, and the multimodal nature of human perception.
 
Detecting context shifts in human experience is critical for applications in cognitive modeling, human-AI interaction, and adaptive neurotechnology. However, formalizing and identifying these shifts in real-world settings remains challenging due to annotation inconsistencies, data sparsity, and the multimodal nature of human perception. In this poster, we explore the use of multimodal foundation models for detecting context shifts by leveraging neural, wearable, and environmental sensors. Initial findings from a neuroscience-driven annotation study highlight discrepancies in human-labeled transitions, emphasizing the need for a model-driven approach. Given the limited availability of labeled datasets, we examine: 1) Surrogate models trained on synthetic datasets, 2) Sensor fusion techniques to align real-world neural and behavioral signals, and 3) The role of foundation models in interpreting multimodal context shifts. We outline key challenges in sensor data alignment, inter-rater variability, and transfer learning from synthetic to real-world data.
TL;DR:
This poster explores using multimodal foundation models to detect context shifts in human experience by leveraging neural, wearable, and environmental sensors. The research highlights challenges like annotation inconsistencies and data sparsity, proposing to investigate surrogate models, sensor fusion, and the role of foundation models to overcome these issues.
Mmbind: Unleashing the potential of distributed and heterogeneous data for multimodal learning in IoT
Authors:
Publication Venue:

Proceedings of the 23rd ACM Conference on Embedded Networked Sensor Systems

Publication Date:

May 6, 2025

Keywords:

AI Innovation, Adaptive Learningy, Machine Learning,
Multimodal Learning

Related Projects:
Abstract:
Multimodal sensing systems are increasingly prevalent in various real-world applications. Most existing multimodal learning approaches heavily rely on training with a large amount of synchronized, complete multimodal data. However, such a setting is impractical in real-world IoT sensing applications where data is typically collected by distributed nodes with heterogeneous data modalities, and is also rarely labeled. In this paper, we propose MMBind, a new data binding approach for multimodal learning on distributed and heterogeneous IoT data. The key idea of MMBind is to construct a pseudo-paired multimodal dataset for model training by binding data from disparate sources and incomplete modalities through a sufficiently descriptive shared modality. We also propose a weighted contrastive learning approach to handle domain shifts among disparate data, coupled with an adaptive multimodal learning architecture capable of training models with heterogeneous modality combinations. Evaluations on ten real-world multi-modal datasets highlight that MMBind outperforms state-of-the-art baselines under varying degrees of data incompleteness and domain shift, and holds promise for advancing multimodal foundation model training in IoT applications.
TL;DR:
This paper introduces MMBind, a novel framework designed to tackle the key challenges of real-world IoT multimodal learning, where data is distributed across different nodes, often lacks labels, and has missing or unsynchronized modalities. Instead of requiring perfectly aligned data, MMBind’s core innovation is to create a “pseudo-paired” training dataset by intelligently binding disparate data sources using a common, descriptive modality. It employs weighted contrastive learning to handle differences between data sources and a flexible architecture that works with various modality combinations. Evaluations show MMBind outperforms existing methods, making it a promising solution for building powerful multimodal models with messy, real-world IoT data.
Reimagining time series foundation models: Metadata and state-space model perspectives
Authors:

Pengrui Quan, Ozan Mulayim, Liying Han, Dezhi Hong, Mario Berges, Mani Srivastava

Publication Venue:
Publication Date:

December 15, 2024

Keywords:

Time Series Foundation Models, State-Space Models, Metadata, Forecasting, Transformer, Efficient AI.

Related Projects:
Abstract:

The success of foundation models in natural language processing has sparked a growing interest in developing analogous models for time series (TS) analysis. These time series foundation models (TSFM), pre-trained on vast amounts of TS data, demonstrate capabilities of zero-shot and few-shot inference on unseen datasets. However, the intrinsic heterogeneity of TS data presents unique challenges: accurate inference often necessitates a deep understanding of the underlying data-generating process and the sensing apparatus, which cannot be readily inferred from the raw data alone. Furthermore, recent advances in state-space models raise the question of whether they may offer advantages over transformer-based architectures for TS analysis.

This paper investigates these questions in two key areas: (a) a fair comparison of methods for integrating metadata into TSFMs and (b) the comparative effectiveness of state-space models (SSM) versus transformer models for TS forecasting. Our results, based on experiments across 11 datasets, suggest advantages for SSM building blocks as well as for incorporating the notion of real-world timestamps. More specifically, on our curated in-domain and out-of-domain datasets, an SSM approach incorporating timestamps outperforms three existing TSFMs on forecasting tasks while using 6,000 fewer trainable parameters and 10 less training data. The paper aims to highlight the potential for SSM building blocks and general directions for future TSFM research.

TL;DR:
This paper rethinks the architecture for Time Series Foundation Models (TSFMs) by proposing two key ideas: the effective integration of metadata (like timestamps) and the use of State-Space Models (SSMs) as a core building block instead of Transformers. Through extensive experiments, the authors demonstrate that an SSM-based model that incorporates real-world timing information significantly outperforms existing TSFMs on forecasting tasks, while being dramatically more efficient—achieving superior results with 6,000x fewer parameters and 10x less training data.
SensorBench: Benchmarking LLMs in coding-based sensor processing
Authors:

Pengrui Quan, Xiaomin Ouyang, Jeya Vikranth Jeyakumar, Ziqi Wang, Yang Xing, Mani Srivastava

Publication Venue:
Publication Date:

February 26, 2025

Keywords:

Large Language Models, Sensor Data Processing, Benchmarking, Cyber-Physical Systems, Prompting Strategies, Code Generation.

Related Projects:
Abstract:

Effective processing, interpretation, and management of sensor data have emerged as a critical component of cyber-physical systems. Traditionally, processing sensor data requires profound theoretical knowledge and proficiency in signal-processing tools. However, recent works show that Large Language Models (LLMs) have promising capabilities in processing sensory data, suggesting their potential as copilots for developing sensing systems.
To explore this potential, we construct a comprehensive benchmark, SensorBench, to establish a quantifiable objective. The benchmark incorporates diverse real-world sensor datasets for various tasks. The results show that while LLMs exhibit considerable proficiency in simpler tasks, they face inherent challenges in processing compositional tasks with parameter selections compared to engineering experts. Additionally, we investigate four prompting strategies for sensor processing and show that self-verification can outperform all other baselines in 48% of tasks. Our study provides a comprehensive benchmark and prompting analysis for future developments, paving the way toward an LLM-based sensor processing copilot.

TL;DR:
This paper introduces SensorBench, a comprehensive benchmark to evaluate the ability of Large Language Models (LLMs) to act as “copilots” for coding-based sensor data processing. The findings reveal that while LLMs are proficient at straightforward sensor tasks, they struggle with more complex, compositional tasks that require nuanced parameter selection, an area where human experts still excel. The study also evaluates prompting strategies, finding that a self-verification approach performs best in nearly half of the tasks, providing a foundational benchmark for future development of LLM-assisted sensing systems.
ProMind-LLM: Proactive mental health care via causal reasoning with sensor data
Authors:

Xinzhe Zheng, Sijie Ji, Jiawei Sun, Renqi Chen, Wei Gao, Mani Srivastava

Publication Venue:
Publication Date:
Keywords:

Proactive Mental Health Care, Large Language Models (LLMs),
Mental Health Risk Assessment,
Causal Reasoning, Objective Behavior Data Integration, mHealth

Related Projects:
Abstract:
Mental health risk is a critical global public health challenge, necessitating innovative and reliable assessment methods. With the development of large language models (LLMs), they stand out to be a promising tool for explainable mental health care applications. Nevertheless, existing approaches predominantly rely on subjective textual mental records, which can be distorted by inherent mental uncertainties, leading to inconsistent and unreliable predictions. To address these limitations, this paper introduces ProMind-LLM. We investigate an innovative approach integrating objective behavior data as complementary information alongside subjective mental records for robust mental health risk assessment. Specifically, ProMind-LLM incorporates a comprehensive pipeline that includes domain-specific pretraining to tailor the LLM for mental health contexts, a self-refine mechanism to optimize the processing of numerical behavioral data, and causal chain-of-thought reasoning to enhance the reliability and interpretability of its predictions. Evaluations of two real-world datasets, PMData and Globem, demonstrate the effectiveness of our proposed methods, achieving substantial improvements over general LLMs. We anticipate that ProMind-LLM will pave the way for more dependable, interpretable, and scalable mental health case solutions.
TL;DR:
This paper introduces ProMind-LLM, a framework designed to provide more reliable and proactive mental health risk assessments by moving beyond subjective patient reports. It integrates objective behavioral data from sensors with subjective records and uses a specialized pipeline that includes mental health domain pre-training, a self-refine mechanism for processing numerical data, and causal reasoning to make its predictions both accurate and interpretable. Evaluated on real-world datasets, ProMind-LLM significantly outperforms general-purpose LLMs, offering a more dependable and scalable path for AI in mental health care.
Benchmarking spatiotemporal reasoning in LLMs and reasoning models: Capabilities and challenges
Authors:

Pengrui Quan, Brian Wang, Kang Yang, Liying Han, Mani Srivastava

Publication Venue:

Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS); 2025

Publication Date:

May 27, 2025

Keywords:

Spatiotemporal Reasoning, Large Language Models, Benchmarking, Cyber-Physical Systems, Large Reasoning Models, Geometric Reasoning.

Related Projects:
Abstract:

Spatiotemporal reasoning plays a key role in Cyber-Physical Systems (CPS). Despite advances in Large Language Models (LLMs) and Large Reasoning Models (LRMs), their capacity to reason about complex spatiotemporal signals remains underexplored. This paper proposes a hierarchical SpatioTemporal reAsoning benchmaRK, STARK, to systematically evaluate LLMs across three levels of reasoning complexity: state estimation (e.g., predicting field variables, localizing and tracking events in space and time), spatiotemporal reasoning over states (e.g., inferring spatial-temporal relationships), and world-knowledge-aware reasoning that integrates contextual and domain knowledge (e.g., intent prediction, landmark-aware navigation). We curate 26 distinct spatiotemporal tasks with diverse sensor modalities, comprising 14,552 challenges where models answer directly or by Python Code Interpreter. Evaluating 3 LRMs and 8 LLMs, we find LLMs achieve limited success in tasks requiring geometric reasoning (e.g., multilateration or triangulation), particularly as complexity increases. Surprisingly, LRMs show robust performance across tasks with various levels of difficulty, often competing or surpassing traditional first-principle-based methods. Our results show that in reasoning tasks requiring world knowledge, the performance gap between LLMs and LRMs narrows, with some LLMs even surpassing LRMs. However, the LRM o3 model continues to achieve leading performance across all evaluated tasks, a result attributed primarily to the larger size of the reasoning models. STARK motivates future innovations in model architectures and reasoning paradigms for intelligent CPS by providing a structured framework to identify limitations in the spatiotemporal reasoning of LLMs and LRMs.

TL;DR:
This paper introduces STARK, a hierarchical benchmark to systematically evaluate the spatiotemporal reasoning abilities of Large Language Models (LLMs) and Large Reasoning Models (LRMs) across tasks critical for Cyber-Physical Systems, from basic state estimation to complex, knowledge-aware reasoning. The evaluation reveals that while LLMs struggle with geometric reasoning, LRMs demonstrate robust and often superior performance, with the o3 model leading across the board, a success attributed to its larger reasoning model size. The benchmark highlights a narrowing performance gap in knowledge-based tasks but clearly shows the current superiority of specialized reasoning models for complex spatiotemporal problems.
Bayesian Sparse Blind Deconvolution Using MCMC Methods Based on Normal-Inverse-Gamma Prior
Authors:
Publication Venue:

IEEE Transactions on Signal Processing

Publication Date:

March 3, 2022

Keywords:
Bayes methods, computational modeling, deconvolution, distortion, probabilistic logic, Monte Carlo methods, estimation
Abstract:

Bayesian estimation methods for sparse blind deconvolution problems conventionally employ Bernoulli-Gaussian (BG) prior for modeling sparse sequences and utilize Markov Chain Monte Carlo (MCMC) methods for the estimation of unknowns. However, the discrete nature of the BG model creates computational bottlenecks, preventing efficient exploration of the probability space even with the recently proposed enhanced sampler schemes. To address this issue, we propose an alternative MCMC method by modeling the sparse sequences using the Normal-Inverse-Gamma (NIG) prior. We derive effective Gibbs samplers for this prior and illustrate that the computational burden associated with the BG model can be eliminated by transferring the problem into a completely continuous-valued framework. In addition to sparsity, we also incorporate time and frequency domain constraints on the convolving sequences. We demonstrate the effectiveness of the proposed methods via extensive simulations and characterize computational gains relative to the existing methods that utilize BG modeling.

TL;DR:
We propose an alternative MCMC method by modeling the sparse sequences using the Normal-Inverse-Gamma (NIG) prior. We derive effective Gibbs samplers for this prior and illustrate that the computational burden associated with the BG model can be eliminated by transferring the problem into a completely continuous-valued framework.
Machine Learning for Microcontroller-Class Hardware--A Review
Authors:
Publication Venue:

IEEE Sensors Journal

Publication Date:

November 15, 2022

Keywords:

feature projection, Internet of Things, machine learning, microcontrollers, model compression, neural architecture search, neural networks, optimization, sensors, TinyML

Abstract:
The advancements in machine learning opened a new opportunity to bring intelligence to the low-end Internet-of-Things nodes such as microcontrollers. Conventional machine learning deployment has high memory and compute footprint hindering their direct deployment on ultra resource-constrained microcontrollers. This paper highlights the unique requirements of enabling onboard machine learning for microcontroller class devices. Researchers use a specialized model development workflow for resource-limited applications to ensure the compute and latency budget is within the device limits while still maintaining the desired performance. We characterize a closed-loop widely applicable workflow of machine learning model development for microcontroller class devices and show that several classes of applications adopt a specific instance of it. We present both qualitative and numerical insights into different stages of model development by showcasing several use cases. Finally, we identify the open research challenges and unsolved questions demanding careful considerations moving forward.
TL;DR:

This paper highlights the unique requirements of enabling onboard machine learning for microcontroller class devices. We characterize a closed-loopwidely applicable workflow of ML model development for microcontroller-class devices and show that several classes of applications adopt a specific instance of it.

Removing Antenna Effects Using an Invertible Neural Network for Improved Estimation of Multilayered Tissue Profiles Using UWB Radar
Authors:
Publication Venue:

2023 IEEE International Symposium on Antennas and Propagation and USNC-URSI Radio Science Meeting (AP-S/URSI). pp. 53–54.

Publication Date:

July 23-28, 2023

Keywords:

antenna measurements, three-dimensional displays, neural networks, transfer functions, radar antennas, nonhomogeneous media, ultra wideband antennas

Abstract:

Ultrawideband (UWB) radar sensors are an emerging biosensing modality that can be used to assess the dielectric properties of internal tissues. Antenna effects, including antenna body interactions limit the sensors ability to isolate the weak returns from the internal tissues. In this paper we develop a data driven calibration method for recovering Green’s function of the multilayered media model of the tissue profiles using an Invertible Neural Network (INN). The proposed INN structure is trained to invert the antenna transfer function to form estimates of the Green’s function modeling returns from internal tissues. We use simulation experiments to assess the effectiveness of the trained INN in antenna transfer function inversion.

TL;DR:

In this paper we develop a data driven calibration method for recovering Green’s function of the multilayered media model of the tissue profiles using an Invertible Neural Network (INN).

The Validity of MotionSense HRV in Estimating Sedentary Behavior and Physical Activity under Free-Living and Simulated Activity Settings
Authors:
Publication Venue:

Sensors (Basel)

Keywords:

MotionSense HRV, accelerometer, mobile health, physical activity, sedentary behavior

Publication Date:

February 18, 2021

Related Projects:
Abstract:

MotionSense HRV is a wrist-worn accelerometery-based sensor that is paired with a smartphone and is thus capable of measuring the intensity, duration, and frequency of physical activity (PA). However, little information is available on the validity of the MotionSense HRV. Therefore, the purpose of this study was to assess the concurrent validity of the MotionSense HRV in estimating sedentary behavior (SED) and PA. A total of 20 healthy adults (age: 32.5 ± 15.1 years) wore the MotionSense HRV and ActiGraph GT9X accelerometer (GT9X) on their non-dominant wrist for seven consecutive days during free-living conditions. Raw acceleration data from the devices were summarized into average time (min/day) spent in SED and moderate-to-vigorous PA (MVPA). Additionally, using the Cosemed K5 indirect calorimetry system (K5) as a criterion measure, the validity of the MotionSense HRV was examined in simulated free-living conditions. Pearson correlations, mean absolute percent errors (MAPE), Bland-Altman (BA) plots, and equivalence tests were used to examine the validity of the MotionSense HRV against criterion measures. The correlations between the MotionSense HRV and GT9X were high and the MAPE were low for both the SED (r = 0.99, MAPE = 2.4%) and MVPA (r = 0.97, MAPE = 9.1%) estimates under free-living conditions. BA plots illustrated that there was no systematic bias between the MotionSense HRV and criterion measures. The estimates of SED and MVPA from the MotionSense HRV were significantly equivalent to those from the GT9X; the equivalence zones were set at 16.5% for SED and 29% for MVPA. The estimates of SED and PA from the MotionSense HRV were less comparable when compared with those from the K5. The MotionSense HRV yielded comparable estimates for SED and PA when compared with the GT9X accelerometer under free-living conditions. We confirmed the promising application of the MotionSense HRV for monitoring PA patterns for practical and research purposes.

TL;DR:

The purpose of this study was to assess the concurrent validity of the MotionSense HRV in estimating sedentary behavior (SED) and PA.

Hydra: Exploiting Multi-Bounce Scattering for Beyond-Field-of-View mmWave Radar
Authors:
Publication Venue:

Proceedings of the 30th Annual International Conference on Mobile Computing and Networking

Publication Date:

December 4, 2024

Keywords:

millimeter-wave, multi-bounce scattering, radar, sensing, beyond-field-of-view, localization.

Related Projects:

SP6

Abstract:

In this paper, we ask, “Can millimeter-wave (mmWave) radars sense objects not directly illuminated by the radar – for instance, objects located outside the transmit beamwidth, behind occlusions, or placed fully behind the radar?” Traditionally, mmWave radars are limited to sense objects that are directly illuminated by the radar and scatter its signals directly back. In practice, however, radar signals scatter to other intermediate objects in the environment and undergo multiple bounces before being received back at the radar. In this paper, we present Hydra, a framework to explicitly model and exploit multi-bounce paths for sensing. Hydra enables standalone mmWave radars to sense beyond-field-of-view objects without prior knowledge of the environment. We extensively evaluate the localization performance of Hydra with an off-the-shelf mmWave radar in five different environments with everyday objects.

TL;DR:

“Hydra” is a novel framework that significantly enhances millimeter-wave (mmWave) radar capabilities by exploiting multi-bounce scattering to sense objects located beyond the radar’s direct field-of-view (FoV). This includes objects that are not directly illuminated, are behind occlusions, or are even completely behind the radar unit. Unlike traditional single-bounce radar methods, Hydra operates without requiring prior knowledge of the environment or additional hardware. It employs a sequential detection and localization process, using earlier single-bounce detections as anchors to find objects via double and triple bounces. Tested on a commercial mmWave MIMO radar, Hydra demonstrates a 2× to 10× improvement in median beyond-FoV localization error for human targets compared to conventional single-bounce techniques.

Emu: Engagement Modeling for User Studies
Authors:

Bo-Jhang Ho, Nima Nikzad, Bharathan Balaji, Mani Srivastava

Publication Venue:

Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing

 

Proceedings of the 2017 ACM International Symposium on Wearable Computers

Publication Date:

December 2017

Keywords:

Context-aware, mobile
applications, engagement,
just-in-time assessment, push notifications, user studies, mhealth, framework

Related Projects:

CP7

Abstract:

Mobile technologies that drive just-in-time ecological momentary assessments and interventions provide an un-precedented view into user behaviors and opportunities to manage chronic conditions. The success of these meth-ods rely on engaging the user at the appropriate moment, so as to maximize questionnaire and task completion rates. However, mobile operating systems provide little support to precisely specify the contextual conditions in which to notify and engage the user, and study designers often lack the ex-pertise to build context-aware software themselves. To ad-dress this problem, we have developed Emu, a framework that eases the development of context-aware study appli-cations by providing a concise and powerful interface for specifying temporal- and contextual-constraints for task no-tifications. In this paper we present the design of the Emu API and demonstrate its use in capturing a range of scenar-ios common to smartphone-based study applications.

TL;DR:

Emu is a framework designed to simplify the development of context-aware mobile study applications. It provides a concise and powerful interface for specifying temporal and contextual constraints for task notifications, addressing the challenge that mobile operating systems offer limited support for precisely engaging users at appropriate moments. By automating the tracking of contextual states and user responses, Emu maximizes questionnaire and task completion rates in studies, helping to manage chronic conditions and improve study protocol adherence. It significantly reduces code complexity for developers compared to native implementations.

StreamQRE: Modular Specification and Efficient Evaluation of Quantitative Queries over Streaming Data
Authors:

Konstantinos Mamouras, Mukund Raghothaman, Rajeev Alur, Zachary G. Ives, Sanjeev Khanna

Publication Venue:

Proceedings of the 38th ACM SIGPLAN Conference on Programming Language Design and Implementation

Publication Date:

December 2017

Keywords:

Data stream processing, Quantitative Regular
Expressions, IoT applications, real-time decision making, relational query languages, regular expressions, modular specification, efficient
evaluation, streaming composition, key-based
partitioning, pattern-based windows, approximate aggregation.

Related Projects:

CP7

Abstract:

Real-time decision making in emerging IoT applications typ-ically relies on computing quantitative summaries of largedata streams in an efficient and incremental manner. To sim-plify the task of programming the desired logic, we proposeStreamQRE, which provides natural and high-level constructsfor processing streaming data. Our language has a novel in-tegration of linguistic constructs from two distinct program-ming paradigms: streaming extensions of relational querylanguages and quantitative extensions of regular expressions.The former allows the programmer to employ relational con-structs to partition the input data by keys and to integrate datastreams from different sources, while the latter can be used toexploit the logical hierarchy in the input stream for modularspecifications.We first present the core language with a small set ofcombinators, formal semantics, and a decidable type system.We then show how to express a number of common patternswith illustrative examples. Our compilation algorithm trans-lates the high-level query into a streaming algorithm withprecise complexity bounds on per-item processing time andtotal memory footprint. We also show how to integrate ap-proximation algorithms into our framework. We report onan implementation in Java, and evaluate it with respect toexisting high-performance engines for processing streamingdata. Our experimental evaluation shows that (1) StreamQREallows more natural and succinct specification of queriescompared to existing frameworks, (2) the throughput of ourimplementation is higher than comparable systems (for ex-ample, two-to-four times greater than RxJava), and (3) theapproximation algorithms supported by our implementationcan lead to substantial memory savings.

TL;DR:

StreamQRE is a novel programming language designed for real-time decision-making in IoT applications by integrating two powerful paradigms: streaming extensions of relational query languages and quantitative extensions of regular expressions. This allows for natural and modular specification of complex queries over large data streams. Its compiler translates high-level queries into efficient streaming algorithms with guaranteed low memory footprint and fast processing times, outperforming existing high-performance streaming engines like RxJava, Esper, and Flink. StreamQRE also supports approximation algorithms for significant memory savings on computationally intensive tasks like median calculation.

mSelf – Using Mobile Sensors to Self-monitor and Improve Health, Wellness, and Performance
Authors:
Publication Venue:

Proceedings of the 2017 Workshop on Wearable Systems and Applications

Publication Date:

June 19, 2017

Keywords:
Related Projects:

CP7

Abstract:

Mobile sensors can track human states, the surrounding context, daily behaviors, and exposures to environmental risk factors in the natural field environment. Real-time analysis of such sensor data makes it possible to deliver personalized recommendations to improve health, wellness, and performance. Widely used GPS-navigation systems that provide just-in-time directions for traffic-aware navigation and activity trackers that help users set and achieve daily physical activity goals are widely used early examples. The increasing availability of mobile sensors that allow collection of raw sensor data, along with mobile big data software platforms that allow labeled collection, curation, modeling, and visualization of such data for development and validation of new markers and sensor-triggered interventions, is opening up exciting new research directions. They include novel sensor systems for selftracking of health, wellness, and performance.

TL;DR:
mCerebrum: A Mobile Sensing Software Platform for Development and Validation of Digital Biomarkers and Interventions
Authors:

Syed Monowar Hossain, Timothy Hnat, Nazir Saleheen, Nusrat Jahan Nasrin, Joseph Noor, Bo-Jhang Ho, Tyson Condie, Mani Srivastava, Santosh Kumar

Publication Venue:

Proceedings of the 15th ACM Conference on Embedded Network Sensor Systems

Publication Date:

November 6, 2017

Keywords:

mHealth, mobile sensor big data, software architecture, digital biomarkers, wearable sensors

Related Projects:

CP7

Abstract:

The development and validation studies of new multisensory biomarkers and sensor-triggered interventions requires collecting raw sensor data with associated labels in the natural field environment. Unlike platforms for traditional mHealth apps, a software platform for such studies needs to not only support high-rate data ingestion, but also share raw high-rate sensor data with researchers, while supporting high-rate sense-analyze-act functionality in real-time. We present mCerebrum, a realization of such a platform, which supports high-rate data collections from multiple sensors with realtime assessment of data quality. A scalable storage architecture (with near optimal performance) ensures quick response despite rapidly growing data volume. Micro-batching and efficient sharing of data among multiple source and sink apps allows reuse of computations to enable real-time computation of multiple biomarkers …

TL;DR:

mCerebrum is an open-source mobile sensing software platform designed for the development and validation of digital biomarkers and sensor-triggered interventions. Unlike traditional mHealth apps, it’s built to handle high-rate raw sensor data collection in natural environments, supporting over 70 million samples per day. The platform features a scalable storage architecture (Pebbles) with near-optimal performance, an efficient data exchange architecture (DataKit) for real-time biomarker computation, and a reconfigurable, burden- and context-aware scheduler for participant prompts. With a modular design spanning over 23 apps, mCerebrum has evolved through its use in scientific field studies at ten sites, accumulating 106,806 person-days of data. Evaluations show it significantly outperforms other platforms like AWARE, HealthKit, and Google Fit in data rates, storage throughput, and CPU usage, enabling efficient processing of multi-sensor biomarkers.

SeleCon: Scalable IoT Device Selection and Control Using HandGestures
Authors:

Amr Alanwar, Moustafa Alzantot, Bo-Jhang Ho, Paul Martin, Mani Srivastava

Publication Venue:

Proceedings of the Second International Conference on Internet-of-Things Design and Implementation

Publication Date:

April 2017

Keywords:

Pointing, IoT, hand gestures, smartwatch, UWB, device selection, smart home, human-computer interaction (HCI), gesture recognition, wearable devices.

Related Projects:

CP7

Abstract:

Although different interaction modalities have been proposed in the field of human-computer interface (HCI), only a few of these techniques could reach the end users because of scalability and usability issues. Given the popularity and the growing number of IoT devices, selecting one out of many devices becomes a hurdle in a typical smarthome environment. Therefore, an easy-to-learn, scalable, and non-intrusive interaction modality has to be explored. In this paper, we propose a pointing approach to interact with devices, as pointing is arguably a natural way for device selection. We introduce SeleCon for device selection and control which uses an ultra-wideband (UWB) equipped smartwatch. To interact with a device in our system, people can point to the device to select it then draw a hand gesture in the air to specify a control action. To this end, SeleCon employs inertial sensors for pointing gesture detection and a UWB transceiver for identifying the selected device from ranging measurements. Furthermore, SeleCon supports an alphabet of gestures that can be used for controlling the selected devices. We performed our experiment in a 9m-by-10m lab space with eight deployed devices. The results demonstrate that SeleCon can achieve 84.5% accuracy for device selection and 97% accuracy for hand gesture recognition. We also show that SeleCon is power efficient to sustain daily use by turning off the UWB transceiver, when a user’s wrist is stationary.

TL;DR:

SeleCon is a novel system that enables natural and scalable interaction with Internet of Things (IoT) devices in smart homes using pointing gestures and hand gestures. Users wear a custom smartwatch equipped with Ultra-Wideband (UWB) and inertial sensors. By pointing to a device, the system identifies the target using UWB ranging data, and then hand gestures drawn in the air control the selected device. SeleCon prioritizes energy efficiency by only activating the power-hungry UWB when wrist motion indicates a potential pointing action. The system demonstrates high accuracy: 84.5% for device selection and 97% for hand gesture recognition.

PhysioGAN: Training High Fidelity Generative Model for Physiological Sensor Readings
Authors:

Moustafa Alzantot, Luis Garcia, Mani Srivastava

Publication Venue:
Publication Date:

April 25, 2022

Keywords:

Physiological sensor

readings, generative
models, GANs, VAEs,
time-series data,
ECG classification, human activity recognition,
synthetic data generation, data imputation, mode collapse mitigation.

Related Projects:
Abstract:

Generative models such as the variational autoencoder (VAE) and the generative adversarial networks (GAN) have proven to be incredibly powerful for the generation of synthetic data that preserves statistical properties and utility of real-world datasets, especially in the context of image and natural language text. Nevertheless, until now, there has no successful demonstration of how to apply either method for generating useful physiological sensory data. The state-of-the-art techniques in this context have achieved only limited success. We present PHYSIOGAN, a generative model to produce high fidelity synthetic physiological sensor data readings. PHYSIOGAN consists of an encoder, decoder, and a discriminator. We evaluate PHYSIOGAN against the state-of-the-art techniques using two different real-world datasets: ECG classification and activity recognition from motion sensors datasets. We compare PHYSIOGAN to the baseline models not only the accuracy of class conditional generation but also the sample diversity and sample novelty of the synthetic datasets. We prove that PHYSIOGAN generates samples with higher utility than other generative models by showing that classification models trained on only synthetic data generated by PHYSIOGAN have only 10% and 20% decrease in their classification accuracy relative to classification models trained on the real data. Furthermore, we demonstrate the use of PHYSIOGAN for sensor data imputation in creating plausible results.

TL;DR:

PHYSIOGAN is a novel generative model that successfully produces high-fidelity, diverse, and novel synthetic physiological sensor data by combining variational autoencoders (VAEs) and generative adversarial networks (GANs). It significantly outperforms state-of-the-art techniques for conditional generation, proving its utility for downstream tasks like training classification models with only a 10% to 20% decrease in accuracy compared to real data, and effectively imputing missing sensor readings.

Multi-objective deep learning for microwave tomography based lymphedema detection
Authors:

Yuyi Chang, James Wright, Emre Ertin

Publication Venue:
Publication Date:

May 9, 2025

Keywords:

Lymphedema Detection, Microwave Tomography, Deep Learning, Medical Diagnostics, B-Scan, Hybrid Loss Function.

Related Projects:
Abstract:
We propose a method to detect and localize lymphedema using tomographic microwave backscatter measurements of the upper limb. Inspired by object detection techniques in ground-penetrating radar (GPR) systems, we reformat the set of tomography scans into a B-Scan representation for analysis.

Prior approaches relied on forming backprojection images to isolate returns from the lymphedema, but these imaging approaches are limited by the low resolution of the system. In contrast, here we employ a deep learning model that processes B-scan images directly, effectively canceling background signals from the layered limb tissues.
 
The model predicts  lymphedema intensity as a function of the azimuth angle, a representation that ensures robustness against variations in limb profiles and the placement of the limb within the sensor array. We incorporate a hybrid loss function for multi-objective deep learning to control the false alarm rate explicitly. To validate the proposed  detection strategy, we generated a new simulated limb phantom dataset comprising 1,410 distinct profiles. Our results demonstrate that the hybrid loss function can be  successfully used to improve lymphedema intensity estimation for negative cases and challenging, positive cases when the radius of the lymphedema is small.
 
Furthermore, our method bypasses the computationally demanding backprojection
imaging step, providing a computationally efficient solution for microwave-based lymphedema diagnosis.
TL;DR:

This paper presents a novel deep learning method for detecting lymphedema (swelling) using microwave tomography scans of a limb. Instead of creating low-resolution images, the approach reformats the raw sensor data into “B-Scans” and uses a deep learning model to directly identify the lymphedema’s location and intensity. This bypasses a computationally heavy imaging step, improves robustness to variations in limb placement, and uses a specialized hybrid loss function to better control false alarms, demonstrating a more efficient and reliable diagnostic solution.

MPADA: Open source framework for multimodal time series antenna array measurements
Authors:

Yuyi Chang, Yingzhe Zhang, Asimina Kiourti, Emre Ertin

Publication Venue:

HumanSys ’25: Proceedings of the 3rd International Workshop on Human-Centered Sensing, Modeling, and Intelligent Systems

Publication Date:

November 1, 2024

Keywords:

Antenna Measurement, Open-Source Framework, Data Acquisition, Sensor Fusion, Time-Synchronization, Vector Network Analyzer.

Related Projects:
Abstract:
This paper presents an open-source framework for collecting time series S-parameter measurements across multiple antenna elements, dubbed MPADA: Multi-Port Antenna Data Acquisition. The core of MPADA relies on the standard SCPI protocol to be compatible with a wide range of hardware platforms. Time series measurements are enabled through the use of a high-precision real-time clock (RTC), allowing MPADA to periodically trigger the VNA and simultaneously acquire other sensor data for synchronized cross-modal data fusion. A web-based user interface has been developed to offer flexibility in instrumentation, visualization, and analysis. The interface is accessible from a broad range of devices, including mobile ones. Experiments are performed to validate the reliability and accuracy of the data collected using the proposed framework. First, we show the framework’s capacity to collect highly repeatable measurements from a complex measurement protocol using a microwave tomography imaging system. The data collected from a test phantom attain high fidelity where a position-varying clutter is visible through coherent subtraction. Second, we demonstrate timestamp accuracy for collecting time series motion data jointly from an RF kinematic sensor and an angle sensor. We achieved an average of 11.8 ms MSE timestamp accuracy at a mixed sampling rate of 10 to 20 Hz over a total of 16-minute test data. We make the framework openly available to benefit the antenna measurement community, providing researchers and engineers with a versatile tool for research and instrumentation. Additionally, we offer a potential education tool to engage engineering students in the subject, fostering hands-on learning through remote experimentation.
TL;DR:

This paper introduces MPADA, an open-source framework designed to simplify and standardize the collection of synchronized, time-series data from multi-antenna arrays and other sensors. By using a high-precision clock to trigger a Vector Network Analyzer (VNA) and acquire additional sensor data, MPADA enables precise cross-modal data fusion. The framework, which includes a versatile web interface, was validated through experiments showing high measurement repeatability and excellent timestamp accuracy, providing the research and education communities with a powerful and accessible tool for advanced antenna measurements.

Joint target recovery and blind calibration of phased-array radar using deep unrolled model

Lorem ipsum dolor sit amet, consectetur adipisicing elit. Optio, neque qui velit. Magni dolorum quidem ipsam eligendi, totam, facilis laudantium cum accusamus ullam voluptatibus commodi numquam, error, est. Ea, consequatur.

MobiVital: Self-supervised quality estimation for UWB-based contactless respiration monitoring
Authors:

Ziqi Wang, Derek Hua, Wenjun Jiang, Tianwei Xing, Xun Chen, Mani Srivastava

Publication Venue:

HumanSys ’25: Proceedings of the 3rd International Workshop on Human-Centered Sensing, Modeling, and Intelligent Systems

Publication Date:

May 6, 2025

Keywords:

Respiration Monitoring, Ultra-Wideband Radar, Waveform Quality, Contactless Sensing, Vital Signs, Biomedical Signal Processing.

Related Projects:
Abstract:
Respiration waveforms are increasingly recognized as important biomarkers, offering insights beyond simple respiration rates, such as detecting breathing irregularities for disease diagnosis. Previous works in wireless respiration monitoring have primarily focused on estimating respiration rate, where the breath waveforms are often generated as a by-product. As a result, issues such as waveform deformation and phase inversion have largely been overlooked, reducing the signal’s utility for applications requiring breathing waveforms. To address this problem, we present a novel approach, MobiVital, that improves the quality of respiration waveforms obtained from ultra-wideband (UWB) radar data. MobiVital combines a self-supervised autoregressive model for breathing waveform extraction with a biology-informed algorithm to detect and correct waveform inversions. To encourage reproducible research efforts for developing wireless vital signal monitoring systems, we also release a 12-person, 24-hour UWB radar vital signal dataset, with time-synchronized ground truth obtained from wearable sensors. Our results show that the respiration waveforms produced by our system exhibit a 7-34% increase in fidelity to the ground truth compared to the baselines and can benefit downstream tasks such as respiration rate estimation.
TL;DR:
This paper introduces MobiVital, a novel system that significantly improves the quality of detailed respiration waveforms captured by ultra-wideband (UWB) radar, moving beyond simple respiration rate estimation. It combines a self-supervised model for waveform extraction with a biology-informed algorithm to correct common issues like phase inversion, resulting in waveforms that are 7-34% more faithful to the ground truth. The authors also contribute a valuable 24-hour dataset to support further research, demonstrating that their high-fidelity waveforms also improve downstream tasks like rate estimation.
ADMN: A layer-wise adaptive multimodal network for dynamic input noise and compute resources
Authors:

Jason Wu, Kang Yang, Lance Kaplan, Mani Srivastava

Publication Venue:
Publication Date:

February 10, 2025

Keywords:

Multimodal Learning, Dynamic Computation, Resource Efficiency, Adaptive Networks, Input Noise, Model Efficiency.

Related Projects:
Abstract:

Multimodal deep learning systems are deployed in dynamic scenarios due to the robustness afforded by multiple sensing modalities. Nevertheless, they struggle with varying compute resource availability (due to multi-tenancy, device heterogeneity, etc.) and fluctuating quality of inputs (from sensor feed corruption, environmental noise, etc.). Current multimodal systems employ static resource provisioning and cannot easily adapt when compute resources change over time. Additionally, their reliance on processing sensor data with fixed feature extractors is ill-equipped to handle variations in modality quality. Consequently, uninformative modalities, such as those with high noise, needlessly consume resources better allocated towards other modalities. We propose ADMN, a layer-wise Adaptive Depth Multimodal Network capable of tackling both challenges – it adjusts the total number of active layers across all modalities to meet compute resource constraints, and continually reallocates layers across input modalities according to their modality quality. Our evaluations showcase ADMN can match the accuracy of state-of-the-art networks while reducing up to 75% of their floating-point operations.

TL;DR:
This paper presents ADMN, a dynamic multimodal network designed to operate efficiently in real-world conditions where compute resources and input quality from different sensors can vary dramatically. Unlike static models, ADMN intelligently adjusts its computational effort in two ways: it scales the total number of active layers to meet changing resource budgets, and it reallocates these layers across modalities in real-time, favoring those with higher-quality inputs. This approach allows ADMN to match the accuracy of state-of-the-art models while reducing computational cost by up to 75%.
WristPrint: Characterizing User Re-identification Risks from Wrist-Worn Accelerometry Data
Authors:
Publication Venue:

International Conference on Learning Representations (ICLR)

Publication Date:

November 13, 2021

Keywords:

privacy, user re-identification, wrist-worn accelerometers

Abstract:
Public release of wrist-worn motion sensor data is growing. They enable and accelerate research in developing new algorithms to passively track daily activities, resulting in improved health and wellness utilities of smartwatches and activity trackers. But, when combined with sensitive attribute inference attack and linkage attack via re-identification of the same user in multiple datasets, undisclosed sensitive attributes can be revealed to unintended organizations with potentially adverse consequences for unsuspecting data contributing users. To guide both users and data collecting researchers, we characterize the re-identification risks inherent in motion sensor data collected from wrist-worn devices in users’ natural environment. For this purpose, we use an open-set formulation, train a deep learning architecture with a new loss function, and apply our model to a new data set consisting of 10 weeks of daily sensor wearing by 353 users. We find that re-identification risk increases with an increase in the activity intensity. On average, such risk is 96% for a user when sharing a full day of sensor data.
TL;DR:

We characterize the re-identification risks inherent in motion sensor data collected from wrist-worn devices in users’ natural environment. We use an open-set formulation, train a deep learning architecture with a new loss function, and apply our model to a new data set.

Aerogel: Lightweight Access Control Framework for WebAssembly-Based Bare-Metal IoT Devices
Authors:
Publication Venue:

IEEE/ACM Symposium on Edge Computing (SEC)

Publication Date:

December 14, 2021

Keywords:

software security engineering, mobile platform security

Abstract:

Application latency requirements, privacy, and security concerns have naturally pushed computing onto smartphone and IoT devices in a decentralized manner. In response to these demands, researchers have developed micro-runtimes for WebAssembly (Wasm) on IoT devices to enable streaming applications to a runtime that can run the target binaries that are independent of the device. However, the migration of Wasm and the associated security research has neglected the urgent needs of access control on bare-metal, memory management unit (MMU)-less IoT devices that are sensing and actuating upon the physical environment. This paper presents Aerogel, an access control framework that addresses security gaps between the bare-metal IoT devices and the Wasm execution environment concerning access control for sensors, actuators, processor energy usage, and memory usage. In particular, we treat the runtime as a multi-tenant environment, where each Wasm-based application is a tenant. We leverage the inherent sandboxing mechanisms of Wasm to enforce the access control policies to sensors and actuators without trusting the bare-metal operating system. We evaluate our approach on a representative IoT development board: a cortexM4 based development board (nRF52840). Our results show that Aerogel can effectively enforce compute resource and peripheral access control policies while introducing as little as 0.19% to 1.04% runtime overhead and consuming only 18.8% to 45.9% extra energy.

TL;DR:

This paper presents Aerogel, an access control framework that addresses security gaps between the bare-metal IoT devices and the Wasm execution environment concerning access control for sensors, actuators, processor energy usage, and memory usage.

Protecting User Data Privacy with Adversarial Perturbations
Authors:
Publication Venue:

Proceedings of the 20th International Conference on Information Processing in Sensor Networks, Pages 386-387

Keywords:

privacy, computing methodologies, machine learning, neural networks

Publication Date:

May 18, 2021

Abstract:

The increased availability of on-body sensors gives researchers access to rich time-series data, many of which are related to human health conditions. Sharing such data can allow cross-institutional collaborations that create advanced data-driven models to make inferences on human well-being. However, such data are usually considered privacy-sensitive, and publicly sharing this data may incur significant privacy concerns. In this work, we seek to protect clinical time-series data against membership inference attacks, while maximally retaining the data utility. We achieve this by adding an imperceptible noise to the raw data. Known as adversarial perturbations, the noise is specially trained to force a deep learning model to make inference mistakes (in our case, mispredicting user identities). Our preliminary results show that our solution can better protect the data from membership inference attacks than the baselines, while succeeding in all the designed data quality checks.

TL;DR:

In this work, we seek to protect clinical time-series data against membership inference attacks, while maximally retaining the data utility. 

Understanding Factors Behind IoT Privacy—A User’s Perspective on RF Sensors
PrivacyOracle: Configuring Sensor Privacy Firewalls with Large Language Models in Smart Built Environments
Authors:
Publication Venue:

2024 IEEE Security and Privacy Workshops (SPW)

Publication Date:

May 1, 2024

Keywords:

Large Language Models (LLM), Privacy, Contextual Integrity,
Smart Environments

Related Projects:
Abstract:

Modern smart buildings and environments rely on sensory infrastructure to capture and process information about their inhabitants. However, it remains challenging to ensure that this infrastructure complies with privacy norms, preferences, and regulations; individuals occupying smart environments are often occupied with their tasks, lack awareness of the surrounding sensing mechanisms, and are non-technical experts. This problem is only exacerbated by the increasing number of sensors being deployed in these environments, as well as services seeking to use their sensory data. As a result, individuals face an unmanageable number of privacy decisions, preventing them from effectively behaving as their own “privacy firewall” for filtering and managing the multitude of personal information flows. These decisions often require qualitative reasoning over privacy regulations, understanding privacy-sensitive contexts.

TL;DR:

This paper introduces PrivacyOracle, a prototype system that leverages Large Language Models (LLMs) to automatically configure privacy firewalls in smart built environments. The system addresses the challenge of managing numerous privacy decisions for individuals in sensor-rich spaces by enabling automated decision-making regarding personal data flows. PrivacyOracle achieves this by performing qualitative reasoning over privacy regulations and social norms, identifying privacy-sensitive states from sensor data, and selecting appropriate data transformation tools. Evaluations show high accuracy in identifying sensitive states (up to 98%) and moderate agreement (75%) with social acceptability norms for information flows.

Systems and Methods for Using Ultrawideband Audio Sensing Systems
Authors:

Ziqi Wang, Mani Srivastava, Akash Deep Singh, Luis Garcia, Zhe Chen, Jun Luo

Publication Venue:

United States Patent Application 20230288549

Publication Date:

September 14, 2023

Keywords:

Ultrawideband Audio
Sensing, Impulse Radio Ultra-Wideband (IR-UWB), wireless vibrometry, sound source separation, Time-of-Flight (ToF), RF sensing, through-wall detection, static clutter suppression,
vibrating activity
localization, audio recover

Related Projects:
Abstract:

Systems and methods for simultaneously recovering and separate sounds from multiple sources using Impulse Radio Ultra-Wideband (IR-UWB) signals are described. In one embodiment, a device can be configured for generating an audio signal based on audio source ranging using ultrawideband signals. In an embodiment the device includes, a transmitter circuitry, a receiver circuitry, memory and a processor. The processor configured to generate a radio signal. The radio signal including an ultra-wideband Gaussian pulse modulated on a radio-frequency carrier. The processor further configured to transmit the radio signal using the transmitter circuitry, receive one or more backscattered signals at the receiver circuitry, demodulate the one or more backscattered signals to generate one or more baseband signals, and generate a set of data frames based on the one or more baseband signals.

TL;DR:

This publication introduces systems and methods for Ultrawideband Audio Sensing that leverage Impulse Radio Ultra-Wideband (IR-UWB) signals to simultaneously recover and separate sounds from multiple sources. Unlike traditional microphones that blend sounds and struggle with background noise, this wireless vibrometry approach senses sound directly from source vibrations using RF signals. Key advantages include the ability to penetrate building materials for non-line-of-sight (NLOS) operation, immunity to non-target noise, and fine-grained sound source separation based on Time-of-Flight (ToF) ranging, even for sources as close as 25 cm apart. The system employs a signal processing pipeline involving phase noise correction, static clutter suppression, vibrating activity localization, and denoising to achieve robust audio recovery.

RefreshChannels: Exploiting dynamic refresh rate switching for mobile device attacks
Authors:

Gaofeng Dong, Jason Wu, Julian De Gortari Briseno, Akash Deep Singh, Justin Feng, Ankur Sarker, Nader Sehatbakhsh, Mani Srivastava

Publication Venue:
Publication Date:

June 04, 2024

Keywords:

Covert Channel, Mobile Security, Dynamic Refresh Rate, Privacy Attack, Side-Channel, OS Sandboxing.

Related Projects:
Abstract:

Mobile devices with dynamic refresh rate (DRR) switching displays have recently become increasingly common. For power optimization, these devices switch to lower refresh rates when idling, and switch to higher refresh rates when the content displayed requires smoother transitions. However, the security and privacy vulnerabilities of DRR switching have not been investigated properly. In this paper, we propose a novel attack vector called RefreshChannels that exploits DRR switching capabilities for mobile device attacks. Specifically, we first create a covert channel between two colluding apps that are able to stealthily share users’ private information by modulating the data with the refresh rates, bypassing the OS sandboxing and isolation measures. Second, we further extend its applicability by creating a covert channel between a malicious app and either a phishing webpage or a malicious advertisement on a benign webpage. Our extensive evaluations on five popular mobile devices from four different vendors demonstrate the effectiveness and widespread impacts of these attacks. Finally, we investigate several countermeasures, such as restricting access to refresh rates, and find they are inadequate for thwarting RefreshChannels due to DDR’s unique characteristics.

TL;DR:
This paper introduces “RefreshChannels,” a novel attack vector that exploits the Dynamic Refresh Rate (DRR) switching feature on modern mobile displays to create covert communication channels. A malicious app can secretly transmit private user data by modulating the screen’s refresh rate, bypassing OS security to communicate with a colluding app, a phishing webpage, or a malicious ad. The attack was proven effective across numerous popular devices, and the study found that standard countermeasures are currently inadequate to stop it due to the fundamental nature of how DRR functions.
Toward foundation models for online complex event detection in CPS-IoT: A case study
Authors:

Liying Han, Gaofeng Dong, Xiaomin Ouyang, Lance Kaplan, Federico Cerutti, Mani Srivastava

Publication Venue:
Publication Date:

May 06, 2025

Keywords:

Complex Event Detection, Foundation Models, State-Space Models, Cyber-Physical Systems, Long-Term Reasoning, Mamba.

Related Projects:
Abstract:

Complex events (CEs) play a crucial role in CPS-IoT applications, enabling high-level decision-making in domains such as smart monitoring and autonomous systems. However, most existing models focus on short-span perception tasks, lacking the long-term reasoning required for CE detection. CEs consist of sequences of short-time atomic events (AEs) governed by spatiotemporal dependencies. Detecting them is difficult due to long, noisy sensor data and the challenge of filtering out irrelevant AEs while capturing meaningful patterns. This work explores CE detection as a case study for CPS-IoT foundation models capable of long-term reasoning. We evaluate three approaches: (1) leveraging large language models (LLMs), (2) employing various neural architectures that learn CE rules from data, and (3) adopting a neurosymbolic approach that integrates neural models with symbolic engines embedding human knowledge. Our results show that the state-space model, Mamba, which belongs to the second category, outperforms all methods in accuracy and generalization to longer, unseen sensor traces. These findings suggest that state-space models could be a strong backbone for CPS-IoT foundation models for long-span reasoning tasks.

TL;DR:
This paper investigates how to build foundation models for detecting long-duration “Complex Events” (a sequence of smaller atomic events) in Cyber-Physical Systems and IoT. It evaluates three approaches—using Large Language Models (LLMs), training neural models on data, and neurosymbolic methods—and finds that a state-space model called Mamba, which learns the rules directly from data, outperforms all others. Mamba demonstrated superior accuracy and an ability to generalize to longer, unseen data sequences, positioning state-space models as a promising backbone for future CPS-IoT foundation models that require long-term reasoning.
Transforming mental health care with autonomous LLM agents at the edge
Authors:

Sijie Ji, Xinzhe Zheng, Wei Gao, Mani Srivastava

Publication Venue:
Publication Date:

May 06, 2025

Keywords:

Autonomous LLM Agents, Mental Health, Mobile Computing, Edge AI, Personalized Interventions, Privacy.

Related Projects:
Abstract:

The integration of Large Language Models (LLMs) with mobile devices is set to transform mental health care accessibility and quality. This paper introduces MindGuard, an autonomous LLM agent that utilizes mobile sensor data and engages in proactive, personalized conversations while ensuring user privacy through local processing. Unlike traditional mental health AI tools, MindGuard enables real-time, context-aware interventions by dynamically adapting to users’ emotional and physiological states. The real-world implementation demonstrates its effectiveness with the ultimate goal of creating an accessible, scalable, and personalized mental healthcare ecosystem for anyone with smart mobile devices.

TL;DR:
This paper introduces MindGuard, an autonomous LLM agent designed to transform mental health care by running locally on mobile devices. It uses on-device sensor data to enable proactive, context-aware, and personalized conversations and interventions, dynamically adapting to a user’s emotional and physiological state. This approach aims to create a scalable and accessible mental health ecosystem that prioritizes user privacy by processing all data on the device itself.
Tinyodom: Hardware-Aware Efficient Neural Inertial Navigation
Authors:
Publication Venue:

ACM on Interactive, Mobile, Wearable, and Ubiquitous Technologies

Publication Date:

July 7, 2022

License:
Languages:

C++

C

Jupyter Notebook

Python

Auritus: An Open-Source Optimization Toolkit for Training & Development of Human Movement Models & Filters Using Earables
Authors:
Publication Venue:

ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies

Publication Date:

July 7, 2022

License:
Languages:

C++

C

Python

MATLAB

Jupyter Notebook

THIN-Bayes: Platform-Aware Machine Learning for Low-End IoT Devices

https://github.com/ARM-software/mango/tree/main/examples/THIN-Bayes

Authors:
Publication Venue:

tinyML Summit

Publication Date:

March 2022

License:
Languages:

Python

CardiacGen: A Hierarchical Deep Generative Model for Cardiac Signals
TinyNS: Platform-Aware Neurosymbolic Auto Tiny Machine Learning
Authors:
Publication Venue:

ACM Transactions on Embedded Computing Systems (2023)

Publication Date:

May 31, 2023

License:
Languages:

C

C++

Assembly

Python

CMake

Makefile

Mobile Open Observation of Daily Stressors (MOODS)
MMBind: Unleashing the Potential of Distributed and Heterogeneous Data for Multimodal Learning in IoT
Reimagining time series foundation models: Metadata and state-space model perspectives
SensorBench: Benchmarking LLMs in Sensor Processing
Benchmarking spatiotemporal reasoning in LLMs and reasoning models: Capabilities and challenges
Bayesian Sparse Blind Deconvolution Using MCMC Methods Based on Normal-Inverse-Gamma Prior


https://github.com/burakcivek/BayesianSparseBlindDeconvolution
0 forks.
2 stars.
0 open issues.

Recent commits:

Authors:
Publication Venue:

IEEE Transactions on Signal Processing

Publication Date:

March 3, 2022

Language:

MATLAB

The Validity of MotionSense HRV in Estimating Sedentary Behavior and Physical Activity under Free-Living and Simulated Activity Settings
Authors:
Publication Venue:

Sensors (Basel)

Publication Date:

February 18, 2021

License:
Languages:

Jupyter Notebook

Python

Shell

MobiVital: Self-supervised Quality Estimation for UWB-based Contactless Respiration Monitoring
MPADA: Open source framework for multimodal time series antenna array measurements
WristPrint: Characterizing User Re-identification Risks from Wrist-Worn Accelerometry Data
Auritus: An Open-Source Optimization Toolkit for Training & Development of Human Movement Models & Filters Using Earables

The Auritus dataset has 2.43 million inertial samples related to head and full-body movements, consisting of 34 head poses and 9 activities from 45 volunteers.  There are three main folders in the dataset, namely Activity DetectionSimple Head Pose and Complex Head Pose. It also contains physiological information about participants in the file Participant_Info.csv, target information for simple and complex head pose in the file HP_target_info.csv, and the IRB approval in the file IRB_Approval.pdf. Simple head pose corresponds to head movements from a origin marker to a target marker and back. Complex head pose corresponds to head movements from a origin marker to target marker A, target marker A to target marker B, and target marker B to origin. 

 

Activity Detection

Complex Head Pose

Simple Head Pose

Authors:
Publication Venue:

ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies

Publication Date:

July 7, 2022

License:
mORAL: A model for inferring oral hygiene behaviors
Cerebral Cortex: A Cloud Tool For Big Data Analysis
STARK_Benchmark
SensorBench: Benchmarking LLMs in Sensor Processing
MobiVital: Self-supervised Quality Estimation for UWB-based Contactless Respiration Monitoring
Vigorous research activity in mHealth has resulted in an ever-growing list of physiological and behavioral markers. However, translation of these biomarkers into real time intervention lagged behind the observational research studies that led to their development due to computation, storage, and communication bottlenecks faced by wearables and smartphone platforms. Further, the next generation of wearables is emerging with the ability to sample data from multiple sensors at rates several orders of magnitude higher than current-generation devices, exacerbating the computational and communication bottleneck. They can image structure, motion, and function, to provide visibility into physiology previously possible only in clinics.

Traditionally, such imaging sensors use post-processing algorithms for feature identification, co-registration, alignment and enhancement. However, high frequency-high volume imaging data from wearables cannot be transported to cloud computing for post-processing. Finally, researchers have shown that the high-dimensionality sensor data needed to compute biomarkers present immense privacy risks. Advances in machine learning are leading to an ever-growing list of surprising inferences about user identity and activities that can be made from seemingly innocuous sensors, particularly when data are captured over long durations. Simplistic methods such as stripping personally identifiable information and addition of noise that focus on anonymizing the data have been ineffective for mHealth, both from privacy and utility perspectives, particularly with the availability of vast amounts of side information (e.g. metadata), computational power, and advanced algorithms.

To address these growing challenges, we propose a hierarchical computing framework that reduces the data into minimal modular abstractions called Micromarkers computed at the edge devices. Micromarkers can be used directly as features in new biomarker inferences or can be adapted to support legacy algorithms. TR&D3 is developing hardware, software, and computational techniques to implement privacy-aware, efficient, and embedded intelligence support into wearables. They will enable continuous, high-throughput, low latency biomarker captures across wearable, mobile, and cloud platforms to support large scale and long-term research studies, and eventual real-life rollout.

Tushar Agarwal, PhD

Senior ML Researcher


Siddharth Baskar, PhD

Sr. Hardware Engineer


Swapnil Sayan Saha, PhD

Algorithm Dev. Engineer


Md Azim Ullah, PhD

Applied Scientist


  1. Civek, Burak C., and Emre Ertin. Bayesian Sparse Blind Deconvolution Using MCMC Methods Based on Normal-Inverse-Gamma Prior. IEEE Transactions on Signal Processing 70 (2022): 1256-1269. NIHMS1839071.
  2. Saleheen, Nazir, Md Azim Ullah, Supriyo Chakraborty, Deniz S. Ones, Mani Srivastava, and Santosh Kumar. WristPrint: Characterizing User Re-identification Risks from Wrist-worn Accelerometry Data. In Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security, pp. 2807-2823. 2021. NIHMS1839082
  3. Liu, Renju, Luis Garcia, and Mani Srivastava. Aerogel: Lightweight Access Control Framework for WebAssembly-Based Bare-Metal IoT Devices. In 2021 IEEE/ACM Symposium on Edge Computing (SEC), pp. 94-105. IEEE, 2021. NIHMS1839084
  4. Saha, Swapnil Sayan, Sandeep Singh Sandha, Luis Antonio Garcia, and Mani Srivastava. Tinyodom: Hardware-Aware Efficient Neural Inertial Navigation. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, no. 2 (2022): 1-32. NIHMS1839164
  5. Saha, Swapnil Sayan, Sandeep Singh Sandha, Siyou Pei, Vivek Jain, Ziqi Wang, Yuchen Li, Ankur Sarker, and Mani Srivastava. Auritus: An Open-Source Optimization Toolkit for Training and Development of Human Movement Models and Filters Using Earables. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, no. 2 (2022): 1-34. NIHMS1839165
  6. Saha, Swapnil Sayan, Sandeep Singh Sandha, and Mani Srivastava. Machine Learning for Microcontroller-Class Hardware--A Review. arXiv preprint arXiv:2205.14550 (2022). Accepted for IEEE J. Sensors ’22.
  7. Saha, Swapnil Sayan, Sandeep Singh Sandha, Mohit Aggarwal, and Mani Srivastava. THIN-Bayes: Platform-Aware Machine Learning for Low-End IoT Devices. Poster at the tinyML Summit 2022.
  8. Bari R, Rahman MM, Saleheen N, Parsons MB, Buder EH, Kumar S. Automated Detection of Stressful Conversations Using Wearable Physiological and Inertial Sensors. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies. 2020 December;4(4). PubMed PMID: 34099995; PubMed Central PMCID: PMC8180313; DOI: 10.1145/3432210.
  9. Akther, S., Saleheen, N., Saha, M., Shetty, V., & Kumar, S. (2021). mTeeth: Identifying Brushing Teeth Surfaces Using Wrist-Worn Inertial Sensors. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 5 (2) DOI: 10.1145/3463494.
  10. Wang Z, Wang B, Srivastava M. Poster Abstract: Protecting User Data Privacy with Adversarial Perturbations. IPSN. 2021 May;2021:386-387. doi: 10.1145/3412382.3458776. PMID: 34651144; PMCID: PMC8513393.
  11. Agarwal T,  Ertin E. CardiacGen: A Hierarchical Deep Generative Model for Cardiac Signals. 2022. ArXiv, abs/2211.08385.
  12. Wenqiang Chen, Ziqi Wang, Pengrui Quan, Zhencan Peng, Shupei Lin, Mani Srivastava, Wojciech Matusik, and John Stankovic. 2023. Robust Finger Interactions with COTS Smartwatches via Unsupervised Siamese Adaptation. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology (UIST '23). Association for Computing Machinery, New York, NY, USA, Article 25, 1–14. DOI: 10.1145/3586183.3606794.
  13. Y. Chang, N. Sugavanam, E. Ertin. Removing Antenna Effects using an Invertible Neural Network for Improved Estimation of Multilayered Tissue Profiles using UWB Radar. 2023 IEEE USNC-URSI Radio Science Meeting (Joint with AP-S Symposium), Portland, OR, USA, 2023, pp. 53-54, DOI: 10.23919/USNC-URSI54200.2023.10289171.
  14. Kwon S, Wan N, Burns RD, Brusseau TA, Kim Y, Kumar S, Ertin E, Wetter DW, Lam CY, Wen M, Byun W. The Validity of MotionSense HRV in Estimating Sedentary Behavior and Physical Activity under Free-Living and Simulated Activity Settings. Sensors (Basel). 2021 Feb 18;21(4):1411. DOI: 10.3390/s21041411. PMID: 33670507; PMCID: PMC7922785.
  15. Swapnil Sayan Saha, Sandeep Singh Sandha, Mohit Aggarwal, Brian Wang, Liying Han, Julian de Gortari Briseno, and Mani Srivastava. 2023. TinyNS: Platform-Aware Neurosymbolic Auto Tiny Machine Learning. ACM Trans. Embed. Comput. Syst. Just Accepted (May 2023). https://doi.org/10.1145/3603171
  1. S.S. Saha, S.S. Sandha, S. Pei, V. Jain, Z. Wang, Y. Li, A. Sarker, M. Srivastava. Workshop: Excerpt of Auritus: An Open-Source Optimization Toolkit for Training and Development of Human Movement Models and Filters Using Earables. ACM EarComp Workshop, 2022. Cambridge, UK. 9/15/2022.

Santosh Kumar, PhD

Lead PI, Center Director, TR&D1, TR&D2, TR&D3


Yuyi Chang

Doctoral Student


Sugavanam Nithin

Doctoral Student


Akash Sing Deep

Doctoral Student


Brian Wang

Doctoral Student


Soumobrata Ghosh

Doctoral Student


Devan Mallory

Undergraduate Student