Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article
Open access

Detecting Smartwatch-Based Behavior Change in Response to a Multi-Domain Brain Health Intervention

Published: 07 April 2022 Publication History
  • Get Citation Alerts
  • Abstract

    In this study, we introduce and validate a computational method to detect lifestyle change that occurs in response to a multi-domain healthy brain aging intervention. To detect behavior change, digital behavior markers are extracted from smartwatch sensor data and a permutation-based change detection algorithm quantifies the change in marker-based behavior from a pre-intervention, 1-week baseline. To validate the method, we verify that changes are successfully detected from synthetic data with known pattern differences. Next, we employ this method to detect overall behavior change for n = 28 brain health intervention subjects and n = 17 age-matched control subjects. For these individuals, we observe a monotonic increase in behavior change from the baseline week with a slope of 0.7460 for the intervention group and a slope of 0.0230 for the control group. Finally, we utilize a random forest algorithm to perform leave-one-subject-out prediction of intervention versus control subjects based on digital marker delta values. The random forest predicts whether the subject is in the intervention or control group with an accuracy of 0.87. This work has implications for capturing objective, continuous data to inform our understanding of intervention adoption and impact.

    1 Introduction

    The connection between behavior and health is undeniable. Unhealthy behaviors and habits are implicated in as much as 40% of premature deaths, in addition to their unfavorable effects on health disparities, quality of life, and economics [1]. Advances in sensor-driven technologies for behavior monitoring suggest that these technologies can monitor intervention adherence and impact in the wild [24]. However, our current understanding of intervention on a person's entire lifestyle is limited because studies have been based on static “snapshots” or assessment of a specific set of behavior parameters (e.g., sleep quality, activity level) rather than data-driven analysis of overall behavior change. Given this limitation, there is a need to introduce algorithmic methods for quantifying behavior change that is characterized by a person's entire pattern of behavior.
    In this work, we introduce an automated method to perform continuous monitoring and change detection of daily activities, in naturalistic settings, using sensor-based observation and machine learning based data analysis. To do this, we perform the steps shown in Figure 1. We collect wearable sensor data for 2 months while subjects perform their daily routines. We process the collected sensor data by merging data from multiple watches, imputing missing values, and automatically associating data sequences with corresponding activity labels. Next, we extract digital behavior markers from activity-labeled data that summarize the key features of each person's behavior on an hourly and daily basis. Finally, we apply a behavior change detection to the digital behavior markers. This algorithm quantifies and describes behavior differences between time points.
    Fig. 1.
    Fig. 1. Process of detecting behavior change with smartwatches. Data are continuously collected, merged, completed, and labeled with inferred activities. From these data, digital markers are extracted for each hour and day. We quantify the amount of change from beginning to end of data collection for the experiment and non-intervention groups using behavior change detection and compare the amount of change between the two groups.
    To validate our approach, we analyze data from the active condition of a pilot multi-domain Brain Health Intervention (BHI) for midlife and older adults. We also compare the data with an age-matched control group that did not receive an intervention. The 7-week pilot intervention, called B-Fit, was designed to improve engagement in aging brain health behaviors by providing brain health psychoeducation combined with individual behavior goal setting, group problem solving, and social support [5]. Participants wore smartwatches during the intervention that tracked behaviors. Control participants similarly wore smartwatches. Seven risk factors were targeted by the intervention: nutrition, stress, social engagement, cognitive engagement, cardiovascular risk factors, physical activity, and sleep. To promote new patterns of engagement in healthy brain aging behaviors and better sustain behavior change, participants set individualized, intrinsically motivating, and manageable new goals (5–10 minutes) each week (e.g., take stairs rather than the elevator). Results from the primary analysis revealed that self-reported engagement in healthy brain aging behaviors was higher for the B-Fit condition and an education-only condition at post-intervention, whereas there was no difference for the waitlist condition. Furthermore, after adjusting for baseline brain health behaviors, compared to the waitlist condition, participants in the B-Fit condition self-reported greater engagement in healthy brain aging behaviors post-intervention [5].
    Because BHI is a multi-domain intervention where participants set their own diverse goals, computational methods are needed to complement self-report data by providing a method to objectively detect and quantify change in overall behavior patterns rather than analyzing a small set of target markers. We initially apply our proposed computational approach to synthetic data to validate their ability to detect known embedded changes in behavior data. We then apply our methods to smartwatch data collected from BHI and control subjects. We hypothesize that behavior change over time will be exhibited among BHI subjects. Because change can be attributed to natural behavior variability, we further postulate that behavior change will be larger for BHI subjects than for control subjects. In addition to quantifying change over time for the participants, we also employ supervised learning to automatically classify each subject as belonging to the intervention or control group, based solely on changes in the extracted digital markers between baseline and the end of the data collection period.
    This article makes the following contributions. First, we offer a novel method to extract digital markers from time series sensor data that characterize a person's activity routine. Second, we introduce a way to analyze behavior across time based on a permutation detection algorithm. Third, we computationally analyze the impact of a multi-domain brain intervention study, contrasting behavior change of intervention participants with an age-matched control group.

    2 Related Work

    2.1 Activity Learning

    To characterize a person's overall behavior routine, we extract digital behavior markers from sensor data that are automatically labeled with corresponding activity categories. Performing human activity recognition from wearable sensors has become a popular topic for researchers to investigate [6, 7]. Because we can collect uninterrupted data in all locations each person visits, wearable sensors are a natural choice for this investigation. Earlier approaches to activity recognition consider many diverse learning models, including decision trees, nearest neighbors, clusters, and classifier ensembles [810]. With the rising popularity of deep learning, these methods have also been explored for their ability to learn deep features that are useful in expressing sensed activities [1114]. The field has matured to the point that numerous public datasets are now available for comparative analysis of recognition methods [11].
    Limitations with many of these existing methods are that they focus on basic, repetitive movement types and are typically evaluated in laboratory settings. Recognizing postures and ambulatory motions such as sitting, standing, climbing, lying down, and running are useful for monitoring gait characteristics and activity levels. However, we are interested in determining change to a person's entire behavioral routine. This is frequently characterized by more complex patterns, such as basic and instrumental activities of daily living (iADLs) [15, 16]. A few researchers consider activity categories that combine ambulation with context (e.g., standing indoors versus outdoors) [8, 17, 18]. They accomplish this by adding location information. This location information is provided by ambient motion sensors inside a specific physical environment [17] or a grid location within a specific physical region [8]. A challenge with analyzing specific locations is that the insights do not generalize well to new individuals. Lin and Hsu [19] introduced generalizable location features that we employ in our method, including heading change rate, distance covered, and velocity change rate. We further integrate ideas from Boukhechba et al. [20], who cluster location readings to detect person-specific frequented spots. Although we extend these previous methods to perform real-time activity labeling of smartwatch sensor data in naturalistic settings, the success these researchers have documented in recognizing a wide assortment of activities suggests the general applicability of activity learning and labeling for smartwatch data.

    2.2 Digital Behavior Markers

    Sensor data offer substantial insights on a person's behavior as well as their health. Now that wearable sensors are decreasing in cost and are deployable in real-world settings, researchers have discovered that they can link behavior interventions with specific, sensor-observed target behavior outcomes. As wearable sensor systems also become more ubiquitous, researchers have increasingly used these platforms to impact and measure change in human behavior. Typically, these prior works target a specific behavior. Popular target behaviors include activity levels [21, 22] and sleep [23]. In these cases, change can be quantified by a single variable such as step counts, sleep duration, or a clinical score [24]. Jain et al. [25] designed a method to analyze the univariate change in the state of a person or place, such as the amount of activity in a region of the home. Wang et al. [26] calculated the variability of a single variable, then joined the variability of multiple features into a single vector to predict personality traits. In a method that utilizes mobile platforms for intervention delivery as well as evaluation, Costa et al. [27] delivered haptic feedback to participants intended to slow their heartbeats while completing math tests. They observed impacts based on self-reported anxiety, sensor-measured heart rate variability, and test performance. In a study conducted by Stanley et al. [28], individuals with dementia received coaching in behavior techniques designed to reduce anxiety. Here, outcomes were measured through clinician-rated and collateral report clinical scores targeted at assessing anxiety. Additional studies monitor single wearable markers to assess intervention impact for target conditions such as substance use disorder [29] and diabetes management [30].
    In other work, researchers targeted specific sensor-observed behavior markers as a mechanism for assessing the relationship between lifestyle and health. Specifically, Dhana et al. [31] quantify healthy behavior as a combination of non-smoking, physical activity, alcohol consumption, nutrition, and cognitive activities. Individuals who scored higher on this behavior metric had lower risk of Alzheimer dementia. Other researchers have also found that sensor-based behavior patterns are predictive of cognitive health [32, 33]. Li et al. [33] found that physical activity level was predictive of Alzheimer's disease, whereas Aramendi et al. [32] were able to predict cognitive health and mobility from activity-labeled sensor data.
    These studies provide evidence that wearable sensors afford the ability to monitor intervention impact and to assess a person's cognitive health. Within this area of investigation, our proposed approach is unique because we investigate a computational method to analyze intervention impact on the pattern of a person's activity-based routine. This holistic approach to sensor analysis of behavior is motivated by the design of a holistic, customized, sustainable healthy behavior intervention.

    2.3 Time Series Change Detection

    A critical component of this work is automatically detecting and quantifying change in behavior data observed longitudinally. Multiple options are available for computing change between two windows of multivariate time series data [34]. Supervised methods can classify a pair of time windows as “no change” or “change.” Such methods can also identify changes of a specific nature (e.g., a change from sedentary to active behavior), but these algorithms must be supplied with a sufficient amount of labeled training data [35]. Unsupervised approaches have been introduced as well. Some methods rely on pre-designed parametric models [36, 37], whereas others introduce more flexible non-parametric variations by directly estimating the ratio of probability densities for the two samples. In these cases, a larger ratio implies a greater likelihood that a change occurred between the two samples. Methods that fit in this category are KLIEP, which employs Kullback-Leibler (KL) divergence as a ratio estimator [38]; uLSIF, which employs Pearson divergence [38]; and Relative uLSIF (RuLSIF) [38]. In this work, we utilize a Permutation-Based Change Detection (PCD) method. This method is not only suited to comparing larger windows of sensor data but also incorporates a novel permutation mechanism to determine the significance of change over time.

    2.4 Brain Health Intervention

    An accumulating body of research suggests that adopting preventative health behaviors may promote healthy brain aging and slow cognitive decline [39]. Given that the etiology of dementia is heterogeneous and influenced by multiple risk factors, several recent large-scale prevention trials have focused on evaluation of multi-domain interventions [40, 41]. However, these interventions are currently limited by expensive treatment regimens and highly prescriptive goals that are difficult to sustain post-intervention. Furthermore, research has found that intervention success is promoted by healthy behavior education as well as designing interventions that are transferable out of the lab and into daily lives [42, 43]. In contrast with prior methods, the pilot group BHI involved creating individualized behavior change goals across multiple risk factor domains that could be integrated into the everyday lifestyles of participants and more easily sustained over time.

    3 BHI Study

    The objective of the pilot BHI study was to develop and evaluate an intervention designed to promote changes in healthy brain behavior for midlife and older adults that is practical and easily replicated. The rationale is that a preventative BHI designed to reduce disability in older age will be most successful if it meets individuals where they are and promotes healthy behavior change goals that can be more easily sustained. Furthermore, the approach is motivated by findings that multiple risk factors influence the development of dementia. Researchers estimate that almost half of dementia cases are attributable to modifiable risk factors including physical inactivity and depression [44]. Studies [45] indicate that engaging in healthy lifestyles behaviors across multiple domains including physical activity, cognitive engagement, and stress management can slow cognitive decline [4651] and promote brain neuroplasticity [52, 53]. There currently exists a gap in the design of methods to assist midlife and older adults in learning to incorporate such healthy behaviors into their everyday lives in a way that leads to sustained, holistic behavior change.

    3.1 Intervention Design

    Our BHI, called B-Fit, is a 7-week intervention focused on providing brain health psychoeducation combined with individual behavior goal setting, group problem solving, and social support. The primary pilot clinical intervention study included three conditions: B-Fit, education-only, and waitlist [5]. Of the 68 participants allocated to the B-Fit intervention in the primary study, 28 of the participants wore a smartwatch during the intervention. These individuals were members of three separate B-Fit groups. The B-Fit intervention was delivered by two clinicians with groups of 8 to 13 participants. One learning topic was introduced to the group each week. During didactics, information about each topic was discussed with an emphasis on empirical research describing mechanisms by which lifestyles changes can improve brain health (e.g., reduce blood pressure and oxidative stress, increase cerebral blood flow) and contribute to a healthy aging phenotype. This approach is based on literature indicating the importance of linking healthy lifestyle behaviors with brain health and prevention research [54, 55]. The goal of this work is to introduce and validate a mechanism for assessing broad-spectrum lifestyle changes associated with this individualized, multi-domain intervention.
    Table 1 lists the primary health topics covered during the weekly sessions. Each week, following the educational material presentation, participants developed a plan to integrate a personalized, realistic, sustainable, and positive lifestyle change into their everyday life (e.g., take a 10-minute walk daily after lunch). Health goals were built across multiple lifestyle domains consistent with the education topics (e.g., diet, socialization, sleep). Because a new goal was introduced each week, participants were assisted with developing goals for each topic that could be completed in about 5 to 10 minutes. At the beginning of each session, for each defined goal, participants indicate whether they made no change (0), partly met (1), completely met (2), or exceeded (3) their goal(s). Twenty-six of the 28 participants who wore a smartwatch regularly reported these values. The average self-reported adherence over the 8-week intervention was 1.7 (range [0.5, 2.3], standard deviation 0.44).
    Table 1.
    SessionEducation Materials Discussed
    1Overview of the brain, cognitive aging, mild cognitive impairment, and dementia
    2Cognitive engagement
    3Physical activity and cardiovascular risk factors
    4Nutrition and diet
    5Social engagement
    6Sleep and stress reduction
    7Compensatory strategies and assistive technology
    Table 1. Didactic Topics Discussed in Group Sessions

    3.2 Data Collection

    In addition to evaluating self-reported outcomes, determining the impact of interventions such as B-Fit on the pattern of behaviors can be implemented by passively and continuously monitoring behavior across time. Automated approaches to such ecologically valid monitoring and assessment pave the way for designing more effective treatment and quantitatively measuring the outcome of pharmaceutical and behavioral interventions. This ability will be particularly valuable in rural locations where access to the clinic or lab is limited [56]. Ecologically valid monitoring implies that a person's routine behavior is observed in naturalistic settings rather than asking individuals to perform scripted activities in controlled settings. This is particularly important for the B-Fit intervention because we want to assess the impact of the intervention on a person's entire set of activities, not on a few selected markers. Currently, holistically assessing change in overall behavior patterns in such a manner is difficult outside of manually observing a person's daily routine. We determine whether behavior change exists using a computational behavior change detection method. We further evaluate the pattern of behavior change over the course of the intervention.
    To provide objective behavior routine and change measures, we gave 28 of the B-Fit participants two Apple Watches to wear continuously—one during the day and one at night (one watch was charged while the other was worn). We collected timestamped readings at 10 Hz for the following sensors: 3D acceleration, 3D rotation and rotation rate, course, speed, and 3D location. Participants received two motivational quotes through the smartwatch at random times during the day to support their lifestyle changes. Data were continuously collected at baseline (before intervention) and each week of the intervention. All data were collected before quarantine restrictions due to COVID-19.

    3.3 Participants

    All study participants are community-dwelling adults. We recruited participants from eastern Washington and northern Idaho. Exclusion criteria included a clinical diagnosis of dementia; inability to provide own consent; unstable or severely disabling disease (e.g., organ failure); and inability to complete assessment and intervention protocols due to communication, vision, hearing, or other medical difficulties. Prior to participation, participants were screened by phone with a medical/health interview and the Telephone Interview of Cognitive Status (TICS) [57]. Participants were excluded if they fell in the Impaired range on the TICS (score <26) or met study exclusion criteria. The study was approved by the Washington State University Institutional Review Board.
    We analyze data for 28 study participants who received the active brain health education and goal-setting intervention (B-Fit) described in Section 2.1 and continuously wore smartwatches. This group is labeled “BHI.” To provide a comparison (non-intervention) group for behavior change detection, we analyze continuous smartwatch sensor data collected for 17 additional community-dwelling older adults who did not differ in age, t(43) = 1.08, p = .28, education, t(43) = .17, p = .87, or sex, X2 (1, N = 45) = .33, p = .57, from the BHI group. This group is labeled “NI.” These participants were older adults who transformed their homes into smart homes and participated in a non-intervention, monitoring study designed to help validate smart home health algorithms. Table 2 summarizes the demographics for these participants.
    Table 2.
    GroupAge
    (mean/SD)
    Years of Education (mean/SD)MaleFemaleCaucasianOther Ethnicity
    BHI (n = 28)63.18/9.1916.00/1.86525264
    NI (n = 17)67.22/15.8915.89/2.49413161
    Table 2. Participant Demographics

    4 Processing Collected Data

    We collected multiple weeks of smartwatch sensor data for each of the subjects, totaling 28,615,680 sets of readings for BHI participants and 18,464,560 sets of readings for the non-intervention group. We examine the first 9 weeks of data collection for each subject. This length of time corresponds to the 1 week of baseline, 7 weeks of intervention, and 1 post-intervention week for the BHI participants. Our goal is to examine change in behavior that occurs over the course of the intervention. Before analyzing the data, we first merge data from multiple watches, impute missing data, and label collected data with activity labels.

    4.1 Data Preparation

    Participants were given a white watch to wear during the day and a black watch to wear at night, affording time to charge the devices. The smartwatches indicate when they are on and off the charging station, allowing us to determine when data should be analyzed from one device or the other. However, participants were not consistent in wearing the designed watches during the specified times. To ensure we analyze the appropriate data, we examine the acceleration and rotation-based movement throughout each day. We merge day and night data collections into one by identifying “swapping” periods, or time points at which the subject swapped one watch for another. This was defined as a change point in the time series when one watch was at rest (total movement below a threshold amount) and the other was active. We create one merged set of sensor readings for each day by incorporating data at each time unit from the active watch.
    Additionally, gaps occurred during data collection. Data for participants were 94.24% complete. Gaps in the data collection occurred during times when subjects forgot to charge and wear the watch. In these cases, we impute missing feature values based on the median value for the particular sensor given the day of the week, time of day, and neighboring context (readings before and after the gap).

    4.2 Activity Labeling

    We automatically label each set of sensor readings with a corresponding activity name. Activity labeling offers several benefits for our analysis. First, activities provide a vocabulary for expressing human behavior. From activity-labeled sensor data, we can see how a person spends their day and observe changes in these activity routines over time. Second, given activity-labeled data, we can target specific activity categories, such as work, exercise, and sleep, to ascertain whether the amount of time that is spent on these specific activities is changing, if the location where they occur changes, and if they occur at regular times.
    We label sensor data using an activity recognition algorithm. Activity recognition maps a sequence of sensor readings onto an activity label. A time series data stream is an infinite sequence of elements \(S=\{x_{1},\ldots ,x_{i},\ldots\) , where xi is a d-dimensional data vector arriving at timestamp i. Given a set of pre-defined activity categories, A, an activity recognition algorithm maps a sequence of time series data points \(\{x_{m},\ldots ,x_{k},\ldots , x_{n}\) , observed as part of time series S, onto an activity label aj ∈ A. We map a fixed-length sequence of elements onto the activity label using the process illustrated in Figure 2.
    Fig. 2.
    Fig. 2. Activity labeling. A feature vector is extracted from 5 seconds of sensor data. A random forest maps the feature vector onto one of the activity choices, based on training data provided in real time by smartwatch users.
    To ensure that we label and monitor complex activities as well as basic movement-based activities, we also incorporate location information. Because we want to ensure that the model generalizes to multiple persons, we do not utilize specific location values. Instead, we construct person-invariant location features. This approach supports generalization without needing to employ techniques such as domain adaptation, which requires a large number of labeled training instances for the initial group of subjects [5860]. One such person-invariant location feature is the location type, obtained through reverse geocoding using an OpenStreetMap. We group these location categories into residence, workplace, road, and other. Other location features are calculated for each sensor sequence and include the number of changes in orientation in the sequence, number of stops/starts, the sequence overall trajectory, and the distance traversed.
    Attempting to perform real-time human activity recognition “in the wild” in naturalistic settings raises numerous challenges. One such challenge is collecting ground-truth labeled data. Users can be queried for their activities. However, these queries need to be answered in the moment to avoid retrospective error. Although active learning and semi-supervised learning can reduce the number of queries [61, 62], the resulting interactions will interrupt many of the activities we wish to monitor for our brain health study. To make the intervention as naturalistic as possible, we do not ask the subjects to provide labels for their daily activities. Instead, we train an activity recognition algorithm using past data collected from 388 different smartwatch wearers. These earlier volunteers provided labels for their current activity when prompted at random times. These prompts resulted in 59,610 labeled instances for 12 activity categories. These activity categories are reflective of routine behavior and include chores, eat, entertainment, errands, exercise, hobby, hygiene, relax, school, sleep, travel, and work.
    Each 5-second sequence of sensor data is considered as a separate instance. Although shorter sequences have been used with success for ambulation recognition [63], we found that slightly longer durations are needed to capture context for recognizing complex activities in naturalistic settings. A feature vector is extracted from each 5-second sensor sequence. The feature vector consists of standard statistical operations applied to each of the sensor types [64]. The vector also contains person-invariant location features. We experimented with multiple types of supervised learning algorithms and found random forest to perform the best for these data. The random forest also generates a prediction for one sequence before the next sensor reading arrives, allowing the activity recognition to occur in real time. The random forest employs bootstrapping and is configured to contain 100 trees with a depth limit of 10. Because the training data are not uniformly distributed across the activity classes, we weight instances inversely proportionate to the relative size of the corresponding activity class. Applying threefold cross validation to our labeled data, activity recognition yielded an accuracy of 0.86 for the 12 activity classes. For the remainder of this article, we analyze sensor data that is labeled using a random forest model trained on all the labeled instances.

    4.3 Extracting Digital Behavior Markers

    Once continuous wearable sensor data are collected and labeled with activity classes, we extract digital markers that describe the person's behavior patterns. Extracting features from longitudinal sensor data is popular for analyzing specific behavior parameters. As mentioned in Section 2, researchers typically extract indicators that are specific to a behavior of interest. Depending on the goal of the study, they may include amount of screen time, activity level, steps, incoming/outgoing phone calls and texts, number of steps, amount of screen time, and sleep. We are interested in characterizing a person's behavior routine. This is expressed as a set of digital markers that express how a person's time is spent on different activities, the locations a person frequents and the amount of time they spend there, and the traditional measures of overall movement types and durations. Obtaining digital behavior markers that express activity routines as well as direct sensor measurements of location and behavior are possible because we automatically label sensor sequences with activity labels, as described in Section 4.2. Thus, in addition to generating statistical summaries of raw sensor readings as listed in Figure 2 [6568], we also extract markers that indicate how each subject divides their time among the activity categories and where they spend their time on a routine basis.
    Table 3 summarizes all of these digital behavior markers that we extract and analyze. Our activity recognition algorithm applies statistical operators (listed in Figure 2) to raw sensor readings, yielding a feature vector that is mapped to an activity label. When assembling digital behavior markers, we apply these same statistical operators to aggregated daily and hourly sensor values, activity information, and location information. Additionally, we compute a regularity index to each variable as defined by Wang et al. [26]. The regularity index determines how repetitive a person's behavior routine is. This is computed as the difference between hourly statistics occurring during the same hours across two different days. To calculate this index, data are first scaled to the range [–0.5, 0.5]. The regularity index comparing days a and b is then defined as shown in Equation (1).
    \begin{equation} R{I_{a,b}}\mathop \sum \limits_{t = 1}^T \frac{{x_t^ax_t^b}}{T} \end{equation}
    (1)
    Table 3.
    TypeFeatures
    DailyAmount of time (in minutes) spent on each activity Time of day (hours and minutes past midnight) of first and last occurrence for each activity Amount of time spent at each location type Step count, activity level, distance traveled Number of missing values (minutes with no sensor readings)
    HourlyAmount of time spent on each activity type and each location type Total rotation, acceleration, distance traveled
    OverallRegularity index (within week, within weekdays, between weeks)
     Diurnal and circadian rhythms
    Table 3. Behavioral Markers
    For our analysis, T is 24 hours and \(x_t^a\\) represents the value of a behavior marker for hour t of day a. We average the regularity index over all pairs of days occurring within the same week, all pairs of weekdays occurring within the same week, and between different weeks for the same day-of-the-week value. Finally, we compute the diurnal and circadian rhythms for each of the hourly values. We follow guidelines by Depner et al. [69] to define the circadian phase as constancy in the behavior routine. We quantify this definition by generating a periodogram for the marker values and calculating the normalized periodogram values for a 24-hour cycle (circadian) and for the 12 hours following the end of nighttime sleep (diurnal). Based on Fourier analysis, the spectral energy of each cycle frequency can be calculated as shown in Equation (2).
    \begin{equation} R = {\left( {\frac{2}{N}\mathop \sum \limits_{i = 1}^N {x_i}{\rm{cos}}\left( {\frac{{2\pi j{t_i}}}{N}} \right)} \right)^2} + {\left( {\frac{2}{N}\mathop \sum \limits_{i = 1}^N {x_i}{\rm{sin}}\left( {\frac{{2\pi j{t_i}}}{N}} \right)} \right)^2} \end{equation}
    (2)
    In Equation (2), N represents the number of samples in the time series; \({x_i}\) represents the value of the measured variable at time i; and the spectral energy is computed for a particular frequency, j. We calculate the spectral energy for a range of possible frequencies. The resulting periodogram generates values for a set of possible cycle lengths. The corresponding circadian rhythm parameter is then represented by the normalized periodogram-derived value for a 24-hour cycle, whereas the 12-hour post-sleep cycle is computed for the diurnal parameter. At the completion of the feature extraction process, our system has constructed a vector containing 102 behavior features for each observed day, 40 features for each observed hour, and 863 overall features for each person.

    5 Detecting Changes in Activity Patterns

    We next consider the challenge of detecting and quantifying change in behavioral routine patterns. To address this problem, we introduce a method called permutation-based change detection (PCD), which determines if a change exists between two windows of activity-labeled multivariate time series data. PCD quantifies the amount of change and determines the significance of the change. Algorithm 1 outlines the process. Let X denote a sample of multivariate time series data segmented into days. Each day, D, is further segmented into equal-sized time intervals, \(D=\{x_{1},x_{2},\ldots, x_{m}\}.\) We let W denote a window of n days such that WX.
    We compare windows of data within time series X. Let Wi refer to a window that starts at day i. Then Wi = X[i : i + n – 1] = \(\{D_{i},D_{i+1},\ldots, D_{i+n-1}\}.\) The compared windows may represent consecutive time periods (e.g., days, weeks, months) or overlapping windows. In our case, we compare a single baseline window with each subsequent time period. Specifically, we compare the first week of observed behavior data, W1, with each following week Wi, i = {2.9}. In our analysis, W1 data were collected before the intervention began, and W2.W9 were collected during the intervention, and W2 data were collected 1 week post-intervention. We compute the change that occurs between week 1 and weeks 2 through 9.
    We propose a small-window PCD approach to quantifying the amount of change, C, between two windows of time series sensor data. This algorithm is uniquely suited for analyzing activity-labeled sensor data. PCD is based on the notion of an activity curve [70], a compilation of m = 24 probability distributions Rt = \(\{d_{t,1},d_{t,2},\ldots, d_{t,A}\}.\) over A possible activities per time interval t in a day \((t=1, 2, \ldots, m).\) Distribution dt,l represents the probability of activity Al occurring during time interval t. Windows spanning multiple days (in our analysis, spanning 1 week) are averaged into an aggregate activity curve \(\hat{A}\) . To compute a distance between two activity curves \({\hat{A}_i}\) and \({\hat{A}_j}\) , the two curves are first aligned by dynamic time warping. Second, a symmetric version of KL divergence computes the distance between each pair of distributions at time interval t, as shown in Equation (3).
    \begin{equation} K{L_{symmetric}} = \mathop \sum \limits_{k = 1}^a {d_{i,k}} \cdot log\frac{{{d_{i,k}}}}{{{d_{j,k}}}} + {\rm{\ }}\mathop \sum \limits_{k = 1}^a {d_{j,k}} \cdot log\frac{{{d_{j,k}}}}{{{d_{i,k}}}} \end{equation}
    (3)
    We adapt a test introduced by Ojala and Garriga [71] to determine whether the distance results are statistically significant. Statistical significance calculation is not easily available for prior methods such as uLSIF and Relative uLSIF (RuLSIF). Our adaptation allows a permutation test to determine the significance of change scores calculated for a small number (n = 7) of days. As shown in Algorithm 1, change score C is computed between windows Wi and Wj. Next, Wi and Wj are joined to form a new longer window. All time intervals within W are shuffled, then W is re-split into new windows, \(W_i^*\) and \(W_j^*\) . A change score C* is computed between the new windows and added to a permutation-result vector V*. The process is repeated N times. Finally, PCD compares C to the permutation vector V* using boxplot outlier detection. Here, an outlier is defined as a value that is outside the interquartile range (75th percentile – 25th percentile) of V*. The p-value estimated by computing the distance between C is an outlier, then the score is considered statistically significant. We employ boxplot outlier detection for this task because many alternative methods, such as Grubb's test, require that the data follow a normal distribution, an assumption that frequently does not apply to human behavior data. The output of the PCD algorithm is a vector V of change scores and associated significance vectors, each of which compares 1 week of observed human behavior data to the first baseline week.

    6 Experimental Results

    We are interested in analyzing the impact of the B-Fit BHI on the pattern of behaviors for the study participants. We start by analyzing one specific targeted behavior and then look at overall behavior. We hypothesize that study participants changed their behavior in response to the BHI. We further postulate that these changes will be detected by PCD applied to activity-labeled smartwatch sensor data. Additionally, we want to determine whether overall behavior change in the intervention group will reflect a steadier increase over the 9 weeks than for the non-intervention control group, consistent with the integration of new goals and maintenance of the earlier defined goals.

    6.1 Validating PCD on Synthetic Data

    First, we verify the reliability of our computational methods that apply PCD-based change detection to digital behavior markers as a method of quantifying change in overall behavior routine. To verify that PCD computes change scores as expected for digital marker data, we examine scores that are generated for synthetic data. We use real digital markers from one randomly selected intervention subject. We then add white Gaussian noise with monotonically increasing values for the Gaussian distribution standard deviation. Table 4 verifies that the PCD-based change score increases monotonically with the amount of noise, or change, that is added to the real collected data.
    Table 4.
    σ
    0.00.10.20.3
    0.00050.82430.85100.8908
    Table 4. PCD Values for Digital Markers with White Additive Gaussian Noise

    6.2 Assessing Change in Target Activity

    Next, we analyze the collected sensor data to determine how much change occurs for one of the targeted activities. Our primary goal of this work is to determine whether our computational method can be used to detect and quantify change in overall behavior activity patterns as promoted by the intervention. However, we also posit that the same method can be used to detect change in a specific activity. The set of activities that are automatically recognized are listed in Section 2. The digital markers reflect these specific activities as well as parameters such as visited locations and movement statistics. Here, we pick a representative activity to demonstrate that the computational method can be used to analyze individual activities as needed.
    We select exercise as the individual activity to analyze. One rationale for this choice is because the week discussing physical activity occurs early in the intervention. If behavior change is sustained per the intervention goal, change should be sustained throughout the data collection period in comparison with the week 0, pre-intervention baseline. A second motivation is that exercise represents a small set of possibilities for the participants (primarily walking and swimming). In contrast, the intervention topic that occurs the week before, cognitive engagement, is manifest through a large variety of individualized healthy behavior choices (e.g., play guitar, read a book, learn a new language), not all of which can be easily detected by smartwatch sensors. We include a subset of the digital markers for this analysis, namely the amount of time spent per day on the automatically labeled exercise activity, time of day that exercise occurs, the regularity of these activities on a daily basis, and statistical operations applied to these markers.
    Table 5 summarizes the results of this analysis. The PCD values indicate that there is a greater change in exercise habits for the intervention group than for the control group. These changes are statistically significant for more than half of the group. Because PCD provides change amounts without indicating the nature of the change, we also compute linear regression for exercise time and note that the slope for the intervention group is slightly higher than for the non-intervention group. In fact, the coefficient for the non-intervention group is close to zero, indicating that the PCD-detected changes may be due in some cases to natural variation in exercise regimens rather than intentional increase in exercise amount.
    Table 5.
    PCD Intervention (%sig)PCD Control (%sig)Coefficient Intervention (R)Coefficient Control (R)
    0.3500 (53.57%)0.1478 (29.41%)0.1016 (0.25)0.0008 (0.40)
    Table 5. Change in Exercise Activity from Baseline to End of Intervention
    Note: This table summarizes the average PCD value for the intervention group and the control group together with the number of persons who experienced a statistically significant change. The table also summarizes the average linear regression coefficient and R value for each of the groups.

    6.3 Assessing Overall Change in Behavior Routine

    The analysis of change in exercise markers confirms that some behavior changes can be detected by wearable sensors and that our PCD algorithm can quantify these changes. These findings align with prior work that detects changes in exercise patterns [72]. However, we are particularly interested in observing the impact of a behavioral intervention on a person's entire pattern of behavior. Researchers have discovered interplay between many behaviors, including but not limited to exercise and work [73], work, travel time, and social life [74], and diet and sleep [75]. Therefore, we hypothesize that engagement in the BHI will impact overall behavior patterns. We analyze the complete set of digital markers collected for study participants and for the non-intervention group and compute PCD scores that compare the baseline week with the end-of-collection week.
    The results of this analysis appear in Table 6. As anticipated, the changes are greater for the intervention group than for the control group. However, weekly changes were still fairly high for the non-intervention group. When we plot weekly change scores for individuals within the two groups, we gain new insights on the nature of the change. Almost all of the individuals in the intervention group exhibited consistently increasing change from baseline, as shown in Figure 3 (left). In contrast, the trends were highly varied for the non-intervention group. Figure 3 (right) provides one such example illustrating a common theme among non-intervention participants. To quantify these trends, we compute the linear regression coefficient for the change scores.
    Table 6.
    PCD Intervention (%sig)PCD Control (%sig)Coefficient Intervention (R)Coefficient Control (R)
    0.7460 (50.00%)0.4428 (17.64%)0.1598 (0.76)0.0230 (0.37)
    Table 6. Change in Overall Activity Routine from Baseline to End of Intervention
    Note: This table summarizes the average PCD value for the intervention group and the control group together with the number of persons who experienced a statistically significant change. The table also summarizes the average linear regression coefficient and R value for each of the groups.
    Fig. 3.
    Fig. 3. PCD score trends. Left: A plot of the PCD scores for a subject in the intervention group. Right: A plot of PCD scores for a subject in the non-intervention group.

    6.4 Assessing Change in Location

    Next, we examine the widespread impact of a BHI on a person's entire daily routine by assessing change in the types of locations where a person spends their time. To visualize the impact that intentional change can have on location patterns, we plotted the normalized, aggregated time spent in the home and outside the home (work, socializing, hobbies, or exercising) the week before intervention initiation and the last week of the intervention. Figure 4 shows this visualization for a 24-hour timeframe. As the graph shows, end-of-intervention routines include more time spent at home in the late-night and early-morning hours. One possible explanation for this is a change in behavior to incorporate more regular sleep hours and patterns. Furthermore, the midday hours reveal less time spent in travel (driving or riding other public transportation) and more time spent in other locations out of the home. This could potentially be attributed to incorporating more exercise, social times, and hobbies into the person's daily routine. Change in location markers between 0 and 9 were much greater for the intervention group, averaging 0.8469, whereas change for the non-intervention group averaged 0.3279.
    Fig. 4.
    Fig. 4. Visualization of time spent at location types for the intervention group. The x-axis shows a 24-hour time period starting at midnight. Bar sizes indicate the aggregated, normalized duration spent at the corresponding location type.

    6.5 Predicting Subject Group Based on Delta Markers

    In our final analysis, we employ a supervised learning algorithm to classify each subject into their corresponding group (intervention or control) based on digital behavior markers. Because we are focused on behavior changes related to the BHI, we compute delta features to perform the classification. Specifically, we calculate the numeric change in each marker from week 0 to week 9. The set of delta features is mapped to the predicted subject group using a range of alternative supervised learning algorithms. To quantify the impact of the intervention on a person's overall activity routine, we compare the results using all delta features with those derived from using a subset of the available features. Specifically, we compute predictive accuracy using only delta features derived from movement-based sensors, labeled activity features, or location type features.
    The results of this analysis are summarized in Table 7. We utilize a random forest with 100 trees and a varied number of features. Results are computed using leave-one-subject-out validation. As the table reflects, the delta values (calculated between the first and last weeks of data collection) were predictive of the subject group with an accuracy of 0.87 when all the features were included in the learned model. The ROC curve for this problem is shown in Figure 5. As the figure indicates, the corresponding AUC (area under the receiver operating characteristic curve) is 0.95.
    Table 7.
    MovementActivityLocationAll
    0.760.730.800.87
    Table 7. Accuracy of Predicting Whether a Subject Is in the Intervention Group or Control Group Using a Random Forest
    Note: Results for are reported for leave-one-subject-out classification and for alternative sets of features.
    Fig. 5.
    Fig. 5. ROC curve for random forest-based prediction of subject group, using pre- and post-intervention changes in digital markers as predictive features. The corresponding AUC is 0.95.
    Using subsets of features still resulted in an accuracy that was better than random chance. However, considering changes in only movement features, activity features, or location features did not produce as strong results as considering changes in the entire set of features, corresponding to the person's larger behavior routine context.

    7 Discussion and Conclusion

    In this study, we investigate methods to detect targeted and secondary behavior changes in response to a multi-modal brain health behavior intervention. By collecting continuous wearable sensor data, we can now monitor intervention impact in an objective, ecologically valid manner. Through application of PCD, we can quantify change in a subset of target behavior markers or in a person's overall pattern of behavior. For the BHI, we did observe that targeted and secondary behaviors changed throughout the data collection process. Although change was also observed for the non-intervention group, linear regression revealed that the amount of change steadily increased more for the intervention group than the non-intervention group. This may be an indication of intentional introduction of new, healthy behaviors into the subjects’ daily routines.
    The system we describe for detecting and quantifying behavior change relies on numerous interacting components. A limitation of this study is the error that can be introduced by each component in the system. The average self-reported compliance with the intervention was 1.7, falling between “partly met” and “completely met” participant-defined goals. As a result, not all intended healthy behaviors were regularly incorporated into the subjects’ lives. Furthermore, weekly self-report may introduce error due to poor recall and objective assessment of behavior adherence. In future work, we will compare the reliability of smartwatch-detected behavior adherence, weekly paper-based self-reporting, and daily self-reporting using ecological momentary assessment techniques delivered through the watch interface.
    Although the smartwatches are designed to continuously collect data, there are gaps in the data collection when the participant did not charge and/or wear the watch. This may introduce error as a function of collecting data when a person is not wearing the watch or imputing missing data. The sensors themselves can occasionally report erroneous location or movement values [76], affecting the quality of the activity labels and digital markers. Similarly, the activity recognition model is based on a large sample of participant-labeled data. Because recognition accuracy was not perfect even on this sample, activity label error may be introduced into the final set of markers, as may error in labeling location categories.
    The findings from this work have important health care implications. By observing behavior patterns before, during, and after a clinical intervention, we can better understand mechanisms and person-characteristics that lead to short-term and long-term change after an intervention. With continued development, this introduced architecture will provide a more continuous, objective measurement of the impact of behavioral interventions on changes in behavior patterns, something that has been difficult to capture clinically but is an important part of interventions. These insights can help design more adaptive, efficacious, community-based preventive interventions for midlife and older adults at risk of developing dementia.

    References

    [1]
    U. S. Department of Health and Human Services. 2015. Healthy People 2020. U. S. Department of Health and Human Services, Washington, DC.
    [2]
    M. H. Aung, M. Matthews, and T. Choudhury. 2017. Sensing behavioral symptoms of mental health and delivering personalized interventions using mobile technologies. Depression and Anxiety 34, 7 (2017), 603–609.
    [3]
    V. Mishra. 2019. From sensing to intervention for mental and behavioral health. In Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers (UbiComp/ISWC’19 Adjunct). 388–392.
    [4]
    B. Das, D. J. Cook, N. Krishnan, and M. Schmitter-Edgecombe. 2016. One-class classification-based real-time activity error detection in smart homes. IEEE Journal of Selected Topics in Signal Processing 10, 5 (2016), 914–923.
    [5]
    B. Boyd, C. McAlister, K. Arrotta, and M. Schmitter-Edgecombe. 2022. Self-reported behavior change and predictors of engagement with a multidomain brain health intervention for midlife and older adults: A pilot clinical trial. Journal of Aging and Health 34, 1 (2022), 109–119.
    [6]
    K. Chen, D. Zhang, L. Yao, B. Guo, Z. Yu, and Y. Liu. 2020. Deep learning for sensor-based human activity recognition: Overview, challenges and opportunities. Journal of the ACM 37, 4 (2020), 111.
    [7]
    A. Bulling, U. Blanke, and B. Schiele. 2014. A tutorial on human activity recognition using body-worn inertial sensors. ACM Computing Surveys 46, 3 (2014), 107–140.
    [8]
    Y. Asim, M. A. Azam, M. Ehatisham-ul-Haq, U. Naeem, and A. Khalid. 2020. Context-aware human activity recognition (CAHAR) in-the-wild using smartphone accelerometer. IEEE Sensors 8 (2020), 4361–4371.
    [9]
    Y. Tian, J. Zhang, L. Chen, Y. Geng, and X. Wang. 2019. Selective ensemble based on extreme learning machine for sensor-based human activity recognition. Sensors 19, 16 (2019), 3468.
    [10]
    A. Nazabal, P. Garcia-Moreno, A. Artes-Rodriguez, and Z. Ghahramani. 2016. Human activity recognition by combining a small number of classifiers. IEEE Journal of Biomedical and Health Informatics 20, 5 (2016), 1342–1351.
    [11]
    J. Wang, Y. Chen, S. Hao, X. Peng, and L. Hu. 2019. Deep learning for sensor-based activity recognition: A survey. Pattern Recognition Letters 119 (2019), 3–11.
    [12]
    N. Y. Hammerla, S. Halloran, and T. Ploetz. 2016. Deep, convolutional, and recurrent models for human activity recognition using wearables. In Proceedings of the International Joint Conference on Artificial Intelligence.
    [13]
    T. Ploetz and Y. Guan. 2018. Deep learning for human activity recognition in mobile computing. Computer 51, 5 (2018), 50–59.
    [14]
    Y. Guan and T. Ploetz. 2017. Ensembles of deep LSTM leaners for activity recognition using wearables. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies. 1, 2 (2017), Article 11, 28 pages.
    [15]
    G. A. Marshall, D. M. Rentz, M. T. Frey, J. J. Locascio, K. A. Johnson, and R. A. Sperling. 2011. Executive function and instrumental activities of daily living in mild cognitive impairment and Alzheimer's disease. Alzheimer's and Dementia 7 (2011), 300–308.
    [16]
    L. S. Miller, C. L. Brown, M. B. Mitchell, and G. M. Williamson. 2013. Activities of daily living are associated with older adult cognitive status: Caregiver versus self-reports. Journal of Applied Gerontology 32 (2013), 3–30.
    [17]
    P. Bharti, D. De, S. Chellappan, and S. K. Das. 2019. HuMAn: Complex activity recognition with multi-modal multi-positional body sensing. IEEE Transactions on Mobile Computing 18, 4 (2019), 857–870.
    [18]
    L. Peng, L. Chen, M. Wu, and G. Chen. 2019. Complex activity recognition using acceleration, vital sign, and location data. IEEE Transactions on Mobile Computing 18, 7 (2019), 1488–1498.
    [19]
    M. Lin and W.-J. Hsu. 2014. Mining GPS data for mobility patterns: A survey. Pervasive and Mobile Computing 12 (2014), 1–16.
    [20]
    M. Boukhechba, Y. Huang, P. Chow, K. Fua, B. A. Teachman, and L. E. Barnes. 2017. Monitoring social anxiety from mobility and communication patterns. In Proceedings of the ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2017 ACM International Symposium on Wearable Computers (UbiComp’17). 749–753.
    [21]
    A. Kankanhalli, M. Saxena, and B. Wadhwa. 2019. Combined interventions for physical activity, sleep, and diet using smartphone apps: A scoping literature review. International Journal of Medical Informatics 123 (2019), 54–67.
    [22]
    G. Sprint and D. J. Cook. 2016. Unsupervised detection and analysis of changes in everyday physical activity data. Journal of Biomedical Informatics 63 (2016), 54–65.
    [23]
    M. de Zambotti, N. Cellini, A. Goldstone, I. M. Colrain, and F. C. Baker. 2019. Wearable sleep technology in clinical and research settings. Medicine and Science in Sports and Exercise 51, 7 (2019), 1538–1557.
    [24]
    R. Wang, W. Wang, A. daSilva, J. F. Huckins, W. M. Kelley, T. F. Heatherton, and A. T. Campbell. 2018. Tracking depression dynamics in college students using mobile phone and wearable sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 2, 1 (2018), Article 43, 26 pages.
    [25]
    A. Jain, M. Popescu, J. Keller, M. Rantz, and B. Markway. 2019. Linguistic summarization of in-home sensor data. Journal of Biomedical Informatics 96 (2019), 103240.
    [26]
    W. Wang, G. M. Harari, R. Wang, S. R. Muller, S. Mirjafari, K. Masaba, and A. T. Campbell. 2018. Sensing behavioral change over time: Using within-person variability features from mobile sensing to predict personality traits. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 2, 3 (2018), 141.
    [27]
    J. Costa, F. Guimbretier, M. F. Jung, and T. Choudhury. 2019. BoostMeUp: Improving cognitive performance in the moment by unobtrusively regulating emotions with a smartwatch. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3, 2 (2019), Article 40, 23 pages.
    [28]
    M. A. Stanley, J. Calleo, A. L. Bush, N. Wilson, A. L. Snow, C. Kraus-Schuman, A. L. Paukert, et al. 2013. The peaceful mind program: A pilot test of a cognitive-behavioral therapy-based intervention for anxious patients with dementia. American Journal of Geriatric Psychiatry 21, 7 (2013), 696–708.
    [29]
    S. Carreiro, M. Newcomb, R. Leach, S. Ostrowski, E. D. Boudreaux, and D. Amante. 2020. Current reporting of usability and impact of mHealth interventions for substance use disorder: A systematic review. Drug and Alcohol Dependence 215 (2020), 108201.
    [30]
    J. Liu, D. L. Spakowicz, G. I. Ash, R. Hoyd, R. Ahluwalia, A. Zhang, S. Lou et al. 2021. Bayesian structural time series for biomedical sensor data: A flexible modeling framework for evaluating interventions. PLoS Computational Biology 17, 8 (2021), e1009303.
    [31]
    K. Dhana, D. A. Evans, K. B. Rajan, D. A. Bennett, and M. C. Morris. 2020. Healthy lifestyle and the risk of Alzheimer dementia: Findings from 2 longitudinal studies. Neurology 95, 4 (2020), e374–e383.
    [32]
    A. Alberdi Aramendi, A. Weakley, M. Schmitter-Edgecombe, D. J. Cook, A. Aztria, A. Basarab, and M. Barrenechea. 2018. Smart home-based prediction of multi-domain symptoms related to Alzheimer's disease. IEEE Journal of Biomedical and Health Informatics 22, 6 (2018), 1720–1731.
    [33]
    J. Li, Y. Rong, H. Meng, Z. Lu, T. Kwok, and H. Cheng. 2018. TATC: Predicting Alzheimer's disease with actigraphy data. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 509–518.
    [34]
    S. Aminikhanghahi and D. J. Cook. 2017. A survey of methods for time series change point detection. Knowledge and Information Systems 51, 2 (2017), 339–367.
    [35]
    H. Ismail Fawaz, G. Forestier, J. Weber, L. Idoumghar, and P.-A. Muller. 2019. Deep learning for time series classification: A review. Data Mining and Knowledge Discovery 33, 4 (2019), 917–963.
    [36]
    H. Cho and P. Fryzlewicz. 2015. Multiple-change-point detection for high dimensional time series via sparsified binary segmentation. Journal of the Royal Statistical Society 77, 2 (2015), 475–507.
    [37]
    K. Yamanishi and J. Takeuchi. 2002. A unifying framework for detecting outliers and change points from non-stationary time series data. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 676.
    [38]
    S. Liu, M. Yamada, N. Collier, and M. Sugiyama. 2013. Change-point detection in time-series data by relative density-ratio estimation. Neural Networks: The Official Journal of the International Neural Network Society 43 (2013), 72–83.
    [39]
    G. Livingston, J. Huntley, A. Sommerlad, D. Ames, C. Ballard, S. Banerjee, C. Brayne et al. 2020. Dementia prevention, intervention, and care: 2020 report of the Lancet Commission. Lancet 396, 10248 (2020), 413–446.
    [40]
    S. Andrieu, S. Guyonnet, N. Coley, C. Cantet, M. Bonnefoy, S. Bordes, L. Bories et al. 2017. Effect of long-term omega 3 polyunsaturated fatty acid supplementation with or without multidomain intervention on cognitive function in elderly adults with memory complaints (MAPT): A randomized, placebo-controlled trial. Lancet Neurology 16, 5 (2017), 377–389.
    [41]
    A. Rosenberg, R. Ngandu, M. Rusanen, R. Antikainen, L. Backman, S. Havulinna, T. Hanninen et al. 2018. Multidomain lifestyle intervention benefits a large elderly population at risk for cognitive decline and dementia regardless of baseline characteristics: The FINGER trial. Alzheimer's & Dementia 14, 3 (2018), 263–270.
    [42]
    C. Hartmann, S. Dohle, and M. Siegrist. 2015. A self-determination theory approach to adults’ healthy body weight motivation: A longitudinal study focusing on food choices and recreational physical activity. Psychology and Health 30, 8 (2015), 924–948.
    [43]
    C. L. Huang, S. C. Yang, and C. H. Chiang. 2017. The associations between individual factors, eHealth literacy, and health behaviors among college students. Journal of Medical Internet Research 19, 1 (2017), e15.
    [44]
    S. Norton, F. Matthews, D. E. Barnes, K. Yaffe, and C. Brayne. 2014. Potential for primary prevention of Alzheimer's disease: An analysis of population-based data. Lancet Neurology 13, 8 (2014), 788–794.
    [45]
    National Academies of Sciences, Engineering and Medicine. 2017. Preventing Cognitive Decline and Dementia: A Way Forward. National Academies Press, Washington, DC.
    [46]
    C. Phillips. 2017. Lifestyle modulators of neuroplasticity: How physical activity, mental engagement, and diet promote cognitive health during aging. Neural Plasticity 2017 (2017), 3589271.
    [47]
    M. Samadi, S. Moradi, M. Moradinazar, R. Mostafai, and Y. Pasdar. 2019. Dietary pattern in relation to the risk of Alzheimer's disease: A systematic review. Neurological Sciences 40 (2019), 2031–2043.
    [48]
    V. Venkatraman, A. Sanderson, K. L. Cox, K. Ellis, C. Steward, and P. M. Phal. 2020. Effect of a 24-month physical activity program on brain changes in older adults at risk of Alzheimer's disease: The AIBL active trial. Neurobiology of Aging 89 (2020), 132–141.
    [49]
    B. Kilmova, P. Maresova, and K. Kamil. 2019. Alzheimer's disease: Physical activities as an effective intervention tool—A mini-review. Current Alzheimer Research 16, 2 (2019), 166–171.
    [50]
    J. Chen, Y. Duan, H. Li, L. Lu, J. Liu, and C. Tang. 2019. Different durations of cognitive stimulation therapy for Alzheimer's disease: A systematic review and meta-analysis. Clinical Interventions in Aging 14 (2019), 1243–1254.
    [51]
    K. E. Cherry, E. M. Jackson Walker, J. L. Silva Brown, J. Volaufova, L. R. LaMotte, and L. J. Su. 2013. Social engagement and health in younger, older, and oldest-old adults in the Louisiana Healthy Aging Study (LHAS). Journal of Applied Gerontology 32 (2013), 51–75.
    [52]
    J.-F. L. Lin, T. Imada, and P. K. Kuhl. 2019. Neuroplasticity, bilingualism, and mental mathematics: A behavior-MEG study. Brain and Cognition 134 (2019), 122–134.
    [53]
    D. Wahl, S. M. Solon-Biet, V. C. Cogger, L. Fontana, S. J. Simpson, D. G. Le Couteur, and R. V. Ribeiro. 2019. Aging, lifestyle and dementia. Neurobiology of Disease 130 (2019), 104481.
    [54]
    M. Schmitter-Edgecombe and D. Dyck. 2014. A cognitive rehabilitation multi-family group intervention for individuals with mild cognitive impairment and their care-partners. Journal of the International Neuropsychological Society 20, 9 (2014), 897–908.
    [55]
    C. S. Park, M. Troutman-Jordan, and M. A. Nies. 2012. Brain health knowledge in community-dwelling older adults. Educational Gerontology 38, 9 (2012), 650–657.
    [56]
    A. P. S. Guerrero, R. Balon, E. V. Beresin, A. K. Louie, J. H. Coverdale, A. Brenner, and L. W. Roberts. 2019. Rural mental health training: An emerging imperative to address health disparities. Academic Psychiatry 43 (2019), 1–5.
    [57]
    J. Brandt and M. Folstein. 2003. Telephone Interview for Cognitive Status. Psychological Assessment Resources, Lutz. FL.
    [58]
    G. Wilson, J. Doppa, and D. J. Cook. 2020. Multi-source deep domain adaptation with weak supervision for time-series sensor data. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’20). 1768–1778.
    [59]
    J. Wang, V. W. Zheng, Y. Chen, and M. Huang. 2018. Deep transfer learning for cross-domain activity recognition. In Proceedings of the 2018 International Conference on Crowd Science and Engineering. 16.
    [60]
    S. A. Rokni and H. Ghasemzadeh. 2018. Autonomous training of activity recognition algorithms in mobile sensors: A transfer learning approach in context-invariant views. IEEE Transactions on Mobile Computing 17, 8 (2018), 1764–1777.
    [61]
    Y. Ma and H. Ghasemzadeh. 2019. LabelForest: Non-parametric semi-supervised learning for activity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence.
    [62]
    T. Miu, P. Missier, and T. Plotz. 2015. Bootstrapping personalised human activity recognition models using online active learning. In Proceedings of the IEEE International Conference on Computer and Information Technology.
    [63]
    O. Banos, J.-M. Galvez, M. Damas, H. Pomares, and I. Rojas. 2014. Window size impact in human activity recognition. Sensors 14, 4 (2014), 6474–6499.
    [64]
    H. Haresamudram, D. V Anderson, and T. Ploetz. 2019. On the role of features in human activity recognition. In Proceedings of the 2019 International Symposium on Wearable Computers. 78–88.
    [65]
    E. Nosakhare and R. Picard. 2020. Toward assessing and recommending combinations of behaviors for improving health and well-being. ACM Transactions on Computing for Healthcare 1, 1 (2020), 4.
    [66]
    M. Obuchi, J. F. Huckins, W. Wang, A. daSilva, C. Rogers, E. Murphy, E. Hedlund, P. Holtzheimer, S. Mirjafari, and A. Campbell. 2020. Predicting brain functional connectivity using mobile sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 4, 1 (2020), 23.
    [67]
    D. J. Cook, M. Schmitter-Edgecombe, L. Jonsson, and A. V Morant. 2019. Technology-enabled assessment of functional health. IEEE Reviews in Biomedical Engineering, 12 (2019), 319–332.
    [68]
    D. Li, J. Vaidya, M. Wang, B. Bush, C. Lu, M. Kollef, and T. Bailey. 2020. Feasibility study of monitoring deterioration of outpatients using multimodal data collected by wearables. ACM Transactions on Computing for Healthcare 1, 1 (2020), Article 5, 22 pages.
    [69]
    C. M. Depner, P. C. Cheng, J. K. Devine, S. Khosla, M. de Zambotti, R. Robillard, A. Vakulin, and S. P. A. Drummond. 2020. Wearable technologies for developing sleep and circadian biomarkers: A summary of workshop discussions. Sleep 43, 2 (2020), zsz254.
    [70]
    P. N. Dawadi, D. J. Cook, and M. Schmitter-Edgecombe. 2016. Modeling patterns of activities using activity curves. Pervasive and Mobile Computing 28 (2016), 51–68.
    [71]
    M. Ojala and C. G. Garriga. 2010. Permutation tests for studying classifier performance. Journal of Machine Learning Research 11 (2010), 1833–1863.
    [72]
    B. Hu and T. Charniak. 2019. Wearable technological platform for multidomain diagnostic and exercise interventions in Parkinson's disease. International Review of Neurobiology 147 (2019), 75–93.
    [73]
    T. Jindo, Y. Kai, N. Kitano, K. Tsunoda, T. Nagamatsu, and T. Arao. 2020. Relationship of workplace exercise with work engagement and psychological distress in employees: A cross-sectional study from the MYLS study. Preventive Medicine Reports 17 (2020), 101030.
    [74]
    N. Nisic and S. Kley. 2020. Gender-specific effects of commuting and relocation on a couple's social life. Demographic Research 40, 36 (2020), 1047–1062.
    [75]
    H. Binks, G. E. Vincent, C. Gupta, C. Irwin, and S. Khalesi. 2020. Effects of diet on sleep: A narrative review. Nutrients 12, 4 (2020), 936.
    [76]
    P. Alinia, C. Cain, R. Fallahzadeh, A. Shahrokni, D. J. Cook, and H. Ghasemzadeh. 2017. How accurate is your activity tracker? A comparative study of step counts in low-intensity physical activities. Journal of Medical Internet Research 5, 8 (2017), e106.

    Cited By

    View all
    • (2024)Ubi-AD: Towards Ubiquitous, Passive Alzheimer Detection using the SmartwatchACM Transactions on Sensor Networks10.1145/3656174Online publication date: 3-Apr-2024
    • (2024)CAvatarProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36314247:4(1-24)Online publication date: 12-Jan-2024
    • (2024)The Difficulties of Managing Pain in People Living with Frailty: The Potential for Digital PhenotypingDrugs & Aging10.1007/s40266-024-01101-441:3(199-208)Online publication date: 24-Feb-2024
    • Show More Cited By

    Index Terms

    1. Detecting Smartwatch-Based Behavior Change in Response to a Multi-Domain Brain Health Intervention

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image ACM Transactions on Computing for Healthcare
          ACM Transactions on Computing for Healthcare  Volume 3, Issue 3
          July 2022
          251 pages
          EISSN:2637-8051
          DOI:10.1145/3514183
          Issue’s Table of Contents

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          Published: 07 April 2022
          Accepted: 01 December 2021
          Revised: 01 April 2021
          Received: 01 November 2020
          Published in HEALTH Volume 3, Issue 3

          Permissions

          Request permissions for this article.

          Check for updates

          Author Tags

          1. Behavior change detection
          2. machine learning from time series
          3. activity recognition
          4. behavior intervention

          Qualifiers

          • Research-article
          • Refereed

          Funding Sources

          • National Institutes of Health
          • National Science Foundation

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)957
          • Downloads (Last 6 weeks)116
          Reflects downloads up to 11 Aug 2024

          Other Metrics

          Citations

          Cited By

          View all
          • (2024)Ubi-AD: Towards Ubiquitous, Passive Alzheimer Detection using the SmartwatchACM Transactions on Sensor Networks10.1145/3656174Online publication date: 3-Apr-2024
          • (2024)CAvatarProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36314247:4(1-24)Online publication date: 12-Jan-2024
          • (2024)The Difficulties of Managing Pain in People Living with Frailty: The Potential for Digital PhenotypingDrugs & Aging10.1007/s40266-024-01101-441:3(199-208)Online publication date: 24-Feb-2024
          • (2023)Hierarchical Clustering-based Personalized Federated Learning for Robust and Fair Human Activity RecognitionProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/35807957:1(1-38)Online publication date: 28-Mar-2023
          • (2023)Artificial Intelligence for Cognitive Health Assessment: State-of-the-Art, Open Challenges and Future DirectionsCognitive Computation10.1007/s12559-023-10153-415:6(1767-1812)Online publication date: 24-Jun-2023

          View Options

          View options

          PDF

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format.

          HTML Format

          Get Access

          Login options

          Full Access

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media