Abstract
Mobile Health (mHealth) has the potential to be transformative in the management of chronic conditions. Machine learning can leverage self-reported data collected with apps to predict periods of increased health risk, alert users, and signpost interventions. Despite this, mHealth must balance the treatment burden of frequent self-reporting and predictive performance and safety. Here we report how user engagement with a widely used and clinically validated mHealth app, myCOPD (designed for the self-management of Chronic Obstructive Pulmonary Disease), directly impacts the performance of a machine learning model predicting an acute worsening of condition (i.e., exacerbations). We classify how users typically engage with myCOPD, finding that 60.3% of users engage frequently, however, less frequent users can show transitional engagement (18.4%), becoming more engaged immediately (â<â21 days) before exacerbating. Machine learning performed better for users who engaged the most, however, this performance decrease can be mostly offset for less frequent users who engage more near exacerbation. We conduct interviews and focus groups with myCOPD users, highlighting digital diaries and disease acuity as key factors for engagement. Users of mHealth can feel overburdened when self-reporting data necessary for predictive modelling and confidence of recognising exacerbations is a significant barrier to accurate self-reported data. We demonstrate that users of mHealth should be encouraged to engage when they notice changes to their condition (rather than clinically defined symptoms) to achieve data that is still predictive for machine learning, while reducing the likelihood of disengagement through desensitisation.
Similar content being viewed by others
Introduction
Chronic diseases are the leading cause of death and disability worldwide and represent 75% of the cost of healthcare1,2. As well as long-term care plans (with adherence crucial for health outcomes, quality of life and minimising healthcare cost) effective management needs the active participation of patients. Chronic diseases, however, by nature are long-term and carry a psychological burden for individuals aiming to continually manage their condition effectively3,4. Technological advancements in mobile Health (mHealth), healthcare and public health practice supported by mobile devices and websites can help streamline care and provide resources to reduce disease burden. Over 2.5 billion people own a mobile device worldwide highlighting the huge potential for mHealth to facilitate access to effective care5.
mHealth apps have the potential to be powerful platforms for positive behavioural change; both for individuals independently monitoring their health (e.g. smart watches) and for encouraging effective management of chronic disease through clinically established prevention and treatment strategies6. App function can range from symptom and medication diaries, educational resources, to the gamification of self-management7. mHealth apps also have the potential to provide early-warnings of increased risk of poor outcomes from chronic diseases (i.e., Just-in-Time Adaptive Interventions (JITAI)) by making use of clinical data in tandem with self-reported data captured in-app8,9. JITAIs can leverage the data collected in these apps to increase the personalisation of care and ensure treatment is provided in a timely manner through the provision of models involving machine learning (ML)10. As with clinical treatment, the effectiveness of mHealth is reliant on the continued engagement of users11. The safety of machine learning models designed to provide early risk warnings depend on data that has sufficient predictive value and quality. Depending on self-reported data raises the concern that data collection introduces additional self-management treatment burden for users. In the design of mHealth apps, there is a clear need to balance the benefit of prediction against the treatment burden of in self-reporting.
We focussed on a widely used and clinically validated mHealth app; myCOPD12,13,14 which is designed for the self-management of Chronic Obstructive Pulmonary Disease (COPD). COPD is a common, costly, and incurable respiratory disease predicted to be the third most common cause of death by 203015. A key characteristic of managing COPD is mitigating the risk of âexacerbationsâ, defined by an acute worsening of a patientâs condition requiring a change in medication or emergency assistance16. myCOPD is provided to users diagnosed with COPD by clinicians as an explicit and agreed part of their long-term management plan.
The purpose of this research is to explore how user engagement with mHealth apps impacts predictive machine learning using self-reported data, and to discuss implications for balancing safety and treatment burden in mHealth design and engineering. To achieve this, we classified how myCOPD users engaged with the app around an exacerbation and quantified how engagement and data quality impacts the performance and safety of a ML model predicting risk to health (i.e., exacerbations). We supported this with focus group discussions and semi-structured interviews with myCOPD users to identify challenges facing digital approaches using predictive models and highlight factors leading to increased engagement and more insightful data.
Results
App usage and engagement
App usage was quantified by the fraction of days that the user was active (i.e., registered a symptom score) out of the 70-days prior to an exacerbation. A 70-day window was chosen empirically to be long enough to define the userâs typical engagement with the app while still demonstrating trends linked to exacerbation. In myCOPD, a symptom score must be registered before accessing further app functionality (on the first opening per day). A registered symptom score therefore represents a 1-to-1 relationship with app use on a given day. Figure 1 (left) shows the distribution of app usage prior to the 727 registered exacerbations. App usage is divided into three groups: frequent users (green, Nâ=â438) who register app activity â¥66% of the possible days, intermediate users (orange, Nâ=â156) who use the app between 33% and 66% of possible days, and infrequent users (red, Nâ=â132) who are active <33% of possible days.
Reasons for engagement were explored in semi-structured interviews. Despite some participants noting limited use of the app, most found it helpful for logging their medication use and acting as a reminder to take medications regularly. Participants also noted that the app was a source of education around self-management, which motivated engagement.
âI used to take my medicine at all different times, and now I use it at the same time every day. And the breathing exercises and how to clear your chest and that, I didnât know any of that before I started using the app so thatâs been a great help.â [P7âmale]
A further motivator was the opportunity the app offers to monitor symptoms, which provides reassurance that they are not deteriorating.
âI do like to look back when Iâve done the COPD assessment test, am I getting worse, am I getting better, and the answer is usually âno, youâre just the sameâ. Itâs a bit of a comfort thing to have around.â [P4âmale]
This is also evidenced by in-app data with over 60% of in-app interactions being related to medication or symptom monitoring.
Self-reported data quality and transitional engagement
Figure 2 provides a schematic of user groups divided by engagement (as in Fig. 1) and self-reported data quality prior to an exacerbation. The size of each vertical segment is proportional to the size of the group. Engagement and data quality is characterised by self-reported symptom scores.
Frequent users provide self-reports with a range of data quality (i.e., use for predictive models). âReporting with Signalâ corresponds to users who show sufficient variability in their self-reports that the deterioration in condition is clear leading up to the exacerbation (i.e., gradually reporting higher scores). Conversely âFixed Reportingâ corresponds to users who register consistently low or high scores (i.e., only 1âs or 3âs) prior to the exacerbation. Similar proportions of reporting with signal are found for intermediate and infrequent user groups.
Intermediate and infrequent user groups can âtransitionâ to become more engaged closer to exacerbation. We find 21.8% of intermediate and 14.4% of infrequent users (classification based on 70 days prior) transition to increased engagement groups in the 21 days immediately prior to the exacerbation (i.e., âEngaged Near Exacerbationâ). We note that most infrequent users (69.7%) are âRetrospective Reportingâ a rescue pack, registering the medications in-app more than 10 days after the event and providing minimal self-reported symptom scores around the actual exacerbation.
Transitions in behaviour immediately before exacerbations were also reported in semi-structured interviews. Notably, participants reported increased app use when their symptoms were worse, as a way of refreshing their memory on self-management techniques such as breathing or relaxation exercises. This was also true of those who had more mild symptoms and had yet to experience an exacerbation, who believed they would use the app more when necessary.
âwhen I do get worse Iâll use it a lot more I think.â [P6-male]
Conversely, others instead said that they use the app less when they are particularly unwell, as they do not have the capacity to engage with it.
âif I need my salamol I donât even think about it. It doesnât, it doesnât even occur to me to write that down or record itâ [P1-male]
Despite several users becoming more engaged immediately prior to an exacerbation there is no strong evidence that this increased engagement remains short-term after the exacerbation. Figure 1 (right) shows the distribution of app use in a 70-day window post exacerbation. The shading represents the original groupings (i.e., in the 70-days prior) with the histogram being stacked so the overall area matches the left panel. We note a slight increase of infrequent users (16.7%) post exacerbation. For 9% of exacerbations there is either a notable gap in self-reports directly after exacerbation, and/or a registered symptom score of 4 (i.e., needed to seek emergency care) highlighting possible disengagement due to a deteriorated condition.
Usersâ confidence in recognising risk
A key theme from user interviews was a lack of confidence around exacerbations and how to identify one. Particularly, the difficultly to differentiate between an exacerbation, a heavy cold, a chest infection, or otherwise was discussed.
âif thatâs what an exacerbation is, i.e., itâs just a chest infection. Or does it mean that, I donât know, itâs difficulty breathing and you need to take the inhaler? So I donât know what it is noâ [P1-male]
This was especially true for those who also suffer from other health complications, such as asthma or bronchiectasis. Participants noted that sputum changes are not always a reliable indicator.
âI had two exacerbations, late last year, both hospitalised and I didnât have the normal triggers that youâd have with changes, like increased volume, coughing and things like thatâ
A key barrier identified was a lack of explanation from health care professionals (HCPs), with most asserting that they had never had it explained to them.
âThat is all you hear is an exacerbation. Youâre not actually told what it is. Well, they havenât in my circumstances. Yes, it, you know, the nurses said âOhh, itâs an exacerbationâ but it doesnât explain what it actually is.â [P3âfemale]
Moreover, issues accessing HCPs means that myCOPD users had minimal opportunities to clarify or ask questions. Issues accessing HCPs also led to hesitancy about medication adherence (Supplementary Note 4).
âtrying to contact your GP is, well I canât think of a similarity but I could probably get in contact with Madonna better or more easilyâ [P4âmale]
Confidence in identifying risk was also reflected in self-reported data. Figure 3 compares self-reported symptom scores and salbutamol use for those registering their first exacerbation in-app relative to those reporting exacerbations having experienced one before (i.e., âSubsequentâ). Those registering their first exacerbation consistently report lower average symptom scores (top panel; chi-square statistic=726.9, Pâ=â3.05âÃâ10â157) demonstrating that users with previous experience of an exacerbation are more likely to be aware of their symptoms and report them in future events. As users increase confidence in recognising their symptoms, they also engage more frequently with the app in the longer-term (bottom panel). Users experiencing their first exacerbation also typically report lower salbutamol usage (middle panel). Salbutamol (classified as a SABA) is commonly used for immediate relief of symptoms including coughing, wheezing, and breathlessness. Increased usage reflects that the individual has experienced more breathlessness through a given day and may be indicative of a more acute condition. Regardless of experience, peak salbutamol use occurs on the first day of the exacerbation whereas average symptom scores peak days later. This indicates users self-report a deterioration through medication before typically self-recognising the deterioration in symptom scores.
Higher engagement for more experienced users is also found when considering the proportion of frequent, intermediate, and infrequent users by GOLD group (Supplementary Fig. 7). The proportion of users with a history of exacerbations (C and D) increases with engagement, reflecting that users are more likely to engage as their condition becomes more of a burden to self-manage and confidence to identify risk increases with experience of previous exacerbations.
Disease acuity and engagement
Figure 4 shows the proportion of frequent, intermediate, and infrequent users by GOLD group. The GOLD 2022 guidelines use a combined COPD assessment approach to group patients according to exacerbation history and symptoms (Fig. 6B). Overall, the majority of users reporting exacerbations are in higher risk groups, predominately represented by group D. The proportion of users with a history of exacerbations (C and D) increases with engagement, reflecting that users are more likely to engage as their condition becomes more of a burden to self-manage and confidence to identify risk increases with experience of previous exacerbations.
How engagement impacts machine learning
Figure 5 compares the performance of our XGBoost model predicting exacerbation up-to three-days in advance. Model performance, measured by AUROC and AUPR, has been computed from the hold-out test sets of simulated exacerbations for each of the following user groups (darker shaded in Fig. 2): Frequent, Intermediate (Consistent), Intermediate (Engaged Near Exacerbation), Infrequent (Consistent), Infrequent (Engaged Near Exacerbation). Predictions are made daily per user (from 55 days before to 70 days post exacerbation) and exacerbation is the positive class. Performance should only be used for contrastive purposes due to simulation of self-reported symptom scores (see Methods).
Both AUROC and average precision improve with 70-day engagement (i.e., infrequent to frequent), however, for transitional users (Engaged Near Exacerbation) the drop in performance relative to frequent users is minimal. This demonstrates that transitional engagement is more important for the safety of ML models than increasing overall engagement (i.e., regardless of current condition).
Discussion
In this study, we classify user engagement with a self-management app, identify barriers to engagement through interviews and focus groups, and directly quantify how this could impact the safety of a ML model. 60.3% of users engage and self-report information frequently. As predicted by the literature17, perceived usefulness is a good indicator of app usage and adherence to self-reporting. Figure 3 demonstrates that a history of exacerbations lead to markedly higher app engagement as disease burden comes to the forefront in the individualâs day-to-day life. This further demonstrated in Fig. 4 with individuals in higher risk categories typically showing more engagement. This is more nuanced than user monitoring symptoms, instead exploiting the affordances of the app to manage medication. Despite this, adherence is not consistent. App usage increases immediately around exacerbations (Fig. 3), despite users highlighting a lack of clarity about what exacerbation means.
For predictive ML models, it is critical that users report frequently and accurately. However, the human-AI partnership must be carefully balanced to achieve useful data while not subjecting the user to burdensome self-reporting behaviours and possible desensitisation18. Figure 5 clearly demonstrates that transitional behaviour (i.e., becoming more engaged around exacerbations) should be a more important focus than increasing overall engagement. Encouraging users to engage when they are starting to feel unwell will not only benefit the ML model, but potentially help the user see the long-term utility of the app and reduce treatment burden.
The early identification of exacerbations is key for COPD patients to prevent or treat them with medication19. Figure 3 shows a âtime-lagâ where use of reliever medications begins to increase before the self-reported symptom scores. Self-recognition of a decline in health appears to be delayed restricting algorithmic ability to make timely predictions from self-reported scores alone. This suggests apps should prompt users to review and report their condition when an uptick of reliever medications is identified. Despite this, COPD patients report challenges in recognising what an exacerbation specifically is. Research suggests that only around 60% of exacerbations are reported to healthcare professionals, suggesting that they often go unidentified and possibly untreated20. This was supported by our qualitative analysis with users uncertain of what to expect during an exacerbation or misinterpreting symptoms as indicators of other, related conditions. Despite this, users should be encouraged to report abrupt changes in condition, even if they do not understand the specifics. Users noted further concerns and hesitancies around taking medications in interviews (see Supplementary Note 4). Inability to discuss side-effects and medication purpose with HCPs may discourage users to adhere to prescribed treatment.
Strengths of this study include the volume of data collected by myCOPD over five years; identification of a clean sample of exacerbations from rescue-pack medication usage; combination of qualitative and quantitative approaches to articulate the balance between human orientated goals and data requirements of ML models. Limitations of this study include inclusion criteria which likely target more engaged users overall than average (i.e., those who report medications and engage with interviews); however, multiple engagement groups demonstrate diversity in the cohort. Another limitation includes the reliability of the self-reported symptom scores. As shown in Fig. 2, a significant fraction of users report minimal variability in their scores, the impact of which is not considered in our predictive models. Models predicting short-term exacerbations would benefit from a variety of self-reported data including activity and physiological measures from smart devices, oxygen therapy and dietary information.
Our research has implications for the design and engineering of mHealth apps, along with how the public should be encouraged to use it. It is critical that the developers of mHealth apps validate predictive models and JITAIs for different levels of engagement to determine safe conditions for its usage. For ML models predicting short-term risk, users of mHealth should be particularly encouraged to engage when they notice changes in their condition. This likely provides the most predictive data for ML models to maximise digital safety while minimising treatment burden on the individual and risk of disengagement through desensitisation.
Methods
Self-reported in-app data
We retrospectively evaluated self-reported data from users of myCOPD between January 1st, 2017, and October 3rd, 2022. All users of myCOPD are clinically diagnosed with COPD, with usage limited to patients âprescribedâ the app by clinicians as part of agreed care plans. myCOPD facilitates self-management of COPD through providing educational content, pulmonary rehabilitation, localised weather/pollution levels, and digital diaries for users to keep track of medications, symptoms, and exacerbations. Further information on myCOPD and data collection can be found in the Supplementary Material Note 1. Self-reported information included:
-
Daily self-assessed symptom scores prompted on app opening ranked on a 4-point scale (Fig. 6A). Symptom score reporting represents a simplistic but high completeness data source (relative to other data collected in myCOPD). To make short-term predictions of risk, it is critical to include data which is updated frequently.
-
COPD Assessment Tests (CAT): a validated instrument quantifying the long-term disease burden of COPD21,22. Evaluated approximately monthly, the CAT is an eight-question assessment indicating the impact of COPD on a userâs overall health23.
-
Prescribed and reliever medications taken for the treatment of COPD. This included routine medications (e.g., Muscarinic-Antagonists, Long-Acting Beta-Agonists, Inhaled Steroids), along with reliever medications (e.g., Short-Acting Beta-Agonists (SABAs), Rescue Packs) taken as an immediate response to a self-identified worsening of condition.
-
Exacerbation history reported annually. Along with CAT this is used to compute long-term acuity of condition as defined by the Global Initiative for Chronic Obstructive Lung Disease (GOLD) criteria24. The GOLD 2022 guidelines use a combined COPD assessment approach to group patients according to exacerbation history and symptoms (Fig. 6B).
Users also provided basic demographic (e.g., age, sex, postcode) and lifestyle information (e.g., smoking status) along with other clinically validated assessment scores (Modified Medical Research Council Dyspnoea scale). To investigate app usage around exacerbations we identified users who registered the use of a âRescue Packâ in their medication diaries (i.e., short course of oral steroids (Prednisolone) and antibiotics (e.g., Amoxicillin, Doxycycline) taken as a response to deteriorating symptoms as part of their acute exacerbation plan25). We did not include longer courses (â>â10 days) to avoid including weaning/maintenance prescriptions. This resulted in 727 exacerbations by 243 unique users (Age: μâ=â68.8, Ïâ=â8.3; Sex: 60.7% Male, 39.3% Female) who were registered throughout the study period. Figure 7 shows the distribution of exacerbations across the total cohort. Our selection criteria are strict to ensure we are selecting a clean sample of exacerbations with a well-defined start date, however, naturally selects users with higher disease acuity (i.e., having been prescribed a home-use Rescue Pack). We have quantified this difference in acuity in Supplementary Fig. 1. Despite this we find the selected cohort exhibit similar characteristics to the overall userbase (e.g., Age of all users: μâ=â68.4, Ïâ=â10.9).
The study received ethics approval from the University of Southamptonâs Faculty of Engineering and Physical Science Research Ethics Committee (ERGO/FEPS/66535) and was reviewed by the University of Southampton Data Protection Impact Assessment panel, with the decision to support the research.
Predicting exacerbations with machine learning
Our ML model is derived from previous work outlined in Chmiel et al. (2022) where a gradient-boosted decision-tree algorithm (XGBoost26) predicted exacerbations up-to three days in advance from self-reported data in myCOPD. Gradient boosted trees are examples of boosting algorithms which aim to combine an ensemble of weak learners in order to decrease bias whilst preserving or lowering variance in the prediction error, making them typically more desirable over other tree-based algorithms. A three-day window was chosen, based on clinical guidance to enable pre-emptive actions for the user while ensuring the exacerbation could be reliably predicted. Here, we aimed to quantify how app engagement and self-reported data quality impacted the performance of this algorithmic approach.
We stratified each user group by engagement and reporting quality, and then validated the performance of a model using only features relating to self-assessed symptom scores (Table 1). In Chmiel et al. (2022) predictive features were generated from a range of sources (e.g., symptom scores, CAT, demographics), however, in this study we only use symptom scores to ensure differences in model performance are related to engagement and reporting differences. An importance plot of the chosen features is provided in Supplementary Fig. 3.
Variances in performance also result from under/over representation of user groups in the training set (i.e., bias). To normalise, we simulated 1000 exacerbations for each user group creating empirical distributions fit to average symptom score values and frequencies from reporting prior to the real exacerbations of users in this study. Each simulated exacerbation was then sampled directly from the empirical distributions for reporting frequency and score to create complete series from 70 days prior to 70 days post exacerbation. Data was split 75â25 into train and hold-out test sets split at the series level (i.e., a simulated exacerbation appears exclusively in train or test). A binary prediction (of exacerbation in the next three-days) was generated for each day from 55 days prior (ensuring 15-day features are complete) to 70 days post exacerbation. Predictions made during the exacerbation were excluded. The XGBoost model was trained by 5-fold and grouped cross-validation grouped at the series level. Model hyperparameters were found using out-of-fold validation samples by Bayesian optimisation via the Tree-Structured Parzen Estimator (Optuna27). The best model hyperparameters can be found in Supplementary Table 1.
Model performance was estimated through area-under the receiving-operator characteristic curve (AUROC) and area-under the precision-recall curve (AUPR). Due to class imbalance (positive class fraction: 0.04) average precision is considered the key performance metric. We note our approach is designed to contrast impact to model performance from user engagement only and AUROC/AUPR scores are not indicative of performance in practice. We also perform this analysis using a logistic regression model (Supplementary Note 3) to justify the selection of XGBoost and confirm the differential performance trends are consistent.
Qualitative data
Qualitative data were collected to triangulate with the quantitative data during the analysis stage. A mixed-methods design was chosen to understand both the impact of engagement with myCOPD on the machine learning models, and also the subsequent experience of app users receiving the risk prediction. Qualitative exploration of subjective experiences of app users aimed to offer explanations for engagement data, providing context for the objective quantitative data which evidences self-reported experiences at scale.
Qualitative data were obtained through semi-structured interviews (Nâ=â7) and focus groups (Nâ=â8) with myCOPD users held online (via phone/video call) in 2022, recruited and consented through myCOPD (Fig. 8). Those recruited were advised to contact the research team via email to discuss participation, directed to an online consent form, and entered their contact details on an online platform (hosted on Qualtrics) containing the participant information sheet. Participants were each paid £25 for their time. Individuals were eligible to participate if they were (1) aged 18 and above and (2) had a diagnosis of COPD. There were no other exclusion criteria.
Sample size was determined using an information power approach, whereby the level of information provided by these 15 participants was sufficiently detailed and rich to address the research questions, particularly given the specific study population and aims. Interviews and focus groups were conducted by an experienced qualitative researcher, using a topic guide developed by the researchers and study stakeholders to address the study aims. Questions focused on participantsâ experience of using myCOPD, their understanding of exacerbations, and how they may perceive getting information regarding exacerbation risk generated from machine learning. B.C. transcribed recordings of the interviews/groups as the first stage of analysis. Thematic analysis was performed on the data in accordance with the six steps outlined by Braun and Clarke28. Transcripts were coded inductively by B.C., and the codes were developed into themes to present shared meaning within the data. Codes and themes were discussed with the research team (B.A., B.P.), who independently checked transcripts to ensure that the themes were representative of the data. The qualitative data collection received ethical approval from the University of Bath Psychology Research Ethics Committee [ref 22â041].
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Data availability
Aggregated data will be made available upon reasonable request to persons with a university affiliation. Requestors will need appropriate data protection, governance, and ethical review in place. Please contact C.J.Duckworth@soton.ac.uk for quantitative data enquiries and b.cliffe@westminster.ac.uk for qualitative data enquiries.
Code availability
Quantitative analysis and modelling was performed in Python v3.8.12 and made use of the following packages: numpy, pandas, sklearn, optuna, XGBoost, matplotlib. Code will be made available upon reasonable request. Please contact C.J.Duckworth@soton.ac.uk.
References
Murphy S. L., et al. Deaths: final data for 2018. (2021).
Kim, T. K. & Lane, S. R. Government health expenditure and public health outcomes: a comparative study among 17 countries and implications for US health care reform. Am. Int. J. Contemp. Res. 3, 8â13 (2013).
De Ridder, D., Geenen, R., Kuijer, R. & van Middendorp, H. Psychological adjustment to chronic disease. Lancet 372, 246â55 (2008).
Turner, J. & Kelly, B. Emotional dimensions of chronic disease. West. J. Med. 172, 124 (2000).
Organization W. H. Global diffusion of eHealth: making universal health coverage achievable: report of the third global survey on eHealth: World Health Organization; 2017.
Rowland, S. P. et al. What is the clinical value of mHealth for patients? NPJ Digital Med. 3, 1â6 (2020).
Perski, O., Blandford, A., West, R. & Michie, S. Conceptualising engagement with digital behaviour change interventions: a systematic review using principles from critical interpretive synthesis. Transl. Behav. Med. 7, 254â67 (2017).
Nahum-Shani, I. et al. Just-in-time adaptive interventions (JITAIs) in mobile health: key components and design principles for ongoing health behavior support. Ann. Behav. Med. 52, 446â62 (2018).
Wang, L. & Miller, L. C. Just-in-the-moment adaptive interventions (JITAI): a meta-analytical review. Health Commun. 35, 1531â44 (2020).
Chmiel, F. P. et al. Prediction of chronic obstructive pulmonary disease exacerbation events by using patient self-reported data in a digital health app: statistical evaluation and machine learning approach. JMIR Med. Inform. 10, e26499 (2022).
Miller, S. et al. A framework for analyzing and measuring usage and engagement data (AMUsED) in digital interventions. J. Med. Internet Res. 21, e10966 (2019).
North, M. et al. A randomised controlled feasibility trial of E-health application supported care vs usual care after exacerbation of COPD: the RESCUE trial. NPJ Digital Med. 3, 1â8 (2020).
Crooks M. G., et al. Evidence generation for the clinical impact of myCOPD in patients with mild, moderate and newly diagnosed COPD: a randomised controlled trial. ERJ Open Res. 6 (2020).
Cooper, R. et al. Evaluation of myCOPD digital self-management technology in a remote and rural population: real-world feasibility study. JMIR mHealth uHealth 10, e30782 (2022).
McLean, S. et al. Projecting the COPD population and costs in England and Scotland: 2011 to 2030. Sci. Rep. 6, 1â10 (2016).
Rodriguez-Roisin, R. Toward a consensus definition for COPD exacerbations. Chest 117, 398Sâ401S (2000).
MaranguniÄ, N. & GraniÄ, A. Technology acceptance model: a literature review from 1986 to 2013. Univers. Access Inf. Soc. 14, 81â95 (2015).
Ratneswaran, C. et al. A cross-sectional survey investigating the desensitisation of graphic health warning labels and their impact on smokers, non-smokers and patients with COPD in a London cohort. BMJ Open 4, e004782 (2014).
Walters J. A., Turnock A. C., Walters E. H., WoodâBaker R. Action plans with limited patient education only for exacerbations of chronic obstructive pulmonary disease. Cochrane Database of Systematic Reviews (2010).
Wilkinson, T. M. et al. Early therapy improves outcomes of exacerbations of chronic obstructive pulmonary disease. Am. J. Resp. Crit. Care Med. 169, 1298â303 (2004).
Dodd, J. W. et al. The COPD assessment test (CAT): response to pulmonary rehabilitation. A multicentre, prospective study. Thorax 66, 425â29 (2011).
Gupta, N., Pinto, L. M., Morogan, A. & Bourbeau, J. The COPD assessment test: a systematic review. Eur. Resp. J. 44, 873â84 (2014).
Jones, P. et al. Development and first validation of the COPD assessment test. Eur. Resp. J. 34, 648â54 (2009).
GOLD. Global strategy for the diagnosis, management and prevention of COPD. Global Initiative for Chronic Obstructive Lung Disease (GOLD). https://goldcopd.org/.
Hopkinson N. S., Molyneux A., Pink J., Harrisingh M. C. Chronic obstructive pulmonary disease: diagnosis and management: summary of updated NICE guidance. Bmj 366 (2019).
Chen T., Guestrin C. Xgboost: a scalable tree boosting system. Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining:785â94 (2016).
Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining.
Braun, V. & Clarke, V. Reflecting on reflexive thematic analysis. Qualitative Res. Sport Exerc. Health 11, 589â97 (2019).
Acknowledgements
This project âmy Smart COPD exacerbation management (mySmartCOPD)â is funded by the National Institute for Health Research (NIHR) Artificial Intelligence (AI) in Health and Care Award AI_AWARD02200. The views expressed are those of the author(s) and not necessarily those of the NIHR or the UK Governmentâs Department of Health and Social Care. The funders of the study had no role in study design, data analysis, interpretation, or writing.
Author information
Authors and Affiliations
Contributions
C.D., B.C., M.J.B., B.A., and T.M.A.W. conceived the research question. C.D. performed the quantitative analysis of in-app data and modelling with support from M.J.B. B.C. performed the qualitative analysis with support from B.P. and B.A. C.D. and B.C. wrote the first draft of the manuscript. All other authors contributed to the first and future iterations of the manuscript. All authors had access to all data, with C.D. and M.J.B. verifying the quantitative data and B.C., B.A., and B.P. verifying the qualitative data. B.P. obtained ethical and governance approvals. A.B. and A.K. managed the data extraction at my mHealth. T.M.A.W., A.B., and A.K. provided clinical insight.
Corresponding author
Ethics declarations
Competing interests
All authors were supported by the National Institute for Health Research (NIHR). TMAW is Chief Science Officer and cofounder of my mHealth, the developer of the myCOPD app. A.B. is a Senior Research Nurse and Clinical Trial Manager at my mHealth. A.K. is the Medical Director and Data Protection Officer at my mHealth. All other authors declare no competing interests.
Additional information
Publisherâs note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the articleâs Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the articleâs Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Duckworth, C., Cliffe, B., Pickering, B. et al. Characterising user engagement with mHealth for chronic disease self-management and impact on machine learning performance. npj Digit. Med. 7, 66 (2024). https://doi.org/10.1038/s41746-024-01063-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41746-024-01063-2