Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 26 July 2016
Sec. Public Mental Health
This article is part of the Research Topic Computers and games for mental health and well-being View all 35 articles

Development and Clinical Evaluation of an mHealth Application for Stress Management

\r\nBrent D. Winslow*Brent D. Winslow1*George L. ChadderdonGeorge L. Chadderdon1Sara J. DechmerowskiSara J. Dechmerowski1David L. JonesDavid L. Jones2Solomon KalksteinSolomon Kalkstein3Jennifer L. GreeneJennifer L. Greene3Philip Gehrman\r\nPhilip Gehrman3
  • 1Design Interactive, Inc., Orlando, FL, USA
  • 2Quantified Design Solutions LLC, Orlando, FL, USA
  • 3Philadelphia VA Medical Center, University of Pennsylvania, Philadelphia, PA, USA

A large number of individuals experience mental health disorders, with cognitive behavioral therapy (CBT) emerging as a standard practice for reduction in psychiatric symptoms, including stress, anger, anxiety, and depression. However, CBT is associated with significant patient dropout and lacks the means to provide objective data regarding a patient’s experience and symptoms between sessions. Emerging wearables and mobile health (mHealth) applications represent an approach that may provide objective data to the patient and provider between CBT sessions. Here, we describe the development of a classifier of real-time physiological stress in a healthy population (n = 35) and apply it in a controlled clinical evaluation for armed forces veterans undergoing CBT for stress and anger management (n = 16). Using cardiovascular and electrodermal inputs from a wearable device, the classifier was able to detect physiological stress in a non-clinical sample with accuracy greater than 90%. In a small clinical sample, patients who used the classifier and an associated mHealth application were less likely to discontinue therapy (p = 0.016, d = 1.34) and significantly improved on measures of stress (p = 0.032, d = 1.61), anxiety (p = 0.050, d = 1.26), and anger (p = 0.046, d = 1.41) compared to controls undergoing CBT alone. Given the large number of individuals that experience mental health disorders and the unmet need for treatment, especially in developing nations, such mHealth approaches have the potential to provide or augment treatment at low cost in the absence of in-person care.

Introduction

Recent estimates suggest that approximately one-third of individuals globally will experience mental health disorders in their lifetime (1). Individuals in developing nations are particularly vulnerable (2, 3). In developed nations, an especially vulnerable population is military veterans. For example, approximately one-third of US military veterans suffer from some type of psychological distress, including post-traumatic stress disorder (PTSD), major depressive disorder (MDD), and suicidal ideation (4). Diagnostic and subthreshold levels of PTSD are associated with poor quality of life, including anger, stress, alcoholism, depression, poor physical health, and increased suicidality (5, 6), and cause impaired ability to function in social, educational, and work environments (7).

Among various interventions to treat depression and anxiety, cognitive behavioral therapy (CBT) has emerged as standard practice for reduction of psychiatric symptoms (8), with previous studies indicating that CBT has similar therapeutic effects as anti-depressant medication (9). CBT is generally administered by mental health professionals and consists of a structured, collaborative process that helps individuals consider and alter their thought processes and behaviors associated with stress or anxiety, usually administered weekly over several months (10). However, standard CBT for stress and anxiety does not offer the provider information regarding therapeutic efficacy outside of office visits nor does it provide objective information about individuals’ triggers, such as location, time, or severity (11). In addition, high dropout rates from CBT programs have been reported to span from 25% (12) to as high as 40% (13) for individuals suffering from depression. The limitations of CBT, including lack of objective data available for providers and high patient dropout rates, could be mitigated with emerging technologies. To support real-time objective stress monitoring in mental health treatment, wearable physiological sensors and associated mobile health (mHealth) applications (14) have the potential to quantify biological metrics associated with stress (15), support remote monitoring, and alert the wearer or provider to real-time changes in emotional state.

Existing approaches to stress detection use a wide array of features calculated from sensor data measuring various aspects of heartbeat, including pulse photoplethysmography (PPG) or ECG (1517), skin conductance measurement (1820), and measurement of respiration, all of which are responsive to increased sympathetic nervous system activity associated with stress (21). Standard supervised machine learning methods have been used previously to develop stress classifiers, which require subjects to engage in tasks known to induce stress so that stress or non-stress labels can be assigned to the input features. Previous work has emphasized the difficulties imposed on stress classification by individual subject variability in physiological responses to stress (16, 19). Another concern is the physical activity of subjects that triggers similar cardiovascular and electrodermal physiological signals as stress, leading to masking and confounds of stress detection (15, 19). The major challenge in using mobile physiological sensors to quantify stress is the lack of robust and clinically tested algorithms to classify stress in a mobile environment in real time (22).

Previous stress monitoring algorithms have been built with traditional laboratory physiological sensor suites that do not translate well to operational settings (17, 19), such as mental health treatment. New wearable devices with clinical grade sensors and associated mobile applications have the potential to take real-time stress monitoring outside of the laboratory. There is an opportunity to combine foundational mobile stress monitoring algorithm research methods with new mobile physiological sensor suites to create an accurate, quantitative classifier for continuous and objective real-time stress assessment. In the current study, we develop a physiological classifier of stress and apply it in a clinical evaluation of patients undergoing CBT for stress/anger management, following military deployment. It is hypothesized that stress induced using standard methods can be classified with high accuracy using a machine learning algorithm and that the use of such an algorithm in an mHealth application can reduce post-deployment psychiatric symptoms, including stress, anger, and anxiety, in a clinical population undergoing CBT.

Materials and Methods

Classifier Study

Participants

All methods involving participants were approved by a series of Institutional Review Boards [Copernicus Group IRB, Durham, NC, USA; Human Studies Subcommittee, Department of Veteran’s Affairs, Philadelphia, PA, USA; US Army Medical Research and Material Command Human Research Protection Office (HRPO), Fort Detrick, MD, USA].

Thirty-five participants (24 males; average age 25.7 ± 6.2 years) were recruited for the initial classifier-development study, which lasted approximately 1.5 h. Participants were recruited using recruitment flyers posted online and through recruitment fairs at local universities.

Experimental Procedure

Upon arrival, participants provided written-informed consent and completed a series of questionnaires including: demographics; the subjective units of distress scale (SUDS); the depression, anxiety, and stress scale (DASS); and the patient-reported outcomes measurement information scale (PROMIS) anger scale. Wireless physiological sensors were then placed on the participants, followed by a 5-min recording of baseline physiological activity, while participants remained seated. Participants then completed the Trier Social Stress Test [TSST; Ref. (23)]. The TSST was used to elicit physiological stress, consisting of 5 min each of: preparatory anticipatory stress (TSST-P); oral speech (TSST-S); and mental arithmetic (TSST-A). Following data acquisition, participants were debriefed and thanked for their participation. A subset of participants (n = 7) also provided a saliva sample via passive drool for cortisol assessment at baseline and following the TSST.

Qualitative Measures

Participants in the classifier study responded to the SUDS, in which they reported their current level of stress on a scale of 0–100, with 0 indicating that they were completely relaxed and 100 indicating that they were experiencing severe stress (24). Participants then completed the DASS, designed to assess current depression, anxiety, and stress using responses to 21 statements (25). Respondents indicated the degree to which each statement has been true for them over the preceding week on a 4-point Likert scale. Participants also responded to the PROMIS anger scale, which consists of an 8-item measure on which respondents indicate the frequency of each item from the past week on a 5-point Likert scale (26).

Physiological Measures

The Biopac MP-150 system (Goleta, CA, USA) was used for collection of physiological data. Participants were fitted with PPG at the non-dominant thumb and electrodermal activity (EDA) on the fourth and fifth fingers of the non-dominant hand, with band limits set between DC and 10 Hz. All physiological data were sampled at 1000 Hz and wirelessly sent to an MP-150 system running AcqKnowledge software (Biopac Systems, Goleta, CA, USA). Following data collection, PPG data were downsampled to 64 Hz and EDA was downsampled to 4 Hz. Heart rate (HR) was calculated from the R–R interval from the PPG signal, with intervals <40 and >180 bpm excluded from the analysis. Salivary cortisol was measured by standard ELISA (Salimetrics, Carlsbad, CA, USA; intra-assay CV = 4.5%, inter-assay CV = 5.8%).

In a subset of participants (n = 8), the Empatica E3 sensor was also used for physiological data collection. A second system was used to ensure the stress algorithm was compatible across multiple hardware solutions and to provide for mobile stress classification in future studies. Physiological data, consisting of PPG (64 Hz) and EDA (4 Hz), were transmitted via Bluetooth 4.0 to a custom mobile application for data collection implemented in the Android OS on a Samsung Galaxy S4 phone.

Classifier Development

Event times and physiological data were stored in Biopac.acq files. All data were read into Python analysis scripts running under the Enthought Canopy environment. The numpy, scipy, pandas, and matplotlib libraries were used for feature extraction and data analysis (27), and the scikit-learn library (28, 29) was used for classifier development. Visual inspection of the raw data in the Biopac software and the interactive Python environment was used to discard physiologically noisy or missing participant data.

From the raw data, non-overlapping 1-min windows were analyzed to yield feature vectors for the minute blocks. Inter-beat intervals (IBI) were extracted from the PPG data using a signal derivative-based algorithm (30). For minutes with less than 40 valid IBI samples, the block of data was discarded; for remaining blocks, the mean IBI was calculated. For each valid IBI block, the mean HR was estimated by dividing 60 by the IBI mean. For the EDA data, the mean was taken over the minute’s raw data. The HR and EDA means were then normalized separately for each participant by subtracting the average of the 5-min baseline.

Matplotlib boxplots and scatterplots were used to explore the distributions of the task-specific patterns (e.g., baseline and TSST-S) in feature space. A stress vs. non-stress classifier was trained using baseline vs. TSST-S, using baseline-normalized mean HR and EDA features. A 2-feature linear model classifier was trained and tested on the E3 dataset using stochastic gradient descent. The train:test (75:25%) set consisted of data feature vectors taken from the baseline and TSST-S minutes of the participants with E3 data. Five-fold cross-validation was implemented to evaluate the average performance of the algorithm on the train set: this set was divided into fifths and, iteratively, one of the five blocks was left out for testing and the other four were used to train the classifier. While the cross-validation measures were used to compare performance of different learning algorithms (e.g., stochastic gradient descent vs. support vector machines), the training and testing accuracies were calculated from performance from training on the full train set. Performance of the E3 data-trained model was also measured using the entire Biopac data set with good HR and EDA baseline and TSST-S minute data as a test set.

Data Analysis and Statistics

Classifier training accuracy was defined by signal detection theory (31) (hit = classifier correctly identified spike stress state; miss = classifier missed a spike in stress state; false alarm = classifier identified a spike in stress state when one did not actually occur; and correct rejection = classifier did not identify a spike in stress state when one did not occur). Accuracy was defined as the ratio of the sum of hits and correct rejections over the sum of minute blocks classified as either stressed or non-stressed. The hit rate was defined as the number of hit minute blocks divided by the total number true stress blocks (TSST-S), and the false-alarm rate was defined as the number of false-alarm blocks divided by the total number of true non-stress blocks (baseline). SUDS scores between baseline and the TSST were analyzed using paired samples t-tests.

Clinical Evaluation

Participants

Following development and evaluation of the stress classifier, 16 participants [13 males; average age 39.8 ± 10.5 (SD)] were enrolled for participation in the clinical evaluation study of the classifier and associated mHealth application, which lasted 8–10 weeks for each individual. Participants were recruited from patients at the Philadelphia VA Medical Center who reported significant difficulties with anger and/or stress and indicated a willingness to participate in a research study. Exclusion criteria consisted of: currently active duty military; moderate or severe TBI; severe mental impairment as assessed in their electronic medical record; and/or functional limitations preventing use of a mobile device.

Experimental Procedure

Following written-informed consent, participants were randomly assigned to the experimental or control group and completed the DASS, PROMIS-Anger, and PTSD Checklist-Military (PCL-M) questionnaires. The control group (n = 6) received standard CBT; the experimental group (n = 10) received standard CBT integrated with the stress classifier and mobile application (see Physiological Measures). Following initial assessment, an appointment was made to begin treatment within 1–2 weeks. All treatment was administered in an individual format by the study clinicians, who are licensed mental health professionals. The study clinicians were directed to use standard CBT treatment manuals (32) as a foundation for CBT while using clinical judgment to determine what content to cover in each session and how many sessions to schedule. Typical treatment following this approach was expected to last 8–10 weeks. This approach was chosen rather than utilizing a fixed protocol in order to represent routine clinical practice.

Sessions involved a weekly, in-person meeting, which lasted 60 min. Patients were asked to keep a log of daily activities, summarizing key stress/anger events that occurred, and present this written report to therapists during their session. Weekly sessions continued until: (a) the participant and clinician jointly determined that there was significant clinical improvement; (b) it was judged by the therapist that no further improvement was likely to occur; or (c) the participant discontinued therapy. Compliance in the experimental group was quantified by use of the mobile application. One month following the completion of treatment, participants were asked to return for a follow-up visit to complete the DASS, PROMIS Anger, and PCL-M questionnaires.

Qualitative Measures

Participants completed the DASS and PROMIS anger scale at their initial assessment and following therapy completion; the DASS-Stress and PROMIS anger scales were considered primary outcome metrics. Participants in the clinical study also completed the PCL-M, which is a 17-item continuous severity measure that corresponds to the 17 DSM-IV criteria for PTSD (33). Respondents indicated the extent to which they had experienced each symptom described in the past month using a 5-point scale, from 1 (not at all) to 5 (very often). The PCL-M was considered a secondary outcome metric in the clinical study.

Physiological Measures

An mHealth application and stress classifier were used for data collection in the clinical study. The mHealth application was implemented in Android on a Samsung Galaxy S4 phone and received data from the E3 band (Empatica, Milan, Italy), classified stress using the algorithm developed in the classifier study, alerted the user when stress was detected, and presented stress mitigation techniques to the user, such as breathing exercises. The E3 band sent PPG, EDA, temperature, and accelerometer information to the mobile application via Bluetooth 4.0. A web-based provider portal that resided on a secure cloud server was also implemented and allowed the provider to view physiological data for individual patients and enter reminders (e.g., complete your cognitive restructuring homework) or focus points (e.g., practice breathing), which were sent to the mobile application.

Data Analysis and Statistics

Non-parametric statistical analysis was used to compare within groups measures across the two timepoints (initial and final assessment) and consisted of Wilcoxon signed-ranks tests with significance set to 0.05. Between groups differences were assessed using Mann–Whitney U tests with significance set to 0.05. All statistical testing was done in SPSS software version 18.

Results

Classifier Study

The sociodemographic factors in the initial classifier study are listed in Table 1. The average age of the participants was 25.7 ± 6.2 (SD) years, and participants had an average of 3.4 ± 2.0 (SD) years of post-secondary education.

TABLE 1
www.frontiersin.org

Table 1. List of sociodemographic factors in the classifier study sample.

Subjective Units of Distress Scale scores are shown in Figure 1. As compared to baseline, the TSST elicited a significant increase in perceived stress (p < 0.001, d = 1.44). Baseline cortisol in the study subset was 0.39 ± 0.33 (SD) μg/dL, which did not differ following the TSST at a level of 0.44 ± 0.39 (SD) μg/dL (p = 0.29, d = 0.16) via paired samples t-test.

FIGURE 1
www.frontiersin.org

Figure 1. Self-reported distress at baseline and following the TSST. Mean + SD shown. *p < 0.001.

Depression, anxiety, and stress scale scores are shown in Table 2. Stress, depression, and anxiety scores were considered normal (25). The PROMIS anger score was 50.1 (7.2).

TABLE 2
www.frontiersin.org

Table 2. Mean (SD) DASS scores in the classifier-development group.

Noise or data loss affected 4/35 participants’ physiological data, which were removed from analysis. Each task phase in the experiment was regarded as having distinct ground truth values for whether the participant would be considered stressed or not stressed. The TSST-S and TSST-A phases of the TSST were considered to be psychological stress phases. The baseline resting task was considered to be a non-psychological stress phase. For the TSST-P task no assumption of stress vs. non-stress was made. Figure 2 shows the distributions of (non-normalized) HR (left) and skin conductance (right) data vs. task for all participants.

FIGURE 2
www.frontiersin.org

Figure 2. Task-dependent heart rate and skin conductance measures across participants. (Left) Includes heart-rate estimate distributions; TSST-S, TSST-A, have notably high HR distributions, whereas baseline tends to be low. (Right) includes electrodermal activity estimate distributions; the baseline conductance is relatively low, whereas TSST-S and TSST-A, are relatively high. Shown are group means ± SD.

Based on the distributions, baseline-normalized HR and EDA means were used for stress vs. non-stress classification. Figure 3 shows the classification results of training the stress vs. non-stress classifier using 75% of the E3 physiological data. For the E3 data, the training accuracy was 97.1%. Test set accuracy on the remaining 25% of the data were 91.7%. The hit rate on the test set was 100%, and the false-alarm rate was 12.5%. For the Biopac data, the testing accuracy was 95.1%, the hit rate was 89.1%, and the false alarm rate was 1.7%.

FIGURE 3
www.frontiersin.org

Figure 3. Stress vs. non-stress classifier using baseline-normalized HR and EDA features during the baseline and TSST-S segments. Stress classification using the E3-collected data is shown at left and with the Biopac-collected data shown at right. The decision boundary is shown as a line; data points to the left of this boundary were classified as non-stress.

Results from Clinical Assessment

The sociodemographic factors in the clinical evaluation are listed in Table 3. The average age of the participants was 39.7 ± 10.5 (SD) years and most were US Army veterans.

TABLE 3
www.frontiersin.org

Table 3. List of sociodemographic factors of study sample.

Nine individuals in the study dropped out prior to completion of therapy and follow-up visit. A Mann–Whitney test indicated that individuals in the experimental group completed a significantly greater number of therapy sessions (p = 0.016, d = 1.34) at an average of 7.2 ± 1.6 (SD) sessions as compared to 3.4 ± 2.4 (SD) in the control group. The remaining participants that completed the study included five in the experimental group and two in the control group. One participant in the experimental group that completed the study did not use the mHealth application but completed standard CBT and was reassigned to the control group.

Depression, anxiety, and stress scale scores are shown in Table 4. For the initial assessment, stress and depression for the participants was in the 96th percentile, and anxiety was in the 99th percentile as compared to a normative sample (34). Anxiety scores were considered extremely severe, while stress and depression scores were in the severe range (25). No differences between the control and experimental group were observed during the initial assessment for stress scores (p = 0.616, d = 0.21), depression (p = 0.964, d = 0.09), or anxiety (p = 0.682, d = 0.29) as assessed by Mann–Whitney testing.

TABLE 4
www.frontiersin.org

Table 4. Mean (SD) DASS assessment scores.

The follow-up assessment was completed by four participants in the experimental group and three participants in the control group. A significant reduction in stress (p = 0.032, d = 1.61) and anxiety (p = 0.050, d = 1.26) was observed between the experimental and the control group but not for depression (p = 0.719, d = 0.29) per Mann–Whitney testing. Within groups, the control group showed no significant changes in DASS scores for stress (p = 0.593), anxiety (p = 0.109), or depression (p = 1.000) between the initial assessment and follow-up as assessed by Wilcoxon signed-ranks tests. The experimental group trended toward a significant decrease in stress (p = 0.068) and depression (p = 0.068) between timepoints but not anxiety (p = 0.144) as assessed by Wilcoxon signed-rank tests.

Patient-reported outcomes measurement information scale anger scores are shown in Table 5. No differences between the control and experimental group were observed at the initial assessment (p = 0.703, d = 0.20), but follow-up scores indicated a significant reduction in anger for the experimental group (p = 0.046, d = 1.41) as assessed by Mann–Whitney testing. Within groups, no difference in anger was observed for the control group (p = 0.715) or the experimental group (p = 0.109) as assessed by Wilcoxon signed-ranks testing.

TABLE 5
www.frontiersin.org

Table 5. Mean (SD) PROMIS Anger scores.

PTSD checklist-military scores are shown in Table 6. No differences between the control and experimental group were observed at the initial assessment (p = 0.639, d = 0.16) or at follow-up (p = 0.480, d = 0.57) as assessed using Mann–Whitney testing. Within groups, no difference in PCL-M scores was observed for the control group (p = 0.285) or the experimental group (p = 0.144) as assessed by Wilcoxon signed-ranks testing.

TABLE 6
www.frontiersin.org

Table 6. Mean (SD) PCL-M Scores.

Discussion

The current series of studies shows the feasibility of creating an individualized, physiological classifier of stress with a high degree of accuracy compatible with different sensor suites. The use of such an algorithm in an mHealth application (35) may reduce symptoms of stress and anger in a small clinical population, increase the number of CBT sessions individuals will attend, and decrease their dropout rate. Given the large number of individuals that experience mental health disorders and the unmet need for treatment, especially in developing nations, such mobile approaches have the potential to provide or augment treatment in the absence of standard, in-person care (36). However, most commercially available apps targeting mental health remain untested (22, 37).

Classification of stress was based on features gathered from a large user group undergoing the TSST, which has one of the highest effect sizes for eliciting stress and associated cortisol responses in laboratory settings (38). Stress classification was based on cardiovascular and electrodermal inputs (3), which showed high variance due to individual differences (19), and was addressed by individual baseline normalization. The psychoendocrine reaction to life stressors, or stressors outside of the laboratory setting, such as bereavement, declining health, or flashbacks in PTSD, are likely of higher duration and intensity than laboratory stressors (39). Therefore, the algorithm and decision boundary developed using acute socio-evaluative stress in the current work may underperform for more severe stressors associated with MDD, PTSD, or other forms of mental health disorders. For instance, the DASS and PROMIS-anger scores from the classifier-development study sample indicated a relatively low burden of stress, depression, anxiety, and anger as opposed to a relatively high burden of mental health symptoms in the clinical evaluation study sample. In addition, veteran post-traumatic stress is often comorbid with depression, which has recently been shown to be associated with intensified anger (2). Anger has been acknowledged as the most prevalent veteran readjustment problem (14). Interestingly, the use of an mHealth application focused on stress identification and reduction in conjunction with CBT reduced metrics of anger and anxiety in addition to stress in a small clinical sample.

The overall dropout rate from CBT has been reported to be between 25–40% for depression (12, 13). In the current study, over 50% of the participants discontinued therapy, early. This higher dropout rate may be due to characteristics of the veteran population or the high burden of mental health symptoms in the study sample. For example, previous research has indicated that medication compliance among veterans is relatively low (40). The high dropout rate likely also reflects that many of the participants were experiencing periods of acute stress and were often preoccupied with these stressors. The availability of validated mHealth applications that individuals could use within the context of their daily lives would help to address this issue. Within the sample, those who used the mobile application and stress algorithm were more likely to complete the study and demonstrated reduced stress and anger as compared to the control group. This reduced stress may result from an increased awareness of their stressors due to the alerts provided through the mobile application to the user (41), or the use of the guided breathing exercises within the application (42).

Future research will include further accuracy refinement through reduction in environmental noise, and a method to learn individual user stress thresholds (19). Additional operational testing to reduce environmental noise is being conducted in order to determine the changes in classifier false alarms and misses when collecting data in different temperatures and while performing different physical activities (43), ranging from typing on a keyboard to walking or running. The 2-feature linear model trained with stochastic gradient descent employed in this effort has the advantage of including a bias term to tune the decision boundary threshold on the stress vs. non-stress classifier to allow adjusting the tradeoff between hit and false-alarm rates, ultimately generating an individualized threshold for each user.

The low sample size in the clinical evaluation and the high dropout rate represent a limitation in the current study. Even though there have been an estimated 180,000 cases of US military veterans with PTSD over the past two decades (44), many do not seek care (45). Additional challenges include long wait times experienced in the VA medical system (46), low participation rates in clinical studies (47), and a high dropout rate during CBT. Further data and objective outcome measures are needed to validate the observed reductions in stress, anger, and anxiety symptoms in the study sample.

The capability to classify individual physiological stress in a mobile environment has additional uses outside of veteran mental health treatment, including military or medical training (48). For example, training instructors could remotely and simultaneously monitor objective stress status for individual trainees during live training sessions and act on the information in real time (49). In addition, instructors could identify individual trainees that tend to have more intense stress responses than others during training scenarios to provide targeted coping and resilience training interventions (50). Beyond training, additional applications for this capability include stress research for laboratory and field settings, chronic disease monitoring for tracking outpatient health and long-term data capture to inform care, and objective, real-time user experience evaluations. mHealth applications and wearable physiological sensors have the potential to analyze and present meaningful data to better manage and optimize general health and specific health conditions. However, high quality, wearable devices and robust, validated algorithms remain a necessary component to realizing the potential of this technology.

Author Contributions

BW analyzed data and wrote the manuscript. GC and BW developed the algorithm for stress. SD and DJ developed the mobile application and managed the experiments. PG designed the clinical evaluation. SK and JG conducted the clinical evaluation.

Conflict of Interest Statement

GC, BW, SD, DJ (2014) System, method, and computer program for the real-time mobile evaluation of physiological stress. US Patent Pending #10878-023.

Acknowledgments

Portions of the classifier study were reported at the 49th Annual Meeting of the Interservice/Industry Training, Simulation & Education Conference (I/ITSEC), as SD, BW, GC, T. Schmidt-Daly, DJ (2014) Classifying stress in a mobile environment (14.195). This work was funded by the US Army Medical Research and Materiel Command (USAMRMC) grant W81XWH-12-C-0071.

References

1. Steel Z, Marnane C, Iranpour C, Chey T, Jackson JW, Patel V, et al. The global prevalence of common mental disorders: a systematic review and meta-analysis 1980–2013. Int J Epidemiol (2014) 43:476–93. doi:10.1093/ije/dyu038

CrossRef Full Text | Google Scholar

2. Wang PS, Aguilar-Gaxiola S, Alonso J, Angermeyer MC, Borges G, Bromet EJ, et al. Use of mental health services for anxiety, mood, and substance disorders in 17 countries in the WHO world mental health surveys. Lancet (2007) 370:841–50. doi:10.1016/S0140-6736(07)61414-7

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Whiteford HA, Degenhardt L, Rehm J, Baxter AJ, Ferrari AJ, Erskine HE, et al. Global burden of disease attributable to mental and substance use disorders: findings from the Global Burden of Disease Study 2010. Lancet (2013) 382:1575–86. doi:10.1016/S0140-6736(13)61611-6

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Thomas JL, Wilk JE, Riviere LA, Mcgurk D, Castro CA, Hoge CW. Prevalence of mental health problems and functional impairment among active component and National Guard soldiers 3 and 12 months following combat in Iraq. Arch Gen Psychiatry (2010) 67:614–23. doi:10.1001/archgenpsychiatry.2010.54

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Marshall RD, Olfson M, Hellman F, Blanco C, Guardino M, Struening EL. Comorbidity, impairment, and suicidality in subthreshold PTSD. Am J Psychiatry (2001) 158:1467–73. doi:10.1176/appi.ajp.158.9.1467

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Yarvis JS, Schiess L. Subthreshold posttraumatic stress disorder (PTSD) as a predictor of depression, alcohol use, and health problems in veterans. J Workplace Behav Health (2008) 23:395–424. doi:10.1080/15555240802547801

CrossRef Full Text | Google Scholar

7. Schumm JA, Chard KM. Alcohol and stress in the military. Alcohol Res (2012) 34:401.

PubMed Abstract | Google Scholar

8. Hollon SD, Stewart MO, Strunk D. Enduring effects for cognitive behavior therapy in the treatment of depression and anxiety. Annu Rev Psychol (2006) 57:285–315. doi:10.1146/annurev.psych.57.102904.190044

PubMed Abstract | CrossRef Full Text | Google Scholar

9. DeRubeis RJ, Hollon SD, Amsterdam JD, Shelton RC, Young PR, Salomon RM, et al. Cognitive therapy vs medications in the treatment of moderate to severe depression. Arch Gen Psychiatry (2005) 62:409–16. doi:10.1001/archpsyc.62.4.409

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Beck AT, Dozois DJ. Cognitive therapy: current status and future directions. Annu Rev Med (2011) 62:397–409. doi:10.1146/annurev-med-052209-100032

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Rachman S. Psychological treatment of anxiety: the evolution of behavior therapy and cognitive behavior therapy. Annu Rev Clin Psychol (2009) 5:97–119. doi:10.1146/annurev.clinpsy.032408.153635

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Hans E, Hiller W. Effectiveness of and dropout from outpatient cognitive behavioral therapy for adult unipolar depression: a meta-analysis of nonrandomized effectiveness studies. J Consult Clin Psychol (2013) 81:75. doi:10.1037/a0031080

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Fernandez E, Salem D, Swift JK, Ramtahal N. Meta-analysis of dropout from cognitive behavioral therapy: magnitude, timing, and moderators. J Consult Clin Psychol (2015) 83:1108. doi:10.1037/ccp0000044

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Luxton DD, Mccann RA, Bush NE, Mishkind MC, Reger GM. mHealth for mental health: integrating smartphone technology in behavioral healthcare. Prof Psychol Res Pract (2011) 42:505. doi:10.1037/a0024485

CrossRef Full Text | Google Scholar

15. Sun F-T, Kuo C, Cheng H-T, Buthpitiya S, Collins P, Griss M. Activity-aware mental stress detection using physiological sensors. In: Gris M and Yang G, editors. Mobile Computing, Applications, and Services. Springer Berlin Heidelberg: Springer (2010). p. 211–30.

Google Scholar

16. De Santos A, Sánchez-Avila C, Bailador-Del Pozo G, Guerra-Casanova J. Real-time stress detection by means of physiological signals. In: Yang J and Poh N, editors. Recent Application in Biometrics. Rijeka: IntTech (2011) p. 23–44.

Google Scholar

17. Plarre K, Raij A, Hossain SM, Ali AA, Nakajima M, Al’absi M, et al. Continuous inference of psychological stress from sensory measurements collected in the natural environment. Information Processing in Sensor Networks (IPSN), 2011 10th International Conference on: IEEE. Chicago (2011). p. 97–108.

Google Scholar

18. Bakker J, Pechenizkiy M, Sidorova N. What’s your current stress level? Detection of stress patterns from GSR sensor data. Data Mining Workshops (ICDMW), 2011 IEEE 11th International Conference on: IEEE). Vancouver (2011). p. 573–80.

Google Scholar

19. Alamudun F, Choi J, Gutierrez-Osuna R, Khan H, Ahmed B. Removal of subject-dependent and activity-dependent variation in physiological measures of stress. Pervasive Computing Technologies for Healthcare (PervasiveHealth), 2012 6th International Conference on: IEEE). San Diego (2012). p. 115–22.

Google Scholar

20. Choi J, Ahmed B, Gutierrez-Osuna R. Development and evaluation of an ambulatory stress monitor based on wearable sensors. IEEE Trans Inf Technol Biomed (2012) 16:279–86. doi:10.1109/TITB.2011.2169804

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Everly GS, Lating JM. The anatomy and physiology of the human stress response. A Clinical Guide to the Treatment of the Human Stress Response. New York: Springer (2002). p. 15–48.

Google Scholar

22. Martínez-Pérez B, De La Torre-Díez I, López-Coronado M. Mobile health applications for the most prevalent conditions by the World Health Organization: review and analysis. J Med Internet Res (2013) 15:e120. doi:10.2196/jmir.2600

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Kirschbaum C, Pirke KM, Hellhammer DH. The ‘Trier Social Stress Test’ – a tool for investigating psychobiological stress responses in a laboratory setting. Neuropsychobiology (1993) 28:76–81. doi:10.1159/000119004

CrossRef Full Text | Google Scholar

24. Wolpe J. The Practice of Behavior Therapy. Boston: Pergamon (1973).

Google Scholar

25. Lovibond P. Manual for the Depression Anxiety Stress Scales, Sydney Psychology Edition. Sydney: Psychology Foundation of Australia (1995).

Google Scholar

26. Pilkonis PA, Choi SW, Reise SP, Stover AM, Riley WT, Cella D, et al. Item banks for measuring emotional distress from the Patient-Reported Outcomes Measurement Information System (PROMIS®): depression, anxiety, and anger. Assessment (2011) 18:263–83. doi:10.1177/1073191111411667

PubMed Abstract | CrossRef Full Text | Google Scholar

27. McKinney W. Python for Data Analysis: Data Wrangling with Pandas, NumPy, and IPython. Sebastopol: O’Reilly Media, Inc. (2012).

Google Scholar

28. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: machine learning in Python. J Mach Learn Res (2011) 12:2825–30.

Google Scholar

29. Garreta R, Moncecchi G. Learning scikit-learn: Machine Learning in Python. Birmingham: Packt Publishing Ltd (2013).

Google Scholar

30. Johnston WS. Development of a Signal Processing Library for Extraction of SpO2, HR, HRV, and RR from Photoplethysmographic Waveforms [Dissertation]. Worcester: Worcester Polytechnic Institute (2006).

Google Scholar

31. Green D, Swets J. Signal Detection Theory and Psychophysics. 1966. New York: Peninsula Pub (1966). 888,889.

Google Scholar

32. Reilly PM, Shopshire MS. Anger management for substance abuse and mental health clients: a cognitive behavioral therapy manual. J Drug Addict Educ Erad (2014) 10:199.

Google Scholar

33. Weathers F, Ford J. Psychometric review of PTSD checklist (PCL-C, PCL-S, PCL-M, PCL-PR). Measurement of Stress, Trauma, and Adaptation. Brooklandville: Sidran Press (1996). p. 250–1.

Google Scholar

34. Crawford JR, Henry JD. The Depression Anxiety Stress Scales (DASS): normative data and latent structure in a large non-clinical sample. Br J Clin Psychol (2003) 42:111–31. doi:10.1348/014466503321903544

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Kay M, Santos J, Takane M. mHealth: New Horizons for Health through Mobile Technologies. Geneva: World Health Organization (2011). p. 66–71.

Google Scholar

36. WHO. Mental Health Action Plan 2013-2020. World Health Organization (2013).

Google Scholar

37. Donker T, Petrie K, Proudfoot J, Clarke J, Birch MR, Christensen H. Smartphones for smarter delivery of mental health programs: a systematic review. J Med Internet Res (2013) 15:e247. doi:10.2196/jmir.2791

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Dickerson SS, Kemeny ME. Acute stressors and cortisol responses: a theoretical integration and synthesis of laboratory research. Psychol Bull (2004) 130:355–91. doi:10.1037/0033-2909.130.3.355

PubMed Abstract | CrossRef Full Text | Google Scholar

39. Biondi M, Picardi A. Psychological stress and neuroendocrine function in humans: the last two decades of research. Psychother Psychosom (1999) 68:114–50. doi:10.1159/000012323

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Graveley EA, Oseasohn CS. Multiple drug regimens: medication compliance among veterans 65 years and older. Res Nurs Health (1991) 14:51–8. doi:10.1002/nur.4770140108

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Grossman P, Niemann L, Schmidt S, Walach H. Mindfulness-based stress reduction and health benefits: a meta-analysis. J Psychosom Res (2004) 57:35–43. doi:10.1016/S0022-3999(03)00573-7

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Rosenthal T, Alter A, Peleg E, Gavish B. Device-guided breathing exercises reduce blood pressure: ambulatory and home measurements. Am J Hypertens (2001) 14:74–6. doi:10.1016/S0895-7061(00)01235-8

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Turpin G, Shine P, Lader M. Ambulatory electrodermal monitoring: effects of ambient temperature, general activity, electrolyte media, and length of recording. Psychophysiology (1983) 20:219–24. doi:10.1111/j.1469-8986.1983.tb03291.x

CrossRef Full Text | Google Scholar

44. Fischer H. A Guide to US Military Casualty Statistics: Operation Inherent Resolve, Operation New Dawn, Operation Iraqi Freedom, and Operation Enduring Freedom. US Congressional Research Service (2014).

Google Scholar

45. Spoont MR, Nelson DB, Murdoch M, Rector T, Sayer NA, Nugent S, et al. Impact of treatment beliefs and social network encouragement on initiation of care by VA service users with PTSD. Psychiatr Serv (2014) 65:654–62. doi:10.1176/appi.ps.201200324

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Affairs DOV. Expanded access to non-VA care through the veterans choice program. Interim final rule. Federal Register (2014) 79:65571.

Google Scholar

47. Davis MM, Clark SJ, Butchart AT, Singer DC, Shanley TP, Gipson DS. Public participation in, and awareness about, medical research opportunities in the era of clinical and translational research. Clin Transl Sci (2013) 6:88–93. doi:10.1111/cts.12019

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Arora S, Sevdalis N, Nestel D, Woloshynowych M, Darzi A, Kneebone R. The impact of stress on surgical performance: a systematic review of the literature. Surgery (2010) 147:318–30. doi:10.1016/j.surg.2009.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

49. Winslow BD, Carroll MB, Martin JW, Surpris G, Chadderdon GL. Identification of resilient individuals and those at risk for performance deficits under stress. Front Neurosci (2015) 9:328. doi:10.3389/fnins.2015.00328

PubMed Abstract | CrossRef Full Text | Google Scholar

50. Cornum R, Matthews MD, Seligman ME. Comprehensive soldier fitness: building resilience in a challenging institutional context. Am Psychol (2011) 66:4–9. doi:10.1037/a0021420

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: stress, electrodermal response, heart rate, mobile applications, wearable devices, cognitive behavioral therapy, telemedicine

Citation: Winslow BD, Chadderdon GL, Dechmerowski SJ, Jones DL, Kalkstein S, Greene JL and Gehrman P (2016) Development and Clinical Evaluation of an mHealth Application for Stress Management. Front. Psychiatry 7:130. doi: 10.3389/fpsyt.2016.00130

Received: 13 June 2016; Accepted: 12 July 2016;
Published: 26 July 2016

Edited by:

Yasser Khazaal, University of Geneva, Switzerland

Reviewed by:

Ma Angeles Gomez Martínez, Pontifical University of Salamanca, Spain
Renzo Bianchi, University of Neuchâtel, Switzerland

Copyright: © 2016 Winslow, Chadderdon, Dechmerowski, Jones, Kalkstein, Greene and Gehrman. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Brent D. Winslow, brent.winslow@designinteractive.net

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.