Skip to content

CornellPACLab/open-data-exploratory

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Open Data Exploratory Analysis

Corresponding GitHub repository for the paper: Insert title

Maintained by: dadler@infosci.cornell.edu (GitHub ID: dadler6)

Assessing the Potential for an Open Data Community

Task Motivation and Research Questions

For this work, we not only wished to survey literature and identify challenges we hypothesized open PS4MH data could solve. We also wished to empirically assess this hypothesis. We designed an experiment to answer the following research questions (RQs), each corresponding to one of the 4 challenges:

[RQ1: Generalizability Across Data and Studies] Does a model trained with one dataset generalize to another dataset?

[RQ2: Human Behavior Changes over Time] Do modeled sensor feature-mental health relationships created using data collected from one time period, across one age group, apply to data collected in another time period from another age group?

[RQ3: Individual-Level Heterogeneity] How does individual-level heterogeneity impact model performance across two datasets?

[RQ4: Device Uncertainty] Will models created using data collected from one set of devices apply to data collected using a different set of devices?

We used 2 publicly available [1–9] PS4MH datasets to design our feasibility analysis. To the best of our knowledge, these are the only open PS4MH datasets. Both datasets contained sensor data and self-reported mental health symptom EMAs, and the differences between datasets (see Table 2) highlighted all 4 challenges described in the prior section. The datasets were thus ideal to answer the 4 RQs.

Our task explored if an EMA prediction model trained using one dataset could accurately predict EMAs in the other dataset. If the model was generalizable (RQ1) between datasets despite different data collection periods (RQ2), population heterogeneity (RQ3), and data collection devices (RQ4), there is promise in open PS4MH data. In machine learning, applying a model trained with a source dataset to a different target dataset is called transfer learning [10]. Transfer learning is successful if a model trained with the source achieves similar or greater performance when applied to the target, compared to using a model trained exclusively on the target. Negative transfer occurs when model performance does not improve performance on the target, but instead diminishes performance, typically due to differences between the datasets. If negative transfer occurs, only a portion of the source dataset may be useful for transfer learning [11].

Dataset Overview

The CrossCheck and StudentLife datasets are 2 open (see [12,13]) PS4MH datasets that contain smartphone sensing and mental health symptoms collected from individuals. Please refer to the Appendix for an overview of the data collection procedures, and [2,4,14] for more details regarding the studies that collected this data. These publications extensively document the participant recruitment, enrollment criteria, and data collection system development, which were not a part of this work. In our task, the larger CrossCheck dataset was designated as the source dataset, and the smaller StudentLife dataset was the target dataset. The CrossCheck dataset was determined to be “larger”, because data was collected for more participants over a longer period of time compared to the StudentLife dataset. See Table 2 for a comparison of these two datasets.

Prediction Model Training and Assessment

We used leave-one-subject-out cross-validation (LOSO-CV) to assess transfer learning. LOSO-CV is a common procedure to estimate the generalizability of PS4MH prediction models [3,4,15]. In LOSO-CV, each target participant is removed from the data, and the remaining data is used for model training. The trained model predicts EMA scores for the held out participant. The predictions across each held-out participant are concatenated to estimate model performance. Lasso regression models were used for prediction. For each held-out participant, we trained a lasso regression model using the remaining source and target data. Since negative transfer may occur, we also developed a second LOSO-CV procedure where a lasso regression model was trained using each participant’s data (“subpopulation models”). We then matched each held-out participant to the model trained using a different participant’s data that maximized the coefficient of determination (R2). Model performance across participants was measured using the R2 and mean absolute error (MAE). EMA values were non-normally distributed (Shapiro-Wilk P<.001). We performed Wilcoxon signed-rank tests hypothesizing that the squared errors (SE) between the predicted and actual EMA values were significantly greater using the target compared to the source as training data. Model parameters and prediction procedures are detailed in the Appendix “Sensor-EMA Alignment” and “Transfer Learning Cross-Validation” sections.

Task Results

Transferable Sensing and EMA Data

Table 3 describes the aligned data elements, across the 2 datasets. We qualitatively identified EMAs that aligned between each dataset based upon question and response structure. We then required a minimum number of data points per-participant [1,4]. Similar to previous work [4], we required ≥30 EMA values per source and target participant, which was about 50% of the intended target EMA collection. Table 4 shows the resulting aligned dataset. We list characteristics separately for each EMA we predicted in the target population because not all individuals who responded to ≥30 sleep EMAs also responded to ≥30 stress EMAs. Specific feature and EMA distribution alignment statistics can be found in the Appendix “Sensing Distribution Alignment” and “EMA Distribution Alignment” sections. EMA values were non-normally distributed (omnibus test for normality sleep=631 P<.001 and stress=633, P<.001) [16]. Mann-Whitney U tests showed that the EMA distributions were significantly different (α=.05) between source and target sleep (U=1,811,706, P=.002) and stress (U=542,292, P<.001). We calculated Cureton’s matched-pairs rank-biserial correlation (RBC∊[−1, 1]) [17] to show the magnitude of differences for the sleep (RBC=-0.07) and stress (RBC=0.59) EMAs.

LOVO-CV Results Using Entire Population Models

Table 5 describes the LOSO-CV results using prediction models trained with the entire source versus target data. All models were not predictive (highest performing models R2≈0). SE values were found to be non-normally distributed (Shapiro-Wilk P<.001). We found the sleep EMA prediction SE was marginally greater (α=.10) using the target data (W=96,087, P=.052) and differences were small (RBC=0.08). The predicted stress EMA SE was significantly lower (W=7,179, P=1.000, RBC=−0.70) using the target data.

R2 MAE W P RBC
Training Data Target Source Target Source
Sleep -0.01 0.00 0.66 0.66 96,087 0.052 0.08
Stress -0.02 -1.95 0.57 1.05 7,179 1.000 -0.70

LOVO-CV Results Using Subpopulation Models

Table 6 describes the LOSO-CV results using prediction models trained with source versus target subpopulations. Source subpopulation models achieved the highest performance with R2=0.18, MAE=0.57 for the sleep and R2=0.22, MAE=0.48 for the stress EMA. Training with source data significantly (α=.05) decreased the SE compared to training with target for the sleep EMA (W=103,373, P<.001, RBC=0.16) and marginally (α=.10) decreased the stress (W=26,096, P=.067, RBC=0.10) EMA SE.

R2 MAE W P RBC
Training Data Target Source Target Source
Sleep 0.11 0.18 0.62 0.57 103,373 <0.001 0.16
Stress 0.18 0.22 0.49 0.48 26,096 0.057 0.10

Repository Information

Data

The cleaned data used for the lasso regression analysis can be found in the data directory. The cleaned datasets are the CrossCheck and StudentLife datasets, with epoch features computed, and these features are averaged across 3 days to align with the EMAs. The cleaned data is called:

  • Cleaned CrossCheck: data/crosscheck_daily_data_cleaned_w_sameday.csv
  • Cleaned StudentLife: data/studentlife_daily_data_cleaned_w_sameday_03192020.csv

These files are used in the notebooks/lasso_regression.ipynb notebook.

Due to filesize, we are not including the original public datasets in this repository, but they are needed if you wish to run the cleaning code (notebooks/data_cleaning.ipynb). Please download them at the following links:

Code

The code to run the Lasso Regression models can be found in the following notebook:

  • notebooks/lasso_regression.ipynb

The code to clean the original public datasets can be found in:

  • notebooks/data_cleaning.ipynb

Utility code

The following .py files are used throughout the lasso regression and data cleaning notebooks. They are:

  • src/util.py
  • src/cleaning_util.py
  • src/regression_cv.py

Abbreviations

  • API: application programming interface
  • EMA: ecological momentary assessment
  • LOSO-CV: leave-one-subject-out cross-validation
  • MAC: media access control
  • MAE: mean absolute error
  • NA: not applicable
  • PS4MH: personal sensing for mental health
  • RBC: rank-biserial correlation
  • RQ: research questions
  • SE: squared error

References

References

  1. Adler DA, Ben-Zeev D, Tseng VW-S, Kane JM, Brian R, Campbell AT, Hauser M, Scherer EA, Choudhury T. Predicting Early Warning Signs of Psychotic Relapse From Passive Sensing Data: An Approach Using Encoder-Decoder Neural Networks. JMIR MHealth UHealth 2020 Aug 31;8(8):e19962. [doi: 10.2196/19962]
  2. Wang R, Chen F, Chen Z, Li T, Harari G, Tignor S, Zhou X, Ben-Zeev D, Campbell AT. StudentLife: assessing mental health, academic performance and behavioral trends of college students using smartphones. Proc 2014 ACM Int Jt Conf Pervasive Ubiquitous Comput [Internet] Seattle, Washington: Association for Computing Machinery; 2014 [cited 2020 Mar 1]. p. 3–14. [doi: 10.1145/2632048.2632054]
  3. Wang R, Wang W, Aung MSH, Ben-Zeev D, Brian R, Campbell AT, Choudhury T, Hauser M, Kane J, Scherer EA, Walsh M. Predicting Symptom Trajectories of Schizophrenia Using Mobile Sensing. Proc ACM Interact Mob Wearable Ubiquitous Technol 2017 Sep;1(3):110:1-110:24. [doi: 10.1145/3130976]
  4. Wang R, Scherer EA, Tseng VWS, Ben-Zeev D, Aung MSH, Abdullah S, Brian R, Campbell AT, Choudhury T, Hauser M, Kane J, Merrill M. CrossCheck: toward passive sensing and detection of mental health changes in people with schizophrenia. Proc 2016 ACM Int Jt Conf Pervasive Ubiquitous Comput - UbiComp 16 [Internet] Heidelberg, Germany: ACM Press; 2016 [cited 2020 Mar 1]. p. 886–897. [doi: 10.1145/2971648.2971740]
  5. Tseng VW-S, Sano A, Ben-Zeev D, Brian R, Campbell AT, Hauser M, Kane JM, Scherer EA, Wang R, Wang W, Wen H, Choudhury T. Using behavioral rhythms and multi-task learning to predict fine-grained symptoms of schizophrenia. Sci Rep Nature Publishing Group; 2020 Sep 15;10(1):15100. [doi: 10.1038/s41598-020-71689-1]
  6. Morshed MB, Saha K, Li R, D’Mello SK, De Choudhury M, Abowd GD, Plötz T. Prediction of Mood Instability with Passive Sensing. Proc ACM Interact Mob Wearable Ubiquitous Technol 2019 Sep 9;3(3):1–21. [doi: 10.1145/3351233]
  7. Buck B, Scherer E, Brian R, Wang R, Wang W, Campbell A, Choudhury T, Hauser M, Kane JM, Ben-Zeev D. Relationships between smartphone social behavior and relapse in schizophrenia: A preliminary report. Schizophr Res 2019 Jun 1;208:167–172. [doi: 10.1016/j.schres.2019.03.014]
  8. Buck B, Hallgren KA, Scherer E, Brian R, Wang R, Wang W, Campbell A, Choudhury T, Hauser M, Kane JM, Ben-Zeev D. Capturing behavioral indicators of persecutory ideation using mobile technology. J Psychiatr Res 2019 Sep 1;116:112–117. [doi: 10.1016/j.jpsychires.2019.06.002]
  9. Lamichhane B, Ben-Zeev D, Campbell A, Choudhury T, Hauser M, Kane J, Obuchi M, Scherer E, Walsh M, Wang R, Wang W, Sano A. Patient-Independent Schizophrenia Relapse Prediction Using Mobile Sensor Based Daily Behavioral Rhythm Changes. In: Ye J, O’Grady MJ, Civitarese G, Yordanova K, editors. Wirel Mob Commun Healthc Cham: Springer International Publishing; 2021. p. 18–33. [doi: 10.1007/978-3-030-70569-5_2]
  10. Pan SJ, Yang Q. A Survey on Transfer Learning. IEEE Trans Knowl Data Eng 2010 Oct;22(10):1345–1359. [doi: 10.1109/TKDE.2009.191]
  11. Weiss K, Khoshgoftaar TM, Wang D. A survey of transfer learning. J Big Data 2016 Dec;3(1):9. [doi: 10.1186/s40537-016-0043-6]
  12. StudentLife Dataset [Internet]. [cited 2021 Apr 20]. Available from: https://studentlife.cs.dartmouth.edu/dataset.html
  13. crosscheck [Internet]. [cited 2021 Apr 20]. Available from: https://kaggle.com/dartweichen/crosscheck
  14. Ben-Zeev D, Brian R, Wang R, Wang W, Campbell AT, Aung MSH, Merrill M, Tseng VWS, Choudhury T, Hauser M, Kane JM, Scherer EA. CrossCheck: Integrating self-report, behavioral sensing, and smartphone use to identify digital indicators of psychotic relapse. Psychiatr Rehabil J 2017;40(3):266–275. [doi: 10.1037/prj0000243]
  15. Saeb S, Lonini L, Jayaraman A, Mohr DC, Kording KP. The need to approximate the use-case in clinical machine learning. GigaScience [Internet] 2017 May 1 [cited 2021 Apr 23];6(gix019). [doi: 10.1093/gigascience/gix019]
  16. D’Agostino RB. An Omnibus Test of Normality for Moderate and Large Size Samples. Biometrika [Oxford University Press, Biometrika Trust]; 1971;58(2):341–348. [doi: 10.2307/2334522]
  17. Kerby DS. The Simple Difference Formula: An Approach to Teaching Nonparametric Correlation. Compr Psychol SAGE Publications Inc; 2014 Jan 1;3:11.IT.3.1. [doi: 10.2466/11.IT.3.1]
  18. Abdullah S, Matthews M, Murnane EL, Gay G, Choudhury T. Towards circadian computing: “early to bed and early to rise” makes some of us unhealthy and sleep deprived. Proc 2014 ACM Int Jt Conf Pervasive Ubiquitous Comput [Internet] Seattle Washington: ACM; 2014 [cited 2021 Mar 9]. p. 673–684. [doi: 10.1145/2632048.2632100]

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published