Skip to content

Commit 91fe875

Browse files
committed
abstract +++
1 parent 4cc016a commit 91fe875

10 files changed

+1678
-110
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,2 @@
1+
eclipse.preferences.version=1
2+
encoding/discussion.tex=UTF-8

report/REM.bib

+1,330
Large diffs are not rendered by default.

report/discussion.tex

+30-18
Original file line numberDiff line numberDiff line change
@@ -6,15 +6,16 @@ \subsection{Software package for feature computation}
66
it was necessary to compute an exhaustive set of features for all consecutive five second epochs
77
over long (24h) time series.
88
For this purpose, \pr{}, a new \py{} package was developed based on
9-
\texttt{PyEEG}\citationneeded{}, which already implements several algorithms often used to study \gls{eeg}.
9+
\texttt{PyEEG}\cite{bao_pyeeg:_2011}, which already implements several
10+
algorithms often used to study \gls{eeg}.
1011
Very significant improvements in performance were achieved for almost all functions implemented in \texttt{PyEEG}
1112
(table~\ref{tab:benchmark}). These improvements will considerably speed-up prototyping of feature extraction
1213
and may be essential in order to build real time classifiers.
1314
In addition, such modifications will make it possible to compute features for a large number
1415
of recordings in reasonable time.
1516
Further improvements are possible, for instance,
16-
sample entropy was tentatively implemented in Julia programming language and performed 25 times faster than
17-
\pr{}'s implementation\footnote{implementation available at
17+
sample entropy was tentatively implemented in Julia programming language\cite{bezanson_julia:_2012}
18+
and performed 25 times faster than \pr{}'s implementation\footnote{Implementation available at
1819
\href{https://github.com/qgeissmann/Physiology.jl/blob/master/src/univariate.jl}{https://github.com/qgeissmann/Physiology.jl/blob/master/src/univariate.jl}.}
1920
Interestingly, it appears that the new implementation of sample and
2021
absolute entropy does not scale as well as the original implementation.
@@ -25,7 +26,8 @@ \subsection{Software package for feature computation}
2526
definitions (see \pr{} documentation, appendix).
2627
This unfortunatly apperas to be a common issue for academic software.
2728
The general status of the peer-review process and the reproducibility of programs and algorithms have
28-
recently drawn attention (see \citationneeded{Black-box; Can I reproduce your algo} for discussions about this issue).
29+
recently drawn attention (see \cite{morin_shining_2012,crick_can_2014} for
30+
discussions about this issue).
2931

3032
\subsection{Exhaustive feature extraction}
3133

@@ -36,32 +38,37 @@ \subsection{Exhaustive feature extraction}
3638

3739
Discrete wavelet decomposition is an extremely fast an accurate algorithm to filter a periodic
3840
signal into complementary and exclusive frequency sub-bands (fig.~\ref{fig:dwd}).
39-
XXX et al.(cite) \citationneeded{} obtained very promising results by computing a large number of features on the raw \gls{eeg} signal
40-
and a limited subset of features (\ie{} mean power and absolute values) in some wavelet coefficients.
41+
\c{S}en et al.\cite{sen_comparative_2014} obtained very promising results by
42+
computing a large number of features on the raw \gls{eeg} signal and a limited subset of features (\ie{} mean power and absolute values) in some wavelet coefficients.
4143
In contrast, in the present study, all features were computed on all frequency sub-bands.
4244
Interestingly, some of the features that are the most important for prediction would not have
4345
been discovered otherwise (see table~\ref{tab:importances}).
4446

47+
48+
4549
Many authors have modelled time series of epochs as if each epoch was statistically independent from each other.
4650
This assumption makes it straightforward to use classical machine learning techniques such as
47-
\glspl{svm}(\citationneeded{}), \glspl{ann}(\citationneeded{}), random forests(\citationneeded{}) and others.
51+
\glspl{ann}, \glspl{svm}\cite{crisler_sleep-stage_2008},
52+
random forests\cite{breiman_random_2001} and others.
4853
They have the advantage coping very well with non-linearity, can handle a large number of predictors and have many optimised implementations.
4954

5055
However, working with this assumption generally does not allow to account for temporal consistency of vigilance states.
5156
Indeed, prior knowledge of, for instance, the state transition probabilities cannot be modelled.
5257
Manual scorers use contextual information to make decisions.
5358
For example, if a given epoch has ambiguous features between \gls{rem} and awake,
5459
it is likely to be classified as awake given surrounding epochs are, less ambiguously, awake.
55-
For this reason, explicit temporal modelling, using, for instance, Hidden Markov Models has been investigated\citationneeded{}.
60+
For this reason, explicit temporal modelling, using, for instance, Hidden Markov Models has been investigated\cite{doroshenkov_classification_2007,pan_transition-constrained_2012}.
5661

5762
In order to benefit from the classical machine learning
5863
framework whist including temporal information,
59-
it is possible to create, new variables, accounting for the temporal variation\citationneeded{}.
64+
it is possible to create, new variables, accounting for the temporal
65+
variation\cite{dietterich_machine_2002}.
6066
This study demonstrated that addition of temporal context significantly improved predictive accuracy (fig.\ref{fig:temporal_integration}).
6167
The convolution approach (eq.\ref{eq:window}) appeared to provide better results.
6268
Instead of averaging feature after calculation, it may be advantageous to compute features over epochs of different length in a first place.
6369
Thus, the accuracy of local of non additive features, such as median, will be improved. In addition to local mean of feature, other variables, such as local
64-
slope and local variance of each feature may improve classification \citationneeded{(Deng 2013).}
70+
slope and local variance of each feature may improve
71+
classification\cite{deng_time_2013}.
6572

6673
Although addition of time-dependent variables improved accuracy over a time-unaware model, their use can be seen as controversial.
6774
Indeed, including prior information about sleep structure will cause problems if the aim is to find differences in sleep structure.
@@ -79,15 +86,16 @@ \subsection{Exhaustive feature extraction}
7986

8087
\subsection{Random forest classification}
8188

82-
In this study, random forest\citationneeded{} classifiers were exclusively used.
89+
In this study, random forest classifiers\cite{breiman_random_2001} were exclusively used.
8390
In addition to their capacity to model non-linearity, they are very efficient at handling very large number of variables.
84-
Recently very promising classification of sleep stages in human were generated using this algorithm\citationneeded{}.
91+
Recently very promising classification of sleep stages in human were generated
92+
using this algorithm\cite{sen_comparative_2014}.
8593
A very interesting feature of random forest is their
8694
natural ability to generate relative values of importance for the different predictors.
8795
These values quantifies how much each variables contributes to the predictive power of the model.
8896
This feature is extremely useful because it allows using random forests for variable selection.
8997
This can be used to reduce dimensionality of the variable space without losing predictive power (fig.\ref{fig:variable_elimination}),
90-
but also to study conditional variable importance, or, for instance,
98+
but also to study conditional variable importance\cite{strobl_conditional_2008}, or, for instance,
9199
determine which variables are important to segregate pairs of classes.
92100
Whilst random forests are not guaranteed to be the best predictor, they allow fast and in-depth preliminary investigation.
93101
Finally, underlying mechanisms of random forest (\ie{} how variables are combined) is relatively simple to understand in terms of binary logic.
@@ -99,7 +107,8 @@ \subsection{Rigorous and comprehensive model evaluation}
99107

100108
Previous research, using classical statistical learning framework,
101109
have often assessed their classifier through cross-validation.
102-
It often unclear how sampling was performed to generate training and testing sets.
110+
It however often unclear how sampling was performed to generate training and
111+
testing sets\cite{ebrahimi_automatic_2008, chapotot_automated_2010, sen_comparative_2014}.
103112
Time series of epochs are dense and, in general,
104113
the features (and labels) at a given time are very correlated with surrounding features.
105114
Therefore, if random sampling of even 50\% of all epochs, from all time series, was performed,
@@ -111,7 +120,8 @@ \subsection{Rigorous and comprehensive model evaluation}
111120
There are several way to reduce overfitting including limiting the maximal number of splits when growing classification trees, or pruning trees.
112121
However, it never possible to unsure a model will not overfit \emph{a priori}.
113122
Thus it remain necessary to assess the model fairly.
114-
In this study, systematic stratified cross-validation was performed.
123+
In this study, systematic stratified cross-validation was
124+
performed \cite{ding_querying_2008}.
115125
As a result, all predictions made on any 24h time series are generated by models
116126
that did not use any point originating from this same time series. This precaution simulate the the behaviour of the predictor with new recordings.
117127
Cross-validation was not only used to generate overall value of accuracy, but also, to further assess differences in sleep patterns (fig. \ref{fig:struct_assess}).
@@ -127,7 +137,9 @@ \subsection{Quality of the raw data}
127137
The ground truth labels used in this study has been generated by a two pass semi-automatic method.
128138
In a first place, an automatic annotation is performed based on a human-defined variable threshold.
129139
Then, the expert visually inspect the result and correct ambiguities.
130-
The first pass was originally designed to combine, through logical rules, four epochs of five seconds to produce 20s epochs\citationneeded{}.
140+
The first pass was originally designed to combine, through logical rules, four
141+
epochs of five seconds to produce 20s
142+
epochs\cite{costa-miserachs_automated_2003}.
131143
However, it was simplified in-house in order to produce
132144
only five second epochs, ignoring the last step, and has since not been reassessed against manual scoring.
133145
It is expected that this simplification increased divergence with manual scorers.
@@ -142,8 +154,8 @@ \subsection{Quality of the raw data}
142154

143155
\subsection{Overall results}
144156
The predictions of the classifier presented in this research agreed with ground truth for 92\% of epochs (table~\ref{tab:confus}).
145-
Although the limitation of the ground truth annotation make it is hard to put this result into perspective,
146-
this score is very promising.
157+
Although the limitation of the ground truth annotation makes it is difficult to
158+
put this result into perspective, this score is very promising.
147159
In addition, prediction did not result in significant difference in prevalences.
148160
However, there were, on average, much less \gls{rem} episodes in the predicted time series.
149161
The duration of \gls{rem} episodes was also over-estimated by prediction (though this is only marginally significant).

report/matmet.tex

+5-1
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,9 @@ \subsection{Data acquisition}
2020
A first, human assisted, pass was applied to generate preliminary annotations on the basis of logical rules\cite{costa-miserachs_automated_2003}.
2121
Then, the expert visually inspected and, when required, corrected the annotations.
2222
Annotation were generated for consecutive epochs of approximately 5.0s.
23+
24+
Data acquisition and manual annotation were performed by Dr. Valentina
25+
Ferretti and Eleonora Steinberg prior to this project.
2326
%~
2427
\subsection{Data preprocessing}
2528

@@ -91,7 +94,8 @@ \subsection{Stratified Cross Validation and sampling}
9194
originating from all \emph{other} time series \cite{ding_querying_2008}.
9295

9396
Quantification of variable importance and accuracy (figs.~\ref{fig:variable_elimination} and ~\ref{fig:temporal_integration}) was performed in order
94-
not to underestimate the relevance of the minority class (\gls{rem}).
97+
not to underestimate the relevance of the minority class
98+
(\gls{rem})\cite{boulesteix_overview_2012}.
9599
Class unbalancedness was accounted for by fitting predictors on balanced subsamples (750 epochs of each class per tree).
96100

97101

0 commit comments

Comments
 (0)