Skip to content

Commit 4aef02f

Browse files
committed
Update results.tex
1 parent 8727638 commit 4aef02f

File tree

1 file changed

+21
-22
lines changed

1 file changed

+21
-22
lines changed

report/results.tex

+21-22
Original file line numberDiff line numberDiff line change
@@ -2,30 +2,28 @@
22
\section{Results} \label{results}
33

44
\subsection{A new efficient \py{} package}
5-
Several algorithms to extract features from univariate time series had already
5+
Several algorithms to extract features from univariate time series have already
66
been implemented in the \py{} package \pyeeg{}\cite{bao_pyeeg:_2011}.
7-
Unfortunately, some of them were critically slow, and could therefore not realistically have been used in the present study.
7+
Unfortunately, some of them were critically slow, and could therefore not realistically be used in the present study.
88
Preliminary investigation of \pyeeg{} source code revealed that runtimes may be improved mainly by vectorising expressions and pre-allocating of temporary arrays.
99
Therefore, systematic reimplementation of all algorithms in \pyeeg{} was undertaken.
10-
Very significant improvement in performance were achieved for almost all functions(table~\ref{tab:benchmark}).
10+
Very significant improvements in performance were achieved for almost all functions(table~\ref{tab:benchmark}).
1111
Critically, sample and approximate entropies\cite{richman_physiological_2000} became usable in
1212
reasonable time.
1313
For instance, 40h of CPU time were originally required to compute sample entropy on a all 5 second epochs in a 24h recording.
1414
The improved implementation cut this down to 55min.
1515
\input{./tables/benchmark}
1616

17-
Importantly, several mathematical inconsistencies between the original code and the mathematical definitions were also noticed.
17+
Importantly, several mathematical inconsistencies between the original code and the mathematical definitions were also detected.
1818
This affected five of the eight reimplemented functions(table~\ref{tab:benchmark}, rightmost column).
19-
Detail of the corrections performed are provided, as notes, in the documentation of the new package (see appendix, section 2.3).
20-
Numerical results for the three functions were consistent throughout optimisation.
19+
Details of the corrections performed are provided, as notes, in the documentation of the new package (see appendix, section 2.3).
20+
Numerical results for the three functions were consistent throughout optimisation.
2121

2222
In order to facilitate feature extraction, several data structures and routines, which were not
23-
available in \pyeeg{},
24-
were also implemented
25-
in a new \py{} package named \pr{}.
23+
available in \pyeeg{}, were also implemented in a new \py{} package named \pr{}.
2624
Briefly, extensions of \texttt{numpy} arrays\cite{walt_numpy_2011} providing
2725
meta-data, sampling frequency, and other attributes were used to represent time series.
28-
User friendly indexing with string representing time was also developed.
26+
User friendly indexing with strings representing time was also developed.
2927
In addition, a container for time series of discrete annotation levels, each linked to a confidence level, was built.
3028
Importantly, a container for multiple time series, which supports heterogeneous (between time series) sampling frequencies was implemented.
3129
The new package also provides visualization, input/output, and wrappers for resampling and discrete wavelet decomposition.
@@ -34,17 +32,17 @@ \subsection{A new efficient \py{} package}
3432
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
3533

3634
\subsection{Twenty variables can generate accurate predictions}
37-
Including temporal information (see next section) wil result in multiplication of the number of variable, rendering computation difficult, and prediction potentially less accurate.
35+
Including temporal information (see next section) will result in multiplication of the number of variables, rendering computation difficult, and prediction potentially less accurate.
3836
Therefore, recursive feature elimination\cite{menze_comparison_2009} based on
3937
Gini variable importance was undertaken.
4038
Starting with all 164 variables, random forests were trained, and the number of features was reduced by a factor $1.5$ by eliminating the least important variables.
4139
For each iteration, the stratified cross-validation error (see material and methods) was computed (fig.~\ref{fig:variable_elimination}). Five replicates were performed.
4240

4341
\input{./figures/variable_elimination}
4442
The predictive accuracy globally increases with the number of variables.
45-
Interestingly, using only the two most important variable already results in less that 20\% average error.
46-
However, this increase is very moderate for ($p>9$) this indicates that dimensionality can be considerably reduced without largely impacting accuracy.
47-
For further investigation, $p=21$ was considered to be a good compromise between error and computational load. Their relative importances are listed in table~\ref{tab:importances}.
43+
Interestingly, using only the two most important variables already results in less that 20\% average error.
44+
Additionally, this increase in predictive accuracy is very moderate for ($p>9$). This indicates that dimensionality can be considerably reduced without largely impacting accuracy.
45+
For further investigation, $p=21$ was considered to be a good compromise between error and computational load. Their relative importance is listed in table~\ref{tab:importances}.
4846
\input{./tables/importances}
4947
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
5048

@@ -83,14 +81,14 @@ \subsection{Structural differences with ground truth}
8381

8482
\input{./tables/confus}
8583

86-
The global confusion matrix is presented in table \ref{tab:confus}. The overall accuracy was 0.92.
87-
For \gls{nrem} and wake, both specificity and positive predictive value were above 0.92. However
84+
The global confusion matrix is presented in table \ref{tab:confus}. The overall accuracy compared to manual scoring was 0.92.
85+
For \gls{nrem} and wake, both specificity and positive predictive value were above 0.92. However,
8886
for \gls{rem} epochs, the specificity is only 0.74, and the false detection rate is 15\%.
8987

9088
In order to investigate the structural differences between ground truth and the predicted states,
9189
three metrics describing physiological properties of sleep were computed (fig.~\ref{fig:struct_assess}).
9290

93-
Prevalence(fig.~\ref{fig:struct_assess}A) of states is a widely used metric to
91+
Prevalence of states (fig.~\ref{fig:struct_assess}A) is a widely used metric to
9492
describe sleep patterns.
9593
No significant difference was found between the ground truth and predicted prevalences ($p-value > 0.13$ for all, z test on the interaction terms of $\beta$ regression).
9694

@@ -106,16 +104,17 @@ \subsection{Structural differences with ground truth}
106104
(t test on the interaction terms of linear mixed model).
107105

108106
\subsection{Attribution of confidence score to predictions}
107+
109108
Since classification may be inaccurate, it would be interesting to associate `confidence' score to each prediction.
110-
A entropy based confidence metric $c$ (eq. \ref{eq:entropy}) was defined for this purpose.
111-
In order to validate it, the average cross-validation error was computed for different ranges of confidence (fig\ref{fig:error}A).
109+
An entropy-based confidence metric $c$ (eq. \ref{eq:entropy}) was defined for this purpose.
110+
In order to validate it, the average cross-validation error was computed for different range of confidence (fig\ref{fig:error}A).
112111

113112
\input{./figures/error}
114113

115114
As expected, the probability of misclassification decreases monotonically with $c$.
116-
In addition, error rate seem to tend to zero when the confidence value is one, and, for confidences close to zero, the predictor is very inaccurate.
117-
These characteristics indicate that $c$ can be used to as a supporting value for predictions.
118-
One application of such a confidence level could be to provide user with an overall quality assessment.
115+
In addition, error rates tend to zero when the confidence value is one, and, for confidences close to zero, the predictor is very inaccurate.
116+
These characteristics indicate that $c$ can be used as a supporting value for predictions.
117+
One application of such a confidence level could be to provide users with an overall quality assessment.
119118
In addition, it makes it possible to display confidence while visually inspecting a recording (fig\ref{fig:error}C) in order to facilitate resolution of ambiguities.
120119

121120

0 commit comments

Comments
 (0)