You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
train3.csv
in the latest version of the dataset (I have not controlled the 20.07 verision) only contain examples labeled as normal in the last four rows. Attacks seem to exist in the testing portion of the dataset, but I don't think the current proposed splits are useful for learning tasks targeting attack columns (I have not checked the single timestamps with the attacks present in the documentation- maybe they have not been correctly labeled as such?).
The text was updated successfully, but these errors were encountered:
Our assumption is that "training data can only be collected during normal operation".
Therefore, there is no attack data in the training data in the HAI dataset.
Anomaly detection with HAI datasets can use unsupervised learning
(or semi-supervised learning, assuming that the training data are all normal).
I did not notice the assumption you mentioned, thank you for the explanation. I see how the splits are suited for unsupervised learning tasks. Maybe a split for supervised learning can be proposed in future versions of the dataset, but I don't really have any problems with this issue being closed now.
The three training data files,
train1.csv
train2.csv
train3.csv
in the latest version of the dataset (I have not controlled the 20.07 verision) only contain examples labeled as normal in the last four rows. Attacks seem to exist in the testing portion of the dataset, but I don't think the current proposed splits are useful for learning tasks targeting attack columns (I have not checked the single timestamps with the attacks present in the documentation- maybe they have not been correctly labeled as such?).
The text was updated successfully, but these errors were encountered: