Sun Dec 8th through Sat the 14th, 2019 at Vancouver Convention Center
Update after author response --------------------------------- I would like to thank the authors for their response. "most p" is still somewhat not very precise and I would like the author to clarify the issues I raised in the final version. Summary of the main ideas: ---------------------------------- The authors theoretically study Nearest Neighbor methods for anomaly detection in the unsupervised setting using the distance-to-a-measure (DTM) quantity. If I understood correctly the major differences from the related work are - [1, 2, 3] (papers about theoretical analyses of kNN methods for anomaly detection) consider the semi-supervised setting (novelty detection) whereas this paper deals with the unsupervised setting - [1, 2, 3] use different techniques (geometric entropy and conformal prediction) and do not rely on the general DTM quantity. - the DTM related work gives a limiting distribution and confidence sets but no finite sample bounds. The authors start the paper by an extensive benchmark showing that Nearest Neighbor methods perform reasonably well compared to other state-of-the-art methods. They then prove finite sample bounds on the deviation of the empirical k-th Nearest neighbor radius and the empirical distance-to-a-measure (DTM) from their true counterparts. They finally apply these results to anomaly detection using kNN methods. Originality ------------ This paper is the first, according to the authors and the cited related work, to state finite sample bounds on the deviation of the empirical k-th Nearest neighbor radius and the empirical distance-to-a-measure (DTM) from their true counterparts and to apply these results to anomaly detection. Quality and clarity ---------------------- The paper is well-written and easy to read. Having a second section about experiments was a bit unusual but this motivates the analysis of the next sections. However I would add a sentence in the introduction of section 2 to emphasize that the methods/algorithms concerned by the theoretical analysis of this paper are k-NN, kth-NN and DTM2. It could be useful to restate at the end of section 3.1 what is mentioned in the introduction of the paper about the difference with the theoretical studies done in [1, 2, 3]. This would clarify a bit more the difference with the related work when reading the theoretical section. The authors should also clarify the main differences (listed in the Summary, if correct). In section 2, for the experiments of Figure 1, hyperparameters for kNN, kthNN and DTM2 methods are clearly stated (k = 0.03 * sample size). I assume the same choice is made for LOF (I would make this clearer). Hyperparameters for LODA and Iforest are absent (e.g. number of trees and subsampling parameter for Iforest). In section 3.4, we do not know which threshold the authors are using to declare a sample as abnormal or not. I liked the explanations and interpretations of the assumptions/results/definitions given in the paper. I found them very clear and useful. I would like the authors to clarify where in  we can find Assumption A1. Some notations are used at different places in the paper with different meanings: - epsilon : epsilon in [0, 1) is used in the contamination model (~ l.123) to denote the proportion of contamination. it is also used in proposition 3.6 with the same meaning. However it is used in the definition of Assumption A1 with a different meaning. - p : the authors should be more careful when using p. First it is said l.134 that for the rest of the article p = k/n but this is not true everywhere in the paper: for instance it is used in equation (1) as the variable of integration, and for equation (2) in Theorem 3.5 to be true we need Theorem 3.3 to be true for all p in [0, 1), not just k/n. The authors need to better clarify when p = k/n (e.g. in the interpretation of \hat r_p(x) as the kth-NN radius or in Theorem 3.4) or when p is/needs to be an arbitrary number in [0, 1). - eta : eta is used in the definition of assumption A1 and in equation (14), both having a different meaning. These inconsistencies should be fixed for a better clarity and to prevent the reader from being confused. About Theorem 3.5, equation (2) : for the proof to hold, Theorem 3.3 must be true for all p in [0,1), as the integral over all p in [0, m) is considered, m in [0, 1). Note also that the event with probability greater than 1 - delta such that the inequality of Theorem 3.3 is true shall not depend on p, otherwise additional arguments are needed to ensure that we can integrate over [0, m) and still have the result with probability at least 1 - delta ([0, m) contains uncountably many p). The fact that the event does not depend on p is not clear in the paper. Although this can easily be checked in the proof, it would be much better if the authors make this clear when stating the result, e.g. "with probability at least 1 - delta: for all p in [0, 1), ..." and maybe add a sentence to say that the event does not depend on p. About the proof of Theorem 3.3 : the authors could also clarify that this holds for all p in [0, 1) by writing something like "with probability at least 1 - delta: for any ball B and for all p in [0, 1)", especially because as explained above we need the result to be true for all p and the event to not depend on p but also because in the reference given by the authors, equations (1) and (2) are stated only for p=k/n. They could also add in the supplementary material that these results are Lemma C.2 and Theorem C.1 of , except that these results are here stated for all p instead of p=k/n. Besides, the needed reference is  and not .  is stated in . I would consider removing  here or better explain how these inequalities are obtained: e.g. "equations (1) and (2) are proven in  by standard VC theory". About Theorem 3.5, equation (3) : I would need additional arguments/explanations to be convinced that this result holds. No proof is given. I assumed that I could follow the proof of equation (2) but Theorem 3.4 only holds for p=k/n and we need it to hold for all p as explained above. The authors can maybe use the fact that \hat r_p(x) is piecewise constant as a function of p? Minor comments: - l.120 : n i.i.d. realization -> n i.i.d. realizations - l.126 : our goal is devise -> to devise - l.127 : anomalous one -> anomalous ones - l.137 and l.140 (in assumptions (A1) and (A2)) : There exists -> There exist - equation following l.205 -> it would be better to use the same notation under the min as the one used under the max. Besides x_I is undefined. - l.213 : in terms the distance -> in terms of the distance - l.221 : the bulk the distribution -> the bulk of the distribution - The authors wrote in the paper (l.62) that the code for the experiments will be made publicly available so I assume that this means once (and if) the paper is accepted while in the reproducibility checklist they said that they provided a link to a downloadable source code. Significance ------------ The finite sample bounds derived in this paper are likely to be of interest to derive new theoretical results. NN methods for anomaly detection are already famous among practitioners and it is important to have theoretical guarantees.
The main contributions of this paper are described in the previous section. The presented empirical study follows the steps of others (as cited by the authors), but studying the DTM-based models is original. The paper is generally well-written, presenting the results clearly. It is somewhat confusing that Table1 shows failure rates (lower is better), while Table2 shows performance (higher is better). The empirical study is comprehensive, and the theoretical analysis is definitely non-trivial. The significance of the paper is mainly in suggesting the use of DTM2 for anomaly detection and pointing out the possible advantage of nearest neighbor based methods for anomaly detection in high dimensional data. While this is probably not a breakthrough in the field, it is of significant interest to data science practitioners. I would grade this paper as a "good submission".
This paper proposes a new outlierness measure, called DTM, for outlier detection, and theoretically analyze its performance. In addition, the paper performs experiments on various real-world datasets to examine the effectiveness of outlier detection methods. Although the theoretical analysis of the paper is interesting, I have the following concerns. - This paper provides several theoretical results for NN-based outlier detection, which is an interesting contribution. However, empirical evaluation does not directly support the obtained theoretical results. For example, how are the assumptions (A0-A2) natural in real-world datasets? How is the result in Proposition 3.6 realistic? - An important empirical result is that an ensemble method is the most effective. However, theoretical comparison between an ensemble method and the proposed method is not performed. - The proposed method includes several parameters. However, the sensitivity is not examined, which is crucial to show the usefulness of the proposed method. - Important references about NN-based outlier detection are missing. In particular,  and  are known to be the state-of-the-art NN-based methods using sampling with rigorous theoretical analyses. Moreover,  also performs more detailed theoretical analysis of . Please refer and discuss the relationship between the proposed method and such methods.  Sugiyama, M. Borgwardt, K. M.: Rapid Distance-Based Outlier Detection via Sampling, NIPS2013.  Wu, M. and Jermaine, C.: Outlier detection by sampling with accuracy guarantees, SIGKDD22006.  Ting, K. M., Washio, T., Wells, J. R., Aryal, S., Defying the gravity of learning curve: a characteristic of nearest neighbour anomaly detectors, Machine Learning, 2016. - Important reference , which discusses performance for high-dimensional data, is missing.  Zimek, A., Schubert, E., Kriegel, H.-P., A survey on unsupervised outlier detection in high-dimensional numerical data, Statistical Analysis and Data Mining, 2012. ========== Update after Author feedback ========== Although I acknowledge the author feedback, some of my concerns (sensitivity analysis and discussion of high-dimensional data) still remain.
Anomaly detection defines a very important task in data analysis, i.e., outliers often have high semantic value, see anomalous tissue samples in histological slides that can indicate early states of cancer. Well implemented empirical studies can guide practitioners in empirical studies. The finite sample bounds also provide a theoretical basis for the comparison of nearest neighbour methods for anomaly detection. The theoretical findings are based on the assumption that the support of the normal distribution does not overlap with the outlier distribution. This assumption is often violated and I believe that Huber in his original contribution did not assume it. Often an anomaly distribution might overlap with the tail of the normal distribution. Such a situation should be discussed in the paper since it happens in practice and for these cases theory should provide guidance. originality: The paper follows the line of research in anomaly detection and puts its findings in relation to the literature. The empirical study has moderate originality since the techniques have been used elsewhere in the literature. The finite sample bounds are interesting and look original to me. quality: The presentation of the empirical results should be improved. The box plots of Fig 1 are not very informative since the different tasks have quite different complexity as can be seen from the absolute AP and AUC values. It would be much better to report on the performance differences between two methods. It could be that method A always performed better than method B given a data set. Then box plots of performance differences between two methods on various data sets might be significant while the box plots of both methods on the various data sets are highly overlapping. clarity: I didn't see a clear strategy in the study The empirical work and the theoretical findings are only very loosely connected to the empirical results. significance: I am not aware of other sample bounds for anomaly detection and therefore, I consider the results as significant. However, I might not be aware of some works on sample bounds for the studied scenario. minor comment: l95: "have significantly downgraded" -> "have been significantly downgraded"