Skip to content

Review #1 #2

@phillipi

Description

@phillipi

The following peer review was solicited as part of the Distill review process. The review was formatted by the editor to help with readability.

The reviewer chose to keep anonymity. Distill offers reviewers a choice between anonymous review and offering reviews under their name. Non-anonymous review allows reviewers to get credit for the service them offer to the community.

Distill is grateful to the reviewer for taking the time to write the review.


What type of contributions does this article make?

Explanation of existing results

How significant are these contributions

4/5

Communication:

Is the article well-organized, focused and structured?

5/5

Is the article well-written? Does it avoid needless jargon?

5/5

Diagram & Interface Style

5/5

Impact of diagrams / interfaces / tools for thought?

5/5

How readable is the paper, accounting for the difficulty of the topic?

4/5

Comments on communication:

The reviewed version was still slightly unfinished (missing citations, fireboat and streetcar comparisons did not load etc.)

Scientific correctness & integrity:

Are experiments in the article well designed, and interpreted fairly?

3/5

Does the article critically evaluate its limitations? How easily would a lay person understand them?

4/5

How easy would it be to replicate (or falsify) the results?

5/5

Does the article cite relevant work?

4/5

Considering all factors, does the article exhibit strong intellectual honesty and scientific hygiene?

5/5

Comments on scientific correctness & integrity:

In certain cases, the observations are slightly exaggerated. Especially in the "new interfaces", the future of the method seem slightly over-hyped. The work shows an interesting idea of combining existing techniques for better undestarnding of hidden representations, which produce interesting qualitative observations for a curious reader. However, it only helps to understand the inner working of one particular model for image classification, and the relationship to generative models is unclear. Especially when the main message of this work seem to be that the concepts are data-dependent (not only by sub-sampling the patches, but it also clearly shows that the discriminability within the dataset is the key).

General comments:

This submission shows that visualisation of feature vectors obtained by averaging values in a discriminable reprojection of a subset of patches from the training dataset. Authors show empirically on selected examples, that these centroids have interesting semantical properties.

The execution of this submission is exceptional. It caters especially to curious readers who are intrigued by the fact that it is extremely hard to interpret the hidden image representations. However, at certain points it feels that the presentation is slightly exaggerating the results.

While the observations are really interesting, authors might more often point out that it is only cherry-picked evidence, while verifying applicability of this method would require more quantitative evaluation. It is not saying that this method might not become a useful tool for e.g. verifying adversial examples, but it is important to note that it is only one of the possible analyses.

I found especially the future work section misleading. It is not clear how these dictionaries do relate to generative models when they are established on one particular dataset. Especially when most of the results verify the built-in biases of the ILSVRC dataset. As such, it might be advisable to be more cautious in assuming possible applications.

Similarly, for the whole section which investigates the manifold, it is important to note that those are just post-hoc observations and that the curved paths are just intriguing observations on this particular projection.

All in all, I find this work of exceptional quality and some of the presented results are really interesting. However, in my opinion, it would be useful to more point out the caveats of qualitative evaluation and the dangers of purely empirical evidence in the main text. As such, I think some statements are slightly exaggerated. However, authors discuss most of the mentioned limitations in the conclusions.

Minor issues. Please note that some of the might be caused by my misunderstandings:

  • In the reviewed version, many citations were missing, and not all footnotes scripts worked. Similarly, the streetcar/fireboat examples did not load.
  • In the section “Aggregating multiple images” it is not described how the class attributions are computed. Also, the attributions for the earlier layers do not make much sense so it might be useful to describe why is that the case.
  • It is not clear whether the combination of concepts (the sand, dune and sandbar) are selected based on the attribution classes or whether based on the combination of their feature representations. It might be an interesting future direction to estimate these relations programatically?
  • The emphasis on why in the last section not formatted.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions