Join GitHub today
GitHub is home to over 28 million developers working together to host and review code, manage projects, and build software together.Sign up
Review #2 #158
The following peer review was solicited as part of the Distill review process.
The reviewer chose to keep keep anonymity. Distill offers reviewers a choice between anonymous review and offering reviews under their name. Non-anonymous review allows reviewers to get credit for the service them offer to the community.
Distill is grateful to the reviewer, for taking the time to review this article.
Conflicts of Interest: Reviewer disclosed no conflict of interest.
This article introduces a feature-wise transformation layer and especially focuses on a specific type of feature-wise layer called FiLM layer, which is proposed in Perez et al., 2018, and shows this type of layer often appears in the models used in recent literature (style transfer, question-answering, classification, etc…).
Overall the article well explains the concept of feature-wise transformation layer in its introduction. Also, the article covers a broad range of recent works using the layer and explains how those works condition the network by the feature-wise transformation layer for each task, which would help people who are going to apply the layer for their own tasks which need the integration of multiple sources.
However, in terms of the ‘understanding’ of the FiLM, I wanted to see more examples on each interpolation result in Section “Properties of FiLM-ed networks”.