By Anonymous User
Review Details
Reviewer has chosen to be Anonymous
Overall Impression: Good
Content:
Technical Quality of the paper: Excellent
Originality of the paper: Yes
Adequacy of the bibliography: Yes, but see detailed comments
Presentation:
Adequacy of the abstract: Yes
Introduction: background and motivation: Good
Organization of the paper: Satisfactory
Level of English: Satisfactory
Overall presentation: Excellent
Detailed Comments:
The paper is a survey/review of methods for explaining concepts in neural networks. These methods are classified as neural-level and layer-level explanation methods. Within these two broad categories the methods are further classified into methods that exploit similarities and causal relationships between concepts and activations, post-hoc methods such as TCAV and probing, and concept bottleneck models.
The paper is well-written and easy to read. It overviews the aforementioned methods in a clear way.
I do not have specific comments apart from pointing out the recent work in [1] that might be related.
I recommend acceptance for its publication in the inaugural special issue.
[1] Pietro Barbiero, Gabriele Ciravegna, Francesco Giannini, Mateo Espinosa Zarlenga, Lucie Charlotte Magister, Alberto Tonda, Pietro Lió, Frederic Precioso, Mateja Jamnik, and Giuseppe Marra. 2023. Interpretable neural-symbolic concept reasoning. In Proceedings of the 40th International Conference on Machine Learning (ICML'23), Vol. 202. JMLR.org, Article 76, 1801–1825.
Minor points:
- pp.2, line 41: the two probabilities P(d|E) and log P(d) -> the two probabilities P(d|E) and P(d)
- pp.5, line 42: (see Figure 3.3) -> (see Figure 4)
- pp.6, line 29: please define the operator \circ
- pp.7, line 25: paper -> paper.