By Anonymous User
Review Details
Reviewer has chosen to be Anonymous
Overall Impression: Average
Content:
Technical Quality of the paper: Average
Originality of the paper: Yes
Adequacy of the bibliography: Yes, but see detailed comments
Presentation:
Adequacy of the abstract: Yes
Introduction: background and motivation: Good
Organization of the paper: Needs improvement
Level of English: Satisfactory
Overall presentation: Weak
Detailed Comments:
This paper discusses a very interesting topic, namely, an attempt to a neuro-symbolic approach for explanations. Moreover, it focuses on XUI, i.e., on the human-machine interaction component of explanations – which is often neglected in the literature. Finally, they aim at explaining symbolic artifacts (process models), which is also something important and massively neglected. So, I am very sympathetic to what the author are set out to achieve with this work.
That said, I think the paper has many shortcomings regarding its presentation (mostly) and organization. In the sequel, I elaborate on the most important points. Again, I am sympathetic to the general idea of the paper but I think the paper should undergo a major revision and another round of reviews before publication.
1) It is unclear what is the role of machine learning in the proposed approach. I understood that, by machine learning, the authors mean Inductive Logic Programming. That is completely fine. ILP is an useful machine learning method. However, it is unclear to me how ILP is used? Is it in Layer 2 to abstract the horn-like rules they present in the paper? In page 9, the authors write: “The system prompts the user to check and correct the calculation formula of the result node. In this case, the calculation formula r(v ) of the node v(onlineBookingEngine) must be specified in a domain-orientated (SIC. Domain-oriented would be better) manner by expanding it by adding customer data in relation to a booking request. The calculation formula must therefore be adapted as follows (additions in bold): r(vonlineBookingEngine) = ← (r(venterData) && r(vcustomerData) && r(vbookingRequest) == true”. The first sentence of this passage is ambiguous: is the user supposed to update the formula (not realistic) or does learning (formulate repair) kicks in here? Still related to this topic, sometimes in the paper, the authors talk about ML and ML experts, etc. Are they referring to ML in general, to ILP, something else?
2) There is some oddness in reconciling the results and prescriptions made in section 2 and what the authors are set out to do. Section 2 basically reduces the discussion of interaction to features – which is odd when one is primarily focused on explaining symbolic artifacts - I see that features are assembled in layer 3 but this should be much better explained. Another weakness in section 2 is that it assumes too much from the reader who is interested in NeSy systems but who is not necessarily knowledgeable of the literature of XAI. There are many technical terms used here without explanations (e.g., why-explanation, global and local explanation, counterfactual explanation, feature relevance, etc.).
3) The empirical evaluation presented is too thin and preliminary. They have interviewed six experts in a qualitative evaluation without a clear frame of comparison to alternatives. In page 13, they write: “A broader evaluation with various case studies and a larger survey participation is in preparation.” Perhaps we should wait for those results? The issue is that many of the claims of this paper are empirical claims (in p.9, the results are supposed to be “easily readable, comprehensible and understandable”).
4) Section 2 is an interesting result. However, it is frustrating that the mapping study does not look at explanation from a broader and more critical perspective. There is an immense literature on explanation that goes beyond explainable AI. According to this literature, some of these typical XAI strategies should not even be considered valid explanations, since they don't satisfy the most basic requirements for explanation. A paper that could be useful in this respect is (https://www.sciencedirect.com/science/article/pii/S0169023X24000491). In particular, because it also defends the need for explaining symbolic artifacts, as the authors are pursuing here.
5) Now, here is my main issue with the paper as it is: section 4. The explanation given there is way too succinct, cryptic, and hard to understand. The authors should spend time elaborating this (absolutely central) section of the paper. I think a running example (e.g., the one used in section 5) could help. They should better explain why this 5-layered architecture, and what really happens inside each of the layer. How is the analysis from 4 to 3 happens (algorithm 1 and its subsequent description are far from enough), how are the rules in L2 formed, verified and revised, how are the paths constructed what exactly is the algorithm here?), and so on.
6) In 5.1, given that we have two profiles here (administrators and domain experts), I was expecting to see at least an illustration of different explanations to these two profiles
7) P18: “In summary, it can be stated that different forms of explanation can support the explanatory power in the sense of comprehensible results.” -> True. However, the true discussion here shouldn’t be restricted to the concrete syntactical form of explanation presentation (e.g., visual, textural or tabular). The literature of explanation is rich on different understandings of what an explanation is. For example, one could combine pragmatic explanations with metaphorical and unificatory ones (again, the paper I mentioned above can perhaps help connecting to this literature).
5. P6: “ontologies frequently permit modeling flexibility, which enables the interpretation of modeled content and, consequently, makes the generic utilization and extension of the ontology for process analysis more challenging.” -> I honestly don’t know what the authors mean here. Ontologies are supposed to excluded unintended interpretations of content, not permit flexibility. The authors then continue “In such domain-specific cases, Noy and McGuinness propose the creation of an ontology from scratch”. I still don’t understand how this connects to the previous sentence. In any case, Noy and McGuinneess’ paper was meant as a very-basic introduction to ontologies (hence the title ‘Ontology 101’) not as a methodology for ontology engineering.
6. P9: “If, after three evaluation cycles, the value of the weighting of a result node assumes the value −1” -> Why three cycles?