By Anonymous User
Review Details
Reviewer has chosen to be Anonymous
Overall Impression: Average
Content:
Technical Quality of the paper: Good
Originality of the paper: Yes, but limited
Adequacy of the bibliography: Yes, but see detailed comments
Presentation:
Adequacy of the abstract: Yes
Introduction: background and motivation: Good
Organization of the paper: Needs improvement
Level of English: Satisfactory
Overall presentation: Average
Detailed Comments:
This paper provides a survey of the knowledge graph embeddings field, with some focus on the use of LLMs and perspective on semantically enriched embeddings. The authors offer some final recommendations and critically reflect on the recent results.
While I believe that some of the content presented also appears in other survey papers, I think this paper provides an interesting overview. It could be an interesting piece to have in the journal.
I still want to flag a couple of things/problems that I think can be fixed with some additional writing and restructuring.
+ The survey is, in general, very concise.
In general, if this wants to be a survey paper, I believe that the knowledge graph embedding summary could benefit from some additional details on what models actually do and a more general introduction to embeddings (e.g., how is a scoring function used to generate a prediction).
There are some papers that have not been cited (e.g., https://arxiv.org/pdf/1903.05485.pdf on multi-modal kg embeddings). I also understand that this might not be the entire focus of the paper (as some of this info can be found in other survey papers), but it also depends on how much this paper wants to be a survey paper.
If it is not possible to extend the survey due to page constraints, I'd still try to reorganize it and provide more details on the vision and the discussion part (4.3 is one of the few sections with a section-specific discussion). For example, I find the recommendations section useful, but I would again extend it and present the content in a more structured format (e.g., a figure or a table). This kind of format could help in conveying the main ideas quickly.
+ I find the paper a bit terse, as some things are briefly mentioned but not really explained. There are many instances of this problem.
"GenKGC [61] converts the KG completion task to a sequence-to-sequence (Seq2Seq) generation task. The incontext learning paradigm of GPT-3 learns correct output answers by concatenating the selected samples relevant to the input." - what are the selected samples here? In addition to this, the next sentence starts with "GenKGC similarly" but I am not sure what "similarly" refers to.
Section 5 is very short and some paragraphs would require better structure and better organization.
"ELEm has been evaluated on the Protein-Protein Interaction (PPI) dataset for the LP task. However, the successor algorithms introduced more appropriate datasets such as Gene Ontology (GO)" - why are these more appropriate and why does this make a difference?
Then, in the same section, the paper describes the results in Ruffinelli et al., which are important results in the KG embeddings evaluation, but I am not sure if this is the best way or section to introduce them. Also, since there have been some attempts at providing more uniform benchmarking utilities such as https://github.com/pykeen/pykeen, I think it might be worth mentioning these (the authors briefly introduce the paper in ref [80] but since this is an evaluation setup section I think more details could help).
+ It is sometimes unclear to me if the focus is on Knowledge Graph embeddings or LLMs.
It seems to me that LLMs appear mostly in one section, but from the introduction, I'd have expected a larger analysis and more details in the discussion section. From the title the focus should be semantically enriched embeddings, but the focus seems on more general knowledge graph embeddings to me.
Some additional references:
Additional ref and dataset for inductive link prediction: https://arxiv.org/abs/2203.01520
On multi-modality: https://arxiv.org/pdf/1903.05485.pdf
Minor comments:
RoBerta should be RoBERTa
Table 3. I think this is a useful table, as it summarizes many of the different evaluations. Maybe it would be good to add the references for the datasets?