A method is presented for validating RDF triples using LLMs with traceable arguments.
The approach avoids using internal LLM factual knowledge and instead compares verified RDF statements to external documents.
1,719 positive statements from the BioRED dataset were evaluated alongside the same number of newly generated negative statements, resulting in 88% precision and 44% recall, indicating the need for human oversight.
The method was also tested on the SNLI dataset, showing comparison with models tuned for natural language inference task.
The method was demonstrated on Wikidata using a SPARQL query to automatically retrieve statements for verification.
Results suggest that LLMs could be applied for large-scale validation of statements in knowledge graphs, reducing human annotation costs.