In Defense of Inconsistency, On Managing Truth in a Knowledge Graph
Ontologies, Taxonomies, Data Modeling | KGC 2023
•
27m
It is well known that work in AI fails to implement diverse viewpoints and generalize requirements adequately. Results are often unreliable, inaccurate and biased. The scope of this issue has only been magnified as LLMs have gained prominence and we need to be concerned about the trajectory of how AI systems are influencing our perceptions of truth and with the social conditioning and technical processes by which AI systems instantiate truth. As systems designed for integration of heterogeneous sources into a homogeneous system, this challenge is particularly pressing for knowledge graphs. Knowledge graphs are often presented as an AI solution for increasing the understandability of complex data landscapes, facilitating reliable pipelines for sharing or integrating data in order to gain new insights. However, the commonly deployed architecture is characterized by creating a single standardized ontology with the lens of establishing a single source of truth for all data needs. Using a number of real examples, we demonstrate that this imposes a counterproductive inflexibility rendering our graphs less useful. Contrary to this approach, and based on science from fields such as cognitive science, sociology and linguistics, we argue for a perspective that allows for more flexibility in the integration of datasets. Our knowledge graph design needs to allow for a greater plurality of vocabularies and ontologies and even inconsistency while still allowing for the data integration objectives originally conceived for knowledge graphs. We present an architecture utilizing extant RDF/OWL standards and an approach that supports a more pluralistic, efficient and effective knowledge graph development.