QueryPIE: Backward reasoning for OWL Horst over very large knowlege bases
published: Nov. 25, 2011, recorded: October 2011, views: 4064
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Both materialization and backward-chaining as dierent modes of performing inference have complementary advantages and disadvan- tages. Materialization enables very ecient responses at query time, but at the cost of an expensive up front closure computation, which needs to be redone every time the knowledge base changes. Backward-chaining does not need such an expensive and change-sensitive precomputation, and is therefore suitable for more frequently changing knowledge bases, but has to perform more computation at query time. Materialization has been studied extensively in the recent semantic web literature, and is now available in industrial-strength systems. In this work, we focus instead on backward-chaining, and we present an hybrid algorithm to perform efficient backward-chaining reasoning on very large datasets expressed in the OWL Horst (pD*) fragment. As a proof of concept, we have implemented a prototype called QueryPIE (Query Parallel Inference Engine), and we have tested its performance on different datasets of up to 1 billion triples. Our parallel implementation greatly reduces the reasoning complexity of a naive backward-chaining approach and returns results for single query-patterns in the order of milliseconds when running on a modest 8 machine cluster. To the best of our knowledge, QueryPIE is the first reported backward- chaining reasoner for OWL Horst that efficiently scales to a billion triples.
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !