Hi Dave,
Yes, it was a typo :)
Thanks, Thats what i’m doing right now to get around the issues
Thanks
Kamalraj
> On 10 Jan 2015, at 10:45 pm, Dave Reynolds <[email protected]> wrote:
>
> On 09/01/15 22:04, Kamalraj Jairam wrote:
>> Hello Dave and everyone
>>
>> I have run into one more issue
>>
>> I have a class “A” and “B” in my ontology for which i have added an
>> equivalent class from “Schema.org<http://Schema.org>” and “DBPedia”
>> ontologies (This is to provide external context).
>>
>> Now when i run the reasoner to inter data against my ontology (using OWLMINI
>> or OWLMICRO”), takes a very long time to produce results.
>>
>> So, i started using Pellet to reason my ontologies, but pellet doesn’t
>> reason unless i put ontology and the data in the same model .
>>
>> 1) How can i improve the speed of OWLMINI and OWLMICRO to reason DBPEDIA and
>> Schema.org<http://Schema.org>
>
> Don't think you can easily, equivalences are expensive especially for the
> rule reasoner. The only option is to cut down the fractions of the ontologies
> that you include or switch to a reasoner like Pellet.
>
>> 2) Why wouldn’t the following statement work for Pellet ?
>>
>> Reasoner reasoner = ontModelSpec.getReasoner();
>>
>> Reasoner boundReasoner = reasoner.bindSchema(ontModel);
>> infModel = ModelFactory.createInfModel(boundReasoner, model);
>>
>> my infidel does not have inferred statements if i use pellet
>
> [Aside: "infidel" was a great typo :)]
>
> Don't know, you would have to ask the Pellet folks. Perhaps bindSchema isn't
> fully supported. That would be reasonable since I doubt there's any partial
> evaluation that Pellet could do at that stage.
>
> Your alternative is to create a union model (e.g. an OntModel over the base
> model which imports the ontology, or manually create a dynamic union model of
> base and ontology). Then you can call createInfModel over that union and omit
> the step of generating a boundReasoner.
>
> Dave
>