Hi Prerak,
interesting, good.

If you put a demo or share your code somewhere in the future, let us know.

Paolo

prerak pradhan wrote:
> Hey Paolo, 
> Sry for the late reply got quite busy here. I have been using the stanford 
> NLP library along with Stanford NER and their dependency parser. To map user 
> queries into SPARQL queries, the application am working on basically 
> recognizes proper nouns or entities in the user input and tries to use the 
> dependency parser to guess what property/relationship of that entity is the 
> user trying to search for.
> 
>  For example if the user puts in the line: 
> "Managers of Manchester United with their active date"
> The application would recognize Manchester United as a proper noun or an 
> entity and since the keyword "Manager" has a direct relationship here i.e the 
> dependency "prep_of", it would try and find their Managers with the first 
> query and than on the second run would try and add in their active date.
> The algorithm,for this is still very much under development and am currently 
> testing it for more complex queries. So lot of hair pulling and head banging 
> still to go ;)
> 
> Prerak
> 
> 
> 
> ________________________________
>  From: Paolo Castagna <castagna.li...@googlemail.com>
> To: jena-users@incubator.apache.org 
> Sent: Friday, April 6, 2012 1:30 AM
> Subject: Re: Loading DBpedia datasets
>  
> prerak pradhan wrote:
>> Thanks for the input, Paolo and yes I did use tdbloader2 on ubuntu and than 
>> transfered the directory into windows. Loading seems a lot faster in linux 
>> using tdbloader2. 
> 
> Yep.
> 
>> I am working on this mini-project which aims to develop that performs 
>> semantic search on DBpedia data, something like Kngine or Hakia on a very 
>> very small scale, I think I have got the natural language processing part 
>> done and am now trying to work on forming SPARQL queries to be run against 
>> Dbpedia dataset based on NLP output on the user entered query. Do you have 
>> any links or resource in this regard? thanks again appreciate it.
> 
> Oh, well... another "semantic search" project. :-)
> 
> I do not have more links that you would find from Wikipedia page
> on "semantic search" and the references there.
> 
> By the way, Which NLP library are you using?
> How do you generate a SPARQL query from natural language?
> 
> In a distant past, I tried to exploit the prefix:keyword pattern
> that many search engines support, and therefore people might already
> used to, to do something vaguely similar but much (much) simpler
> thing (no NLP involved):
> 
>   type:car color:blue model:touran city:london ...
> 
> It works for very simple queries, but it quickly becomes impractical.
> It would, however, be an improvement for certain type of searches.
> Even supporting just type:{book|person|city|...} can be useful.
> 
> Paolo
> 
>>
>> ________________________________
>>   From: Paolo Castagna <castagna.li...@googlemail.com>
>> To: jena-users@incubator.apache.org 
>> Sent: Thursday, April 5, 2012 8:36 AM
>> Subject: Re: Loading DBpedia datasets
>>   
>> prerak pradhan wrote:
>>> Hello, there am just starting off with Jena and am pretty new to it. I am 
>>> trying to load all the DBpedia datasets so that I can have a local version 
>>> of DBpedia working on my station here. I used the TDB loader to load the 
>>> data sets while doing so I specified a directory on which to load the 
>>> dataset.  I used the following code to query the dataset.
>>>     String directory = "c:/dataset" ;
>>>     DatasetGraphTDB dataset = TDBFactory.createDatasetGraph(directory);
>>>     Graph g1 = dataset.getDefaultGraph();
>>>     Model newModel = ModelFactory.createModelForGraph(g1);
>>>     String q= "SELECT ?p ?o WHERE { 
>>> <http://dbpedia.org/resource/Mendelian_inheritance> ?p ?o . }";
>>>     Query query = QueryFactory.create(q);
>>>     QueryExecution qexec = QueryExecutionFactory.create(query,newModel);
>>>     ResultSet results = qexec.execSelect();
>>>     while (results.hasNext()) {
>>>       QuerySolution result = results.nextSolution();
>>>       RDFNode s = result.get("s");
>>>       RDFNode p = result.get("p");
>>>       RDFNode o = result.get("o");
>>>       System.out.println( " { " + s + " " + p + " " + o + " . }");
>>> }
>>> Now my question is, the DBpedia data dumps come in various files, do I load 
>>> all these files in the same directory using TDB to create one huge model or 
>>> do I need to load it into different directories thus having to create 
>>> different models to query the data. Please not that I do not plan to load 
>>> the whole of DBpedia datasets onto the datastore just the english version 
>>> of Ontology Infobox properties, Titles and Ontology Infobox types. Forgive 
>>> me for my very amateur question but i am just getting started with it ;). 
>> Hi Prerak,
>> first of all, welcome on the Jena mailing list.
>>
>> DBPedia is one of the "not so small" RDF data dumps around, so it's better 
>> you
>> check you are using a 64 bits OS and JVM and you have a decent amount of RAM 
>> on
>> that machine. A few more details here:
>> http://incubator.apache.org/jena/documentation/tdb/jvm_64_32.html
>>
>> Then, allow me to suggest you to read about 'RDF dataset' here:
>> http://www.w3.org/TR/sparql11-query/#rdfDataset
>>
>> TDB supports RDF datasets, documentation is here:
>> http://incubator.apache.org/jena/documentation/tdb/datasets.html
>>
>> So, you can load the entire DBPedia data into a single TDB location on disk
>> (i.e. a single directory). This way, you can run SPARQL queries over it.
>> This is in my opinion the best option.
>>
>> You could use named graphs, read more about N-Quads serialization format 
>> here:
>> http://sw.deri.org/2008/07/n-quads/
>>
>> And, in relation to DBPedia, here:
>> http://wiki.dbpedia.org/Datasets#h18-18
>>
>> You might decide to create your own named graphs and load parts of DBPedia 
>> in it
>> to support your data management needs, rather than taking the named graphs
>> given to you by DBPedia to track provenance.
>>
>> Finally, with datasets of the size of DBPedia, tdbloader2 should be a better
>> choice than tdbloader, but you seems to be using Windows therefore 
>> tdbloader2 is
>> not a good choice for you. You could have a look at tdbloader3 as well, if 
>> you
>> have problems with tdbloader. But, try first with tdbloader. You can also 
>> load
>> the data of a server with a decent amount of RAM and move the files around as
>> you need.
>>
>> What are you planning to do with DBPedia loaded locally?
>>
>> I hope this helps and let me know how it goes,
>> Paolo

Reply via email to