Hi Rupert,

After last failure, I am only using language=en and it still fails.

Thanks for the timely answer. Just to double confirm, if I re-started the index 
command this am again with higher -Xmx option is it too late to run finalise 
correct? 

With best regards,
Rajan

Sent from my iPhone

> On May 26, 2015, at 7:47 AM, Rupert Westenthaler 
> <rupert.westentha...@gmail.com> wrote:
> 
> Hi Rajan
> 
>> On Mon, May 25, 2015 at 6:15 AM, Rajan Shah <raja...@gmail.com> wrote:
>> Hi Rupert,
>> 
>> Thanks for the reply.
>> 
>> As per your suggestion, I made necessary changes however it failed with
>> "OutOfMemory" errors. At present, I am running with -Xmx48g however at this
>> point it's a trial and error approach with several days effort being
>> wasted.
> 
> I guess you are getting the OutOfMemory while optimizing the Solr
> Index (right?). The README [1] explicitly notes that a high amount of
> memory is needed by exactly this step of the indexing process.
> 
> If the indexing fails at this step you can call the indexing tool with
> the `finalise` command (instead if `indexing`) (seeSTANBOL-1047 [2]
> for details). This will prevent the indexing to be repeated and only
> execute the finalization steps (optimizing the Solr Index and creating
> the freebase.solrindex.zip file).
> 
> 
>> I am just throwing out an idea, but wanted to see
>> 
>> a. Is it possible to publish set of constraints and required parameters.
>> i.e. with minimal set of entities within mappings.txt, one need to set
>> these parameters?
> 
> I do not understand this question. Do you want to filter entities
> based on their information? If so you might want to have a look at the
> `org.apache.stanbol.entityhub.indexing.core.processor.FieldValueFilter`.
> The generic RDF indexing tool as an example on how to use this
> processor to filter entities based on their rdf:type values.
> 
> See also the "Entity Filters" section of [3]
> 
>> 
>> b. Is it possible to split the file based on subject? generate smaller
>> index for each subject and merge afterwards?
> 
> Yes. You can split up the dump (by subject). Import those parts in
> different Indexing Tool instances (meaning different Jena TDB
> instances). Importing 4*500million triples to Jena TDB is supposed to
> be much faster as 1*2Billion.
> 
> If you still want to have all data in a single Entityhub Site you need
> to script the indexing process.
> 
> * call indexing for the first part
> * after this finishes link the {part1}/indexing/destination/indexes
> folder to {part2..n}/indexing/destination/indexes
> * call indexing for the 2..n parts.
> 
> As the indexing tool only adds additional information to the Solr
> Index you will get the union over all parts at the end of the process.
> All parts need to use the full incoming_links.txt file because
> otherwise the rankings would not be correct.
> 
> The "Indexing Datasets separately" section of [3] describes a similar
> trick of creating an union index over multiple datasets.
> 
> 
> best
> Rupert
> 
>> c. Work with BaseKB guys to also make it available at nominal charge?
>> 
>> d. Maybe apply some Map/Reduce - extension of idea b
>> 
>> With best regards,
>> Rajan
> 
> 
> 
> [1] 
> http://svn.apache.org/repos/asf/stanbol/trunk/entityhub/indexing/freebase/README.md
> [2] https://issues.apache.org/jira/browse/STANBOL-1047
> [3] 
> http://svn.apache.org/repos/asf/stanbol/branches/release-0.12/demos/ehealth/README.md
> 
>> 
>> 
>> 
>> On Fri, May 22, 2015 at 9:29 AM, Rupert Westenthaler <
>> rupert.westentha...@gmail.com> wrote:
>> 
>>> Hi Rajan,
>>> 
>>>> *05:11:41,851 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl - Indexed 0 items in 2059467sec (Infinityms/item):
>>> 
>>> 'You have not indexed a single entity. So something in your indexing
>>> configuration is wrong. Most likely you are not correctly building the
>>> URIs of the entities from the incoming_links.txt file. Can you provide
>>> me an example line of the 'incoming_links.txt' file and the contents
>>> of the 'iditerator.properties' file. Those specify how Entity URIs are
>>> built.
>>> 
>>> Short answers to the other questions
>>> 
>>> 
>>>> On Fri, May 22, 2015 at 2:10 PM, Rajan Shah <raja...@gmail.com> wrote:
>>>> it ran for almost 3 days and generated index.
>>> 
>>> Thats good. It means you do have now the Freebase dump in your Jena
>>> TDB triple store. You will not need to repeat this (until you want to
>>> use a newer dump. On the next call to the indexing tool it will
>>> immediately start with the indexing step.
>>> 
>>> 
>>>> 
>>>> Couple questions come to mind:
>>>> 
>>>> a. Is there any particular log/error file the process generates besides
>>>> printing out on stdout/stderr?
>>> 
>>> The indexer writes a zip archive with the IDs of all the indexed
>>> entities. Its in the indexing/destination folder.
>>> 
>>>> b. Is it a must-have to have stanbol full launcher running all the time
>>>> while indexing is going on?
>>> 
>>> No Stanbol instance is needed by the indexing process.
>>> 
>>>> c. Is it possible that, if the machine is not connected to internet for
>>>> couple minutes could cause some issues?
>>> 
>>> No Internet connectivity is needed during indexing. Only if you want
>>> to use the namespace prefix mappings of prefix.cc you need to have
>>> internet connectivity when starting the indexing tool.
>>> 
>>> best
>>> Rupert
>>> 
>>>> 
>>>> I would really appreciate, if you can shed some light on "what could be
>>>> wrong" or "potential approach to nail down this issue"? If you need, I am
>>>> happy to share any additional logs/properties.
>>>> 
>>>> With best regards,
>>>> Rajan
>>>> 
>>>> *1. Configuration changes*
>>>> 
>>>> a. set ns-prefix-state=false*
>>>> [within /indexing/config/iditerator.properties]*
>>>> b. add empty space mapping to   http://rdf.freebase.com/ns/*
>>>> [within namespaceprefix.mappings]*
>>>> c. enable bunch of properties within mappings.txt such as following
>>>> 
>>>> fb:music.artist.genre
>>>> fb:music.artist.label
>>>> fb:music.artist.album
>>>> 
>>>> *2. Contents of indexing/dist directory*
>>>> 
>>>> -rw-r--r--  108899 May 22 05:11 freebase.solrindex.zip
>>>> -rw-r--r--  3457 May 22 05:11
>>>> org.apache.stanbol.data.site.freebase-1.0.0.jar
>>>> 
>>>> *3. Contents of /tmp/freebase/indexing/resources/imported directory*
>>>> 
>>>> -rw-r--r--  1 31026810858 May 20 07:32 freebase.nt.gz
>>>> 
>>>> *4. Contents of /tmp/freebase/indexing/resources directory*
>>>> 
>>>> -rw-r--r--   1 1206745360 May 19 09:38 incoming_links.txt
>>>> 
>>>> *5. The indexer log*
>>>> 
>>>> *04:31:57,236 [Thread-3] INFO  jenatdb.RdfResourceImporter - Add:
>>>> 570,850,000 triples (Batch: 2,604 / Avg: 3,621)*
>>>> *04:32:00,727 [Thread-3] INFO  jenatdb.RdfResourceImporter - Filtered:
>>>> 2429800000 triples (80.97554853864854%)*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - -- Finish
>>>> triples data phase*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - ** Data:
>>>> 570,859,352 triples loaded in 157,619.39 seconds [Rate: 3,621.76 per
>>>> second]*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - -- Start
>>>> triples index phase*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - -- Finish
>>>> triples index phase*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - -- Finish
>>>> triples load*
>>>> *04:32:01,157 [Thread-3] INFO  jenatdb.RdfResourceImporter - **
>>> Completed:
>>>> 570,859,352 triples loaded in 157,619.39 seconds [Rate: 3,621.76 per
>>>> second]*
>>>> 04:32:56,880 [Thread-3] INFO  source.ResourceLoader -    ... moving
>>>> imported file freebase.nt.gz to imported/freebase.nt.gz
>>>> 04:32:56,883 [Thread-3] INFO  source.ResourceLoader -    - completed in
>>>> 157675 seconds
>>>> 04:32:56,883 [Thread-3] INFO  source.ResourceLoader -  > loading
>>>> '/private/tmp/freebase/indexing/resources/rdfdata/fixit.sh' ...
>>>> 04:32:56,944 [Thread-3] WARN  jenatdb.RdfResourceImporter - ignore File
>>> {}
>>>> because of unknown extension
>>>> 04:32:56,958 [Thread-3] INFO  source.ResourceLoader -    - completed in 0
>>>> seconds
>>>> 04:32:56,958 [Thread-3] INFO  source.ResourceLoader -  ... 2 files
>>> imported
>>>> in 157675 seconds
>>>> 04:32:56,958 [Thread-3] INFO  source.ResourceLoader - Loding 0 File ...
>>>> 04:32:56,958 [Thread-3] INFO  source.ResourceLoader -  ... 0 files
>>> imported
>>>> in 0 seconds
>>>> 04:32:56,971 [main] INFO  impl.IndexerImpl -  ... delete existing
>>>> IndexedEntityId file
>>>> /private/tmp/freebase/indexing/destination/indexed-entities-ids.zip
>>>> 04:32:56,982 [main] INFO  impl.IndexerImpl - Initialisation completed
>>>> 04:32:56,982 [main] INFO  impl.IndexerImpl -   ... initialisation
>>> completed
>>>> 04:32:56,982 [main] INFO  impl.IndexerImpl - start indexing ...
>>>> 04:32:56,982 [main] INFO  impl.IndexerImpl - Indexing started ...
>>>> 
>>>> 
>>>> 
>>>> 04:45:48,075 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'nsogi' valid , namespace '
>>>> http://prefix.cc/nsogi:' invalid -> mapping ignored!
>>>> 04:45:48,076 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'category' valid , namespace '
>>>> http://dbpedia.org/resource/Category:' invalid -> mapping ignored!
>>>> 04:45:48,077 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'chebi' valid , namespace '
>>>> http://bio2rdf.org/chebi:' invalid -> mapping ignored!
>>>> 04:45:48,077 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'hgnc' valid , namespace '
>>>> http://bio2rdf.org/hgnc:' invalid -> mapping ignored!
>>>> 04:45:48,077 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'dbptmpl' valid , namespace '
>>>> http://dbpedia.org/resource/Template:' invalid -> mapping ignored!
>>>> 04:45:48,077 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'dbc' valid , namespace '
>>>> http://dbpedia.org/resource/Category:' invalid -> mapping ignored!
>>>> 04:45:48,078 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'pubmed' valid , namespace '
>>>> http://bio2rdf.org/pubmed_vocabulary:' invalid -> mapping ignored!
>>>> 04:45:48,078 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'dbt' valid , namespace '
>>>> http://dbpedia.org/resource/Template:' invalid -> mapping ignored!
>>>> 04:45:48,078 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'dbrc' valid , namespace '
>>>> http://dbpedia.org/resource/Category:' invalid -> mapping ignored!
>>>> 04:45:48,078 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'call' valid , namespace '
>>>> http://webofcode.org/wfn/call:' invalid -> mapping ignored!
>>>> 04:45:48,078 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'dbcat' valid , namespace '
>>>> http://dbpedia.org/resource/Category:' invalid -> mapping ignored!
>>>> 04:45:48,084 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'affymetrix' valid , namespace '
>>>> http://bio2rdf.org/affymetrix_vocabulary:' invalid -> mapping ignored!
>>>> 04:45:48,084 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'bgcat' valid , namespace '
>>>> http://bg.dbpedia.org/resource/Категория:' invalid -> mapping ignored!
>>>> 04:45:48,084 [pool-1-thread-1] WARN  impl.NamespacePrefixProviderImpl -
>>>> Invalid Namespace Mapping: prefix 'condition' valid , namespace '
>>>> http://www.kinjal.com/condition:' invalid -> mapping ignored!
>>>> 05:11:41,836 [Indexing: Entity Source Reader Deamon] INFO
>>> impl.IndexerImpl
>>>> - Indexing: Entity Source Reader Deamon completed (sequence=0) ...
>>>> 05:11:41,838 [Indexing: Entity Source Reader Deamon] INFO
>>> impl.IndexerImpl
>>>> -  > current sequence : 0
>>>> 05:11:41,838 [Indexing: Entity Source Reader Deamon] INFO
>>> impl.IndexerImpl
>>>> -  > new sequence: 1
>>>> 05:11:41,838 [Indexing: Entity Source Reader Deamon] INFO
>>> impl.IndexerImpl
>>>> - Send end-of-queue to Deamons with Sequence 1
>>>> 05:11:41,839 [Indexing: Entity Processor Deamon] INFO  impl.IndexerImpl -
>>>> Indexing: Entity Processor Deamon completed (sequence=1) ...
>>>> 05:11:41,839 [Indexing: Entity Processor Deamon] INFO  impl.IndexerImpl -
>>>>> current sequence : 1
>>>> 05:11:41,839 [Indexing: Entity Processor Deamon] INFO  impl.IndexerImpl -
>>>>> new sequence: 2
>>>> 05:11:41,839 [Indexing: Entity Processor Deamon] INFO  impl.IndexerImpl -
>>>> Send end-of-queue to Deamons with Sequence 2
>>>> 05:11:41,839 [Indexing: Entity Perstisting Deamon] INFO
>>> impl.IndexerImpl -
>>>> Indexing: Entity Perstisting Deamon completed (sequence=2) ...
>>>> 05:11:41,839 [Indexing: Entity Perstisting Deamon] INFO
>>> impl.IndexerImpl -
>>>>> current sequence : 2
>>>> 05:11:41,839 [Indexing: Entity Perstisting Deamon] INFO
>>> impl.IndexerImpl -
>>>>> new sequence: 3
>>>> 05:11:41,839 [Indexing: Entity Perstisting Deamon] INFO
>>> impl.IndexerImpl -
>>>> Send end-of-queue to Deamons with Sequence 3
>>>> *05:11:41,851 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl - Indexed 0 items in 2059467sec (Infinityms/item):
>>>> processing:  -1.000ms/item | queue:  -1.000ms*
>>>> 05:11:41,851 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl -   - source   :  -1.000ms/item
>>>> 05:11:41,851 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl -   - processing:  -1.000ms/item
>>>> 05:11:41,851 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl -   - store     :  -1.000ms/item
>>>> 05:11:41,906 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl - Indexing: Finished Entity Logger Deamon completed
>>>> (sequence=3) ...
>>>> 05:11:41,906 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl -  > current sequence : 3
>>>> 05:11:41,906 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl -  > new sequence: 4
>>>> 05:11:41,906 [Indexing: Finished Entity Logger Deamon] INFO
>>>> impl.IndexerImpl - Send end-of-queue to Deamons with Sequence 4
>>>> 05:11:41,910 [Indexer: Entity Error Logging Daemon] INFO
>>> impl.IndexerImpl
>>>> - Indexer: Entity Error Logging Daemon completed (sequence=4) ...
>>>> 05:11:41,910 [Indexer: Entity Error Logging Daemon] INFO
>>> impl.IndexerImpl
>>>> -  > current sequence : 4
>>>> 05:11:41,910 [main] INFO  impl.IndexerImpl -   ... indexing completed
>>>> 05:11:41,910 [main] INFO  impl.IndexerImpl - start post-processing ...
>>>> 05:11:41,910 [main] INFO  impl.IndexerImpl - PostProcessing started ...
>>>> 05:11:41,910 [main] INFO  impl.IndexerImpl -   ... post-processing
>>> finished
>>>> ...
>>>> 05:11:41,911 [main] INFO  impl.IndexerImpl - start finalisation....
>>>> 
>>>> 
>>>> 
>>>> On Wed, May 20, 2015 at 8:19 AM, Rupert Westenthaler <
>>>> rupert.westentha...@gmail.com> wrote:
>>>> 
>>>>>> On Tue, May 19, 2015 at 7:04 PM, Rajan Shah <raja...@gmail.com> wrote:
>>>>>> Hi Rupert and Antonio,
>>>>>> 
>>>>>> Thanks a lot for the reply.
>>>>>> 
>>>>>> I start to follow Rupert's suggestion, however it failed again at
>>>>>> 
>>>>>> 10:56:34,152 [Thread-3] ERROR jena.riot - [line: 8722294, col: 88]
>>>>> illegal
>>>>>> escape sequence value: $ (0x24) -- Is there anyway it can be resolved
>>> for
>>>>>> the entire file?
>>>>> 
>>>>> The indexing tool uses Apache Jena. An those are Jena parsing errors.
>>>>> So the Jena Mailing lists would be the better place to look for
>>>>> answers.
>>>>> This specific issue looks like an invalid URI that is not fixed by the
>>>>> fixit script.
>>>>> 
>>>>> 
>>>>>> I requested an access to latest BaseKB bucket, as it doesn't seem to
>>> be
>>>>>> open.
>>>>>> 
>>>>>> s3cmd ls s3://basekb-now/2015-04-15-18-54/
>>>>>> --add-header="x-amz-request-payer: requester"
>>>>>> ERROR: Access to bucket 'basekb-now' was denied
>>>>>> 
>>>>>> 
>>>>>> *Couple additional questions:*
>>>>>> 
>>>>>> *1. indexing enhancements:*
>>>>>> What settings/properties one can tweak to gain most out of the
>>> indexing.
>>>>> 
>>>>> In general you do only want information as needed for your application
>>>>> case in the index.
>>>>> For EntityLinking only labels and type are required.
>>>>> Additional properties will only be used for dereferencing Entities. So
>>>>> this will depend on your application needs (your dereferencing
>>>>> configuration).
>>>>> 
>>>>> In general I try to exclude as much information as possible form the
>>>>> index to keep the size of the Solr Index as small as possible.
>>>>> 
>>>>>> a. for ex. domain specific such as Pharmaceutical, Law etc... within
>>>>>> freebase
>>>>>> b. potential optimizations to speed up the overall indexing
>>>>> 
>>>>> Most of the time will be needed to load the Freebase dump into Jena
>>>>> TDB. Even with an SSD equipped Server this will take several days.
>>>>> Assigning more RAM will speed up this process as Jena TDB can cache
>>>>> more things in RAM.
>>>>> 
>>>>> Usually it is a good Idea to cancel the indexing process after the
>>>>> importing of the RDF data has finished (and the indexing of the
>>>>> Entities has started). This is because after indexing all the RAM will
>>>>> be used by Jena TDB for caching stuff that is no longer needed in the
>>>>> read-only operations during indexing. So a fresh start can speed up
>>>>> the indexing part of the process.
>>>>> 
>>>>> Also have a look at the Freebase Indexing Tool Readme
>>>>> 
>>>>>> 
>>>>>> *2. demo:*
>>>>>> I see that, in recent github commit(s) the eHealth and other demos
>>> have
>>>>>> been commented out. How can I get demo source code and other
>>> components
>>>>> for
>>>>>> these demos. I prefer to build it myself to see the power of stanbol.
>>>>> 
>>>>> The eHealth demo is still in the 0.12 branch [1]. This is fully
>>>>> compatible to the trunk version.
>>>>> 
>>>>>> *3. custom vocabulary:*
>>>>>> Suppose, I have custom vocabulary in CSV format. Is there a preferred
>>> way
>>>>>> to upload it to Stanbol and have it recognize my entities?
>>>>> 
>>>>> Google Refine[2] with the RDF extension [3]. You can also try to use
>>>>> the (newer) Open Refine [4] with the RDF Refine 0.9.0 Alpha version
>>>>> but AFAIK this combination is not so stable and might not work at all.
>>>>> 
>>>>> * Google Refine allows you to import your CSV file.
>>>>> * Clean it up (if necessary)
>>>>> * The RDF extension allows you to map your CSV data to RDF
>>>>> * based on this mapping you can save your data as RDF
>>>>> * after that you can import the RDF data to Apache Stanbol
>>>>> 
>>>>> hope this helps
>>>>> best
>>>>> Rupert
>>>>> 
>>>>>> 
>>>>>> Thanks in advance,
>>>>>> Rajan
>>>>> 
>>>>> 
>>>>> 
>>>>> [1]
>>> http://svn.apache.org/repos/asf/stanbol/branches/release-0.12/demos/ehealth/
>>>>> [2] https://code.google.com/p/google-refine/
>>>>> [3] http://refine.deri.ie/
>>>>> [4] http://openrefine.org/
>>>>> 
>>>>>> On Tue, May 19, 2015 at 3:01 AM, Rupert Westenthaler <
>>>>>> rupert.westentha...@gmail.com> wrote:
>>>>>> 
>>>>>>> Hi Rajan,
>>>>>>> 
>>>>>>> I think this is because you named you file
>>>>>>> "freebase-rdf-latest-fixed.gz". Jena assumes RDF/XML if the RDF
>>> format
>>>>>>> is not provided by the file extension. Renaming the file to
>>>>>>> "freebase-rdf-latest-fixed.nt.gz" should fix this issue.
>>>>>>> 
>>>>>>> The suggestion of Antonio to use BaseKB is also a valid option.
>>>>>>> 
>>>>>>> best
>>>>>>> Rupert
>>>>>>> 
>>>>>>> On Tue, May 19, 2015 at 8:32 AM, Antonio David Perez Morales
>>>>>>> <ape...@zaizi.com> wrote:
>>>>>>>> Hi Rajan
>>>>>>>> 
>>>>>>>> Freebase dump contains some things that does not fit very well with
>>>>> the
>>>>>>>> indexer.
>>>>>>>> I advise you to use the dump provided by BaseKB (http://basekb.com
>>> )
>>>>>>> which
>>>>>>>> is a curated Freebase dump.
>>>>>>>> I did not have any problem indexing it using that dump.
>>>>>>>> 
>>>>>>>> Regards
>>>>>>>> 
>>>>>>>> On Mon, May 18, 2015 at 8:48 PM, Rajan Shah <raja...@gmail.com>
>>>>> wrote:
>>>>>>>> 
>>>>>>>>> Hi,
>>>>>>>>> 
>>>>>>>>> I am working on indexing Freebase data within EntityHub and
>>> observed
>>>>>>>>> following issue:
>>>>>>>>> 
>>>>>>>>> 01:06:01,547 [Thread-3] ERROR jena.riot - [line: 1, col: 7 ]
>>> Element
>>>>> or
>>>>>>>>> attribute do not match QName production:
>>> QName::=(NCName':')?NCName.
>>>>>>>>> 
>>>>>>>>> I would appreciate any help pertaining to this issue.
>>>>>>>>> 
>>>>>>>>> Thanks,
>>>>>>>>> Rajan
>>>>>>>>> 
>>>>>>>>> *Steps followed:*
>>>>>>>>> 
>>>>>>>>> *1. Initialization: *
>>>>>>>>> java -jar
>>>>>>> org.apache.stanbol.entityhub.indexing.freebase-1.0.0-SNAPSHOT.jar
>>>>>>>>> init
>>>>>>>>> 
>>>>>>>>> *2. Download the data:*
>>>>>>>>> Download data and copy it to
>>>>>>> https://developers.google.com/freebase/data
>>>>>>>>> 
>>>>>>>>> *3. Performed execution of fbrankings-uri.sh*
>>>>>>>>> It generated incoming_links.txt under resources directory as
>>> follows
>>>>>>>>> 
>>>>>>>>> 10888430 m.0kpv11
>>>>>>>>> 3741261 m.019h
>>>>>>>>> 2667858 m.0775xx5
>>>>>>>>> 2667804 m.0775xvm
>>>>>>>>> 1875352 m.01xryvm
>>>>>>>>> 1739262 m.05zppz
>>>>>>>>> 1369590 m.01xrzlb
>>>>>>>>> 
>>>>>>>>> *4. Performed execution of fixit script*
>>>>>>>>> 
>>>>>>>>> gunzip -c ${FB_DUMP} | fixit | gzip > ${FB_DUMP_fixed}
>>>>>>>>> 
>>>>>>>>> *5. Rename the fixed file to freebase.rdf.gz and copy it *
>>>>>>>>> to indexing/resources/rdfdata
>>>>>>>>> 
>>>>>>>>> *6. config/iditer.properties file has following setting*
>>>>>>>>> #id-namespace=http://freebase.com/
>>>>>>>>> ns-prefix-state=false
>>>>>>>>> 
>>>>>>>>> *7. Performed run of following command:*
>>>>>>>>> java -jar -Xmx32g
>>>>>>>>> org.apache.stanbol.entityhub.indexing.freebase-1.0.0-SNAPSHOT.jar
>>>>> index
>>>>>>>>> 
>>>>>>>>> The error dump on stdout is as follows:
>>>>>>>>> 
>>>>>>>>> 01:37:32,884 [Thread-0] INFO
>>> solryard.SolrYardIndexingDestination -
>>>>>>> ...
>>>>>>>>> copy Solr Configuration form
>>>>>>> /private/tmp/freebase/indexing/config/freebase
>>>>>>>>> to
>>>>> /private/tmp/freebase/indexing/destination/indexes/default/freebase
>>>>>>>>> 01:37:32,895 [Thread-3] INFO  jenatdb.RdfResourceImporter -     -
>>>>> bulk
>>>>>>>>> loading File freebase.rdf.gz using Format Lang:RDF/XML
>>>>>>>>> 01:37:32,896 [Thread-3] INFO  jenatdb.RdfResourceImporter - --
>>> Start
>>>>>>>>> triples data phase
>>>>>>>>> 01:37:32,896 [Thread-3] INFO  jenatdb.RdfResourceImporter - **
>>> Load
>>>>>>> empty
>>>>>>>>> triples table
>>>>>>>>> *01:37:32,948 [Thread-3] ERROR jena.riot - [line: 1, col: 7 ]
>>>>> Element or
>>>>>>>>> attribute do not match QName production:
>>> QName::=(NCName':')?NCName.*
>>>>>>>>> 01:37:32,948 [Thread-3] INFO  jenatdb.RdfResourceImporter - --
>>> Finish
>>>>>>>>> triples data phase
>>>>>>>>> 01:37:32,948 [Thread-3] INFO  jenatdb.RdfResourceImporter - --
>>> Finish
>>>>>>>>> triples load
>>>>>>>>> 01:37:32,960 [Thread-3] INFO  source.ResourceLoader - Ignore Error
>>>>> for
>>>>>>> File
>>>>>>>>> /private/tmp/freebase/indexing/resources/rdfdata/freebase.rdf.gz
>>> and
>>>>>>>>> continue
>>>>>>>>> 
>>>>>>>>> Additional Reference Point:
>>>>>>>>> 
>>>>>>>>> *Original Freebase dump size:*  31025015397 May 14 18:10
>>>>>>>>> freebase-rdf-latest.gz
>>>>>>>>> *Fixed Freebase dump size:* 31026818367 May 15 12:45
>>>>>>>>> freebase-rdf-latest-fixed.gz
>>>>>>>>> *Incoming Links size: *1206745360 May 17 00:42 incoming_links.txt
>>>>>>>> 
>>>>>>>> --
>>>>>>>> 
>>>>>>>> ------------------------------
>>>>>>>> This message should be regarded as confidential. If you have
>>> received
>>>>>>> this
>>>>>>>> email in error please notify the sender and destroy it immediately.
>>>>>>>> Statements of intent shall only become binding when confirmed in
>>> hard
>>>>>>> copy
>>>>>>>> by an authorised signatory.
>>>>>>>> 
>>>>>>>> Zaizi Ltd is registered in England and Wales with the registration
>>>>> number
>>>>>>>> 6440931. The Registered Office is Brook House, 229 Shepherds Bush
>>>>> Road,
>>>>>>>> London W6 7AN.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> --
>>>>>>> | Rupert Westenthaler             rupert.westentha...@gmail.com
>>>>>>> | Bodenlehenstraße 11                              ++43-699-11108907
>>>>>>> | A-5500 Bischofshofen
>>>>>>> | REDLINK.CO
>>> ..........................................................................
>>>>>>> | http://redlink.co/
>>>>> 
>>>>> 
>>>>> 
>>>>> --
>>>>> | Rupert Westenthaler             rupert.westentha...@gmail.com
>>>>> | Bodenlehenstraße 11                              ++43-699-11108907
>>>>> | A-5500 Bischofshofen
>>>>> | REDLINK.CO
>>> ..........................................................................
>>>>> | http://redlink.co/
>>> 
>>> 
>>> 
>>> --
>>> | Rupert Westenthaler             rupert.westentha...@gmail.com
>>> | Bodenlehenstraße 11                              ++43-699-11108907
>>> | A-5500 Bischofshofen
>>> | REDLINK.CO
>>> ..........................................................................
>>> | http://redlink.co/
> 
> 
> 
> -- 
> | Rupert Westenthaler             rupert.westentha...@gmail.com
> | Bodenlehenstraße 11                              ++43-699-11108907
> | A-5500 Bischofshofen
> | REDLINK.CO 
> ..........................................................................
> | http://redlink.co/

Reply via email to