Hi Mario, The reason I wanted the view job output is because there are multiple ways you can do forced metadata with a we connection. There's a Forced Metadata tab, a Metadata tab, and you can add a Metadata Transformer to the pipeline as well.
I will have a look at why Forced Metadata is no longer working, but I suggest that you try the other two possibilities while I do that. Thanks, Karl Sent from my Windows Phone ------------------------------ From: Bisonti Mario Sent: 8/29/2014 2:50 AM To: [email protected] Subject: R: Populate field Solr Ok, thanks. Tab Name Name:ScanPdftatankamNEW Tab Connection Stage Type Precedent Description Connection name 1. Repository ConnessioneWeb 2. Output 1. Solr Tab Forced metadata Parameter name:category Parameter value: manuale Tab Seeds http://tatankam.herobo.com/prova/sotto/ Tab Inclusions Include in crawl: .*sotto* Include in index: .*sotto* Tab Security, Metadata, Solr Field Mapping Empty I omit the scheduled Tab because I start it manually. I am using ManifoldCF 1.7 Thanks a lot for your support Mario *Da:* Karl Wright [mailto:[email protected]] *Inviato:* giovedì 28 agosto 2014 17:54 *A:* [email protected] *Oggetto:* Re: Populate field Solr Hi Mario, No metadata whatsoever is getting through to Solr. Can you cut/paste the data on the view page of your job please? View your job, and then select the output so I can see how everything is configured. Karl On Thu, Aug 28, 2014 at 11:30 AM, Bisonti Mario <[email protected]> wrote: INFO - 2014-08-28 17:26:47.372; org.apache.solr.update.processor.LogUpdateProcessor; [collection1] webapp=/solr path=/update/extract params={commitWithin=10000&literal.id= http://tatankam.herobo.com/prova/sotto/&resource.name=index.html&wt=xml&version=2.2} {add=[http://tatankam.herobo.com/prova/sotto/ (1477694830537605120)]} 0 5 INFO - 2014-08-28 17:26:48.976; org.apache.solr.update.processor.LogUpdateProcessor; [collection1] webapp=/solr path=/update/extract params={commitWithin=10000&literal.id= http://tatankam.herobo.com/prova/sotto/Using%2520the%2520various%2520optional%2520Film%2520Adapters.pdf&resource.name=Using%2520the%2520various%2520optional%2520Film%2520Adapters.pdf&wt=xml&version=2.2} {add=[ http://tatankam.herobo.com/prova/sotto/Using%20the%20various%20optional%20Film%20Adapters.pdf (1477694832220569600)]} 0 4 INFO - 2014-08-28 17:26:51.409; org.apache.solr.update.processor.LogUpdateProcessor; [collection1] webapp=/solr path=/update/extract params={commitWithin=10000&literal.id= http://tatankam.herobo.com/prova/sotto/DopoFullCrawl.pdf&resource.name=DopoFullCrawl.pdf&wt=xml&version=2.2} {add=[http://tatankam.herobo.com/prova/sotto/DopoFullCrawl.pdf (1477694834770706432)]} 0 67 INFO - 2014-08-28 17:26:51.747; org.apache.solr.update.processor.LogUpdateProcessor; [collection1] webapp=/solr path=/update/extract params={commitWithin=10000&literal.id= http://tatankam.herobo.com/prova/sotto/SAP%2520SSO%2520Authentication%2520with%2520verify.pdf&resource.name=SAP%2520SSO%2520Authentication%2520with%2520verify.pdf&wt=xml&version=2.2} {add=[ http://tatankam.herobo.com/prova/sotto/SAP%20SSO%20Authentication%20with%20verify.pdf (1477694835126173696)]} 0 58 INFO - 2014-08-28 17:26:57.372; org.apache.solr.update.DirectUpdateHandler2; start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=true,prepareCommit=false} INFO - 2014-08-28 17:26:57.377; org.apache.solr.search.SolrIndexSearcher; Opening Searcher@45d1f61c[collection1] main INFO - 2014-08-28 17:26:57.377; org.apache.solr.core.QuerySenderListener; QuerySenderListener sending requests to Searcher@45d1f61c[collection1] main{StandardDirectoryReader(segments_alc:42455:nrt _ex1(4.9):C4)} INFO - 2014-08-28 17:26:57.378; org.apache.solr.core.QuerySenderListener; QuerySenderListener done. INFO - 2014-08-28 17:26:57.378; org.apache.solr.core.SolrCore; [collection1] Registered new searcher Searcher@45d1f61c[collection1] main{StandardDirectoryReader(segments_alc:42455:nrt _ex1(4.9):C4)} INFO - 2014-08-28 17:26:57.378; org.apache.solr.update.DirectUpdateHandler2; end_commit_flush INFO - 2014-08-28 17:27:01.329; org.apache.solr.update.DirectUpdateHandler2; start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false} INFO - 2014-08-28 17:27:01.344; org.apache.solr.core.SolrDeletionPolicy; SolrDeletionPolicy.onCommit: commits: num=2 commit{dir=NRTCachingDirectory(MMapDirectory@/usr/share/solr/example/solr/collection1/data/index lockFactory=NativeFSLockFactory@/usr/share/solr/example/solr/collection1/data/index; maxCacheMB=48.0 maxMergeSizeMB=4.0),segFN=segments_alc,generation=13728} commit{dir=NRTCachingDirectory(MMapDirectory@/usr/share/solr/example/solr/collection1/data/index lockFactory=NativeFSLockFactory@/usr/share/solr/example/solr/collection1/data/index; maxCacheMB=48.0 maxMergeSizeMB=4.0),segFN=segments_ald,generation=13729} INFO - 2014-08-28 17:27:01.344; org.apache.solr.core.SolrDeletionPolicy; newest commit generation = 13729 INFO - 2014-08-28 17:27:01.345; org.apache.solr.core.SolrCore; SolrIndexSearcher has not changed - not re-opening: org.apache.solr.search.SolrIndexSearcher INFO - 2014-08-28 17:27:01.346; org.apache.solr.update.DirectUpdateHandler2; end_commit_flush INFO - 2014-08-28 17:27:01.346; org.apache.solr.update.processor.LogUpdateProcessor; [collection1] webapp=/solr path=/update/extract params={commit=true&wt=xml&version=2.2} {commit=} 0 17 *Da:* Karl Wright [mailto:[email protected]] *Inviato:* giovedì 28 agosto 2014 17:21 *A:* [email protected] *Oggetto:* Re: Populate field Solr Hi Mario, Can you post the Solr log INFO message for the indexing of the document in question? Thanks, Karl On Thu, Aug 28, 2014 at 11:18 AM, Bisonti Mario <[email protected]> wrote: Hallo. I have web repository containing pdf files. So from Manifold I scan that directory and index the output connector : solr I need to populate the field “category” of solr index. I tried to use a job on ManifoldCF to do this. Tab: Forced Metadata Parameter name: category Parameter value: manuale Buti t doesn’t work. So I don’t understand if the problemi s Tika that executes the scan of pdf documents and passes to Solr field the values not using the parameter name : category Colud you help me? Thanks a lot
