setup is probably reasonable.
>
> Can you give us details about your pipeline? What repository connector is
> this?
>
> Karl
>
>
> On Thu, Jan 25, 2018 at 1:49 AM, Nikita Ahuja <nik...@smartshore.nl>
> wrote:
>
>> Hello,
>>
>> I am using
> connector if you want ACLs to be generated. The FileSystem connector does
> not generate ACLs.
>
> Karl
>
>
> On Thu, Jan 25, 2018 at 3:56 AM, Nikita Ahuja <nik...@smartshore.nl>
> wrote:
>
>> Hi Karl,
>>
>>
>> I have tried for Elasti
Hi Karl,
I have created a job to fetch the data in output connector of
"ElasticSearch" and it is returning the data in base_64 encoded format
which is not readable or searchable, like in the image.
[image: Inline image 1]
Also, I should mention that I am using Tika Transformation connector ,
Hi Karl,
I am trying to connect through api and fetch the data inside that, but
there are many issues while creating the conncetor also the connector never
stops running. Will you please provide any example for the Generic API
connection.
;>
>>> Hi Nikita,
>>>
>>> I do not understand your question.
>>>
>>> The Generic Connector was written by a committer who has since become
>>> unavailable, and nobody here knows how it is supposed to work. All that we
>>> have is the
at we
> have is the code and the documentation.
>
> Karl
>
>
> On Thu, Feb 15, 2018 at 5:58 AM, Nikita Ahuja <nik...@smartshore.nl>
> wrote:
>
>> Hi Karl,
>>
>>
>> I am trying to connect through api and fetch the data inside that, but
>>
there is a coding error in it that is responsible.
>
> Karl
>
>
> On Mon, Aug 20, 2018 at 8:49 AM Nikita Ahuja wrote:
>
>> Hi Karl,
>>
>> Thanks for reply.
>>
>> I am using in the same sequence. The allowed document is added first and
>> then
it's there but you are not
> sending in the actual document length.
>
> Karl
>
>
>
>
> On Tue, Aug 28, 2018 at 8:10 AM Nikita Ahuja wrote:
>
>> Hi Karl,
>>
>> These methods are already in use with the connector in the code where
>> file is need to read a
indexable.
> */
> public boolean checkLengthIndexable(long length)
> throws ManifoldCFException, ServiceInterruption;
>
> Your connector should call this and honor the response.
>
> Thanks,
> Karl
>
>
>
> On Fri, Aug 24, 2018 at 9:55 AM Nikita Ahuja wrote:
>
&
and see why it is rejecting your documents.
>
> All of our shipping connectors use this logic and it does work, so I am
> rather certain that the problem is in your connector.
>
> Thanks,
> Karl
>
>
> On Tue, Aug 28, 2018 at 8:54 AM Nikita Ahuja wrote:
>
>>
Hi Karl,
There is a custom job executing for Aconex in the ManifoldCF environment.
But while executing it is not able to crawl complete set of documents. It
crashes in the middle of the execution.
Also it is not compatible with the Allowed Documents and Metadata Adjuster
Connectors.
The custom
erefore should be placed in the
> pipeline before the Tika Extractor.
>
> "Also it is not compatible with the Allowed Documents and Metadata
> Adjuster Connectors."
>
> This is a huge red flag. Why not?
>
> Karl
>
>
> On Mon, Aug 20, 2018 at 6:47 AM Nikit
Hello,
I am executing a job to crawl Sharepoint 2013 data using ManifoldCF
.
I am able to crawl the data from library and get ingested it in Elastic
Search index. But when the case comes for a infopath form stored in
Sharepoint 2013 library it generates following error:
*Manifoldcf Error:
crawl. Sounds like infopath forms are one such data type.
>
> It's not clear that crawling a form is a good idea in any case. What
> content do you expect this to yield?
>
> Karl
>
>
> On Fri, Jul 6, 2018 at 7:59 AM Nikita Ahuja wrote:
>
>> Hello,
>>
>>
&g
Hi Karl,
I am trying to ingest the data from website ans Sharepoint to Elastic
Search output in different pipelines in same index.
But the ManifoldCF is not able to ingest all the data. It only put image
files present in the source to ElasticSearch output.
Is there anything which is being
Wri/manifoldcfinaction/tree/master/pdfs
>
> Karl
>
> On Mon, Apr 16, 2018 at 4:48 AM, Nikita Ahuja <nik...@smartshore.nl>
> wrote:
>
>> Hi,
>>
>> I need to connect to Aconex API and fetch the data from that in
>> ManifoldCF and ingest the documents in t
re what you mean
>> beyond that. Tika's output is also completely represented in the extracted
>> metadata from the document, so if it does language detection, it will be in
>> there.
>>
>> Karl
>>
>>
>> On Tue, Apr 17, 2018 at 3:08 AM, Nikita Ahuj
Hi,
I need to connect to Aconex API and fetch the data from that in ManifoldCF
and ingest the documents in the Elastic Search.
Please suggest a way out for this.
Thanks and Regards,
Nikita
tent. I suspect you would need to run the output
> of the Tika translator into the NullOutputConnector in order to see what
> attributes are being generated to know for sure.
>
> Karl
>
>
> On Wed, Nov 21, 2018 at 4:45 AM Nikita Ahuja wrote:
>
>> HI All,
>>
>> T
Hi,
I have query related to detect the language of the records/data which is
going to be ingest in the Output Connector.
OpenNLP connector is not working for the detection as per the user
documentation, but this is not working appropriately. Please suggest is NLP
has to be used if yes, then how
ManifoldCF transformer. It
>> uses a specific directory to deliver the models that OpenNLP uses to match
>> and extract content from documents. Thus, you can provide any models you
>> want that are compatible with the OpenNLP version we're including.
>>
>> Can you desc
Hi Mates,
I am executing a Web connector job with the ManifoldCF version 2.12 on the
linux server, with given specification:
Repository Connector: Web
Transformation: Tika Extractor
Transformation: Metadata Adjuster
Output: ElasticSearch
But the issue is after crawling some records, I am
Hi,
I am executing a web connector job on MAnifoldCF 2.12 and ingesting the
records in ElasticSearch.
It was running fine earlier. But from past few days it is getting paused
after 2 minutes of starting and to restart the job I have to start the
process again.
The Postgresql is used as DB.
But
Hi Bisonti,
Will you please explain other configurations of the system which you are
using and number of worker threads and Max DB connections or any other
related parameters?
Actually I am also facing similar type of issue of slow crawling the job.
That will be really helpful for me.
Thanks And
24 matches
Mail list logo