Error in Dataimport without reason or log

2019-08-27 Thread Daniel Carrasco
Hello,

I write because I'm having problems importing some data from a MariaDB
database to my Solr Cloud cluster, and I'm not able to see the data or
where's the import problem.
My Solr has a dataimport that query a MariaDB database and index the data,
but seems to be not working.
When the dataimport ends, I see a message saying that was OK (in green
letters) with this message:
*Indexing completed. Added/Updated: 10 documents. Deleted 0 documents.*

But after activate the debug checkbox I've seen an error at end:
"Full Import failed": "2019-08-27 11:32:27"

So looks like the import is not right.
I've tried to look at the Solr logs but I only get GC messages, so I don't
know what else I can do to debug the problem.

Someone can help me to debug the problem?

I'm using the Solr version 7.7.2 (7.7.2
d4c30fc2856154f2c1fefc589eb7cd070a415b94 - janhoy - 2019-05-28 23:37:48),
I've 11 nodes in NTR and all looks healthy.

Thanks, and greetings.

-- 
_____

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Open file limit warning when starting solr

2018-12-14 Thread Daniel Carrasco
Hello,

How did you installed Solr?, have you followed this instructions?:
https://lucene.apache.org/solr/guide/7_0/taking-solr-to-production.html#taking-solr-to-production

On that instructions you first extracts an script file from inside the
tar.gz. Then running that script file with a few options, it install the
required files, creates the solr user and folders and starts (or not if you
use that option) the solr daemon from /etc/init.d/solr.
After all, you'll have a solr user, a file in /etc/init.d/solr that runs
the daemon at boot, and a file in /etc/default/solr (i don't rememeber the
entire name), that is for configure the daemon.

I'm not fully sure if is different on Ubuntu, but showld be similar.

Greetings!

El vie., 14 dic. 2018 a las 14:47, Armon, Rony () escribió:

> I don’t have a file named solr n etc/init.d  and I followed the
> instructions in the link that you sent.
> Should I uninstall and re-install?
>
> -Original Message-
> From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> Sent: Wednesday, December 12, 2018 5:45 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Open file limit warning when starting solr
>
> Hello,
>
> Strange... Solr user is created during the installation... What user is
> your Solr running?
>
> > cat /etc/init.d/solr |grep -i "RUNAS="
> >
>
> Have you followed all the info in the link I've sent?, because they talk
> also about this:
>
> You also need to edit /etc/pam.d/common-session* and add the following
> line to the end:
>
> session required
> https://urldefense.proofpoint.com/v2/url?u=http-3A__pam-5Flimits.so=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=p6Cv_eEgtezTZl9gDqIBoHVoL8I6zsy6C97uH4KauOE=KT98Xa93AgxUFtCr1pR_abkYgq3cViJfpad6bbvnmfY=
>
> I've not done this on my Debian, but maybe Ubuntu need it.
>
> Greetings!.
>
> El mié., 12 dic. 2018 a las 16:30, Armon, Rony ()
> escribió:
>
> > Probably not solr:
> > rony@rony-VirtualBox:~$ sudo su -
> > root@rony-VirtualBox:~# su solr -
> > No passwd entry for user 'solr'
> >
> > I tried the solution he suggested placing the following in limits.conf
> > root soft nofile 65000 root hard nofile 65000
> >
> > And then with the asterix:
> > *   soft nofile 65000
> > *  hard nofile 65000
> >
> > The result is the same
> >
> > -Original Message-
> > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > Sent: Wednesday, December 12, 2018 5:00 PM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Open file limit warning when starting solr
> >
> > Hello,
> >
> > I mean change to solr user using su as sudo. For your system will be
> > something like:
> >
> > $ sudo su -
> > pasword
> > # su solr -
> > $ ulimit -n
> > ...a file limit number...
> >
> > Your file limit in Ubuntu is fine, so looks like a problem with file
> > limit for that user, that's why I ask you if the user you've using for
> > run the Solr daemon is "solr" instead "root".
> >
> > Take a look at this:
> >
> > https://urldefense.proofpoint.com/v2/url?u=https-3A__askubuntu.com_que
> > stions_162229_how-2Ddo-2Di-2Dincrease-2Dthe-2Dopen-2Dfiles-2Dlimit-2Df
> > or-2Da-2Dnon-2Droot-2Duser=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOz
> > vVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=VbOekCwtf82rMzNmRFMfBCbyWyvenum
> > eFABhdfeKkcM=ZyP8I32R-VUBIINch0q8HuGBrRFFxvTSC72UP4RB6mA=
> >
> > He is changing the limit to 4096 for all users using an asterisk, but
> > you can put solr instead asterisk to change the limit to solr user
> > only, just like my first message.
> >
> > Greetings!
> >
> >
> >
> > El mié., 12 dic. 2018 a las 15:47, Armon, Rony ()
> > escribió:
> >
> > > Tried it as well...
> > >
> > > rony@rony-VirtualBox:~$ sudo su -
> > > [sudo] password for rony:
> > > root@rony-VirtualBox:~# sysctl -a|grep -i fs.file-max fs.file-max =
> > > 810202
> > >
> > > -Original Message-
> > > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > > Sent: Wednesday, December 12, 2018 4:04 PM
> > > To: solr-user@lucene.apache.org
> > > Subject: Re: Open file limit warning when starting solr
> > >
> > > Hello,
> > >
> > > The *su solr* command is important, because you change to Solr user
> > > before check the limits again, then it shows its limits. Are you
> > > running the daemon as solr user?
> > >
> > > Other command to check is:
> > >
> > > > # sysctl -a|grep

Re: Open file limit warning when starting solr

2018-12-12 Thread Daniel Carrasco
Hello,

Strange... Solr user is created during the installation... What user is
your Solr running?

> cat /etc/init.d/solr |grep -i "RUNAS="
>

Have you followed all the info in the link I've sent?, because they talk
also about this:

You also need to edit /etc/pam.d/common-session* and add the following line
to the end:

session required pam_limits.so

I've not done this on my Debian, but maybe Ubuntu need it.

Greetings!.

El mié., 12 dic. 2018 a las 16:30, Armon, Rony () escribió:

> Probably not solr:
> rony@rony-VirtualBox:~$ sudo su -
> root@rony-VirtualBox:~# su solr -
> No passwd entry for user 'solr'
>
> I tried the solution he suggested placing the following in limits.conf
> root soft nofile 65000
> root hard nofile 65000
>
> And then with the asterix:
> *   soft nofile 65000
> *  hard nofile 65000
>
> The result is the same
>
> -Original Message-
> From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> Sent: Wednesday, December 12, 2018 5:00 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Open file limit warning when starting solr
>
> Hello,
>
> I mean change to solr user using su as sudo. For your system will be
> something like:
>
> $ sudo su -
> pasword
> # su solr -
> $ ulimit -n
> ...a file limit number...
>
> Your file limit in Ubuntu is fine, so looks like a problem with file limit
> for that user, that's why I ask you if the user you've using for run the
> Solr daemon is "solr" instead "root".
>
> Take a look at this:
>
> https://urldefense.proofpoint.com/v2/url?u=https-3A__askubuntu.com_questions_162229_how-2Ddo-2Di-2Dincrease-2Dthe-2Dopen-2Dfiles-2Dlimit-2Dfor-2Da-2Dnon-2Droot-2Duser=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=VbOekCwtf82rMzNmRFMfBCbyWyvenumeFABhdfeKkcM=ZyP8I32R-VUBIINch0q8HuGBrRFFxvTSC72UP4RB6mA=
>
> He is changing the limit to 4096 for all users using an asterisk, but you
> can put solr instead asterisk to change the limit to solr user only, just
> like my first message.
>
> Greetings!
>
>
>
> El mié., 12 dic. 2018 a las 15:47, Armon, Rony ()
> escribió:
>
> > Tried it as well...
> >
> > rony@rony-VirtualBox:~$ sudo su -
> > [sudo] password for rony:
> > root@rony-VirtualBox:~# sysctl -a|grep -i fs.file-max fs.file-max =
> > 810202
> >
> > -Original Message-
> > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > Sent: Wednesday, December 12, 2018 4:04 PM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Open file limit warning when starting solr
> >
> > Hello,
> >
> > The *su solr* command is important, because you change to Solr user
> > before check the limits again, then it shows its limits. Are you
> > running the daemon as solr user?
> >
> > Other command to check is:
> >
> > > # sysctl -a|grep -i fs.file-max
> > > fs.file-max = 6167826
> >
> >
> > If is low then you may increase it.
> >
> > Greetings!
> >
> > El mié., 12 dic. 2018 a las 14:44, Armon, Rony ()
> > escribió:
> >
> > > rony@rony-VirtualBox:~$ ulimit -n
> > > 1024
> > > rony@rony-VirtualBox:~/solr-7.5.0$ ulimit -n
> > > 1024
> > >
> > > -Original Message-
> > > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > > Sent: Wednesday, December 12, 2018 3:31 PM
> > > To: solr-user@lucene.apache.org
> > > Subject: Re: Open file limit warning when starting solr
> > >
> > > Hello,
> > >
> > > What output you get with this commands?:
> > >
> > > > root@solr-temp01:/# ulimit -n
> > > > 1024
> > > > root@solr-temp01:/# su solr
> > > > solr@solr-temp01:/$ ulimit -n
> > > > 65000
> > >
> > >
> > > Greetings!
> > >
> > > El mié., 12 dic. 2018 a las 12:53, Armon, Rony ()
> > > escribió:
> > >
> > > > Hi Daniel and thanks for the prompt reply. I tried that but I'm
> > > > still getting the file limit warning.
> > > >
> > > > -Original Message-
> > > > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > > > Sent: Wednesday, December 12, 2018 12:14 PM
> > > > To: solr-user@lucene.apache.org
> > > > Subject: Re: Open file limit warning when starting solr
> > > >
> > > > Hello,
> > > >
> > > > Try creating a file in /etc/security/limits.d/solr.conf with this:
> > > > solr softnofile  65000
> > > 

Re: Open file limit warning when starting solr

2018-12-12 Thread Daniel Carrasco
Hello,

I mean change to solr user using su as sudo. For your system will be
something like:

$ sudo su -
pasword
# su solr -
$ ulimit -n
...a file limit number...

Your file limit in Ubuntu is fine, so looks like a problem with file limit
for that user, that's why I ask you if the user you've using for run the
Solr daemon is "solr" instead "root".

Take a look at this:
https://askubuntu.com/questions/162229/how-do-i-increase-the-open-files-limit-for-a-non-root-user

He is changing the limit to 4096 for all users using an asterisk, but you
can put solr instead asterisk to change the limit to solr user only, just
like my first message.

Greetings!



El mié., 12 dic. 2018 a las 15:47, Armon, Rony () escribió:

> Tried it as well...
>
> rony@rony-VirtualBox:~$ sudo su -
> [sudo] password for rony:
> root@rony-VirtualBox:~# sysctl -a|grep -i fs.file-max
> fs.file-max = 810202
>
> -Original Message-
> From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> Sent: Wednesday, December 12, 2018 4:04 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Open file limit warning when starting solr
>
> Hello,
>
> The *su solr* command is important, because you change to Solr user before
> check the limits again, then it shows its limits. Are you running the
> daemon as solr user?
>
> Other command to check is:
>
> > # sysctl -a|grep -i fs.file-max
> > fs.file-max = 6167826
>
>
> If is low then you may increase it.
>
> Greetings!
>
> El mié., 12 dic. 2018 a las 14:44, Armon, Rony ()
> escribió:
>
> > rony@rony-VirtualBox:~$ ulimit -n
> > 1024
> > rony@rony-VirtualBox:~/solr-7.5.0$ ulimit -n
> > 1024
> >
> > -Original Message-
> > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > Sent: Wednesday, December 12, 2018 3:31 PM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Open file limit warning when starting solr
> >
> > Hello,
> >
> > What output you get with this commands?:
> >
> > > root@solr-temp01:/# ulimit -n
> > > 1024
> > > root@solr-temp01:/# su solr
> > > solr@solr-temp01:/$ ulimit -n
> > > 65000
> >
> >
> > Greetings!
> >
> > El mié., 12 dic. 2018 a las 12:53, Armon, Rony ()
> > escribió:
> >
> > > Hi Daniel and thanks for the prompt reply. I tried that but I'm
> > > still getting the file limit warning.
> > >
> > > -Original Message-
> > > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > > Sent: Wednesday, December 12, 2018 12:14 PM
> > > To: solr-user@lucene.apache.org
> > > Subject: Re: Open file limit warning when starting solr
> > >
> > > Hello,
> > >
> > > Try creating a file in /etc/security/limits.d/solr.conf with this:
> > > solr softnofile  65000
> > > solr hardnofile  65000
> > > solr softnproc   65000
> > > solr hardnproc   65000
> > >
> > > This worked for me on Debian 9.
> > >
> > > Greetings!
> > >
> > > El mié., 12 dic. 2018 a las 11:09, Armon, Rony ()
> > > escribió:
> > >
> > > > Hello, When launching solr (Ubuntu 16.04) I'm getting:
> > > > *   [WARN] *** Your open file limit is currently 1024.
> > > >It should be set to 65000 to avoid operational
> > > > disruption.
> > > >If you no longer wish to see this warning, set
> > > > SOLR_ULIMIT_CHECKS to false in your profile or
> > > https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwI
> > > Fa
> > > Q=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP
> > > 0K
> > > Fw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfL
> > > L8
> > > I599seXXPaNeEmSZuUSmKTEo=
> > > > *   [WARN] ***  Your Max Processes Limit is currently 15058.
> > > >  It should be set to 65000 to avoid operational disruption.
> > > >  If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS
> > > > to false in your profile or
> > > > https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=D
> > > > wI
> > > > Fa
> > > > Q=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37
> > > > tP
> > > > 0K
> > > > Fw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCR
> > > > fL
> > > > L8
> > > > I599seXXPaNeEmSZuUSmKTEo=
>

Re: Open file limit warning when starting solr

2018-12-12 Thread Daniel Carrasco
Hello,

The *su solr* command is important, because you change to Solr user before
check the limits again, then it shows its limits. Are you running the
daemon as solr user?

Other command to check is:

> # sysctl -a|grep -i fs.file-max
> fs.file-max = 6167826


If is low then you may increase it.

Greetings!

El mié., 12 dic. 2018 a las 14:44, Armon, Rony () escribió:

> rony@rony-VirtualBox:~$ ulimit -n
> 1024
> rony@rony-VirtualBox:~/solr-7.5.0$ ulimit -n
> 1024
>
> -Original Message-
> From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> Sent: Wednesday, December 12, 2018 3:31 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Open file limit warning when starting solr
>
> Hello,
>
> What output you get with this commands?:
>
> > root@solr-temp01:/# ulimit -n
> > 1024
> > root@solr-temp01:/# su solr
> > solr@solr-temp01:/$ ulimit -n
> > 65000
>
>
> Greetings!
>
> El mié., 12 dic. 2018 a las 12:53, Armon, Rony ()
> escribió:
>
> > Hi Daniel and thanks for the prompt reply. I tried that but I'm still
> > getting the file limit warning.
> >
> > -Original Message-
> > From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> > Sent: Wednesday, December 12, 2018 12:14 PM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Open file limit warning when starting solr
> >
> > Hello,
> >
> > Try creating a file in /etc/security/limits.d/solr.conf with this:
> > solr softnofile  65000
> > solr hardnofile  65000
> > solr softnproc   65000
> > solr hardnproc   65000
> >
> > This worked for me on Debian 9.
> >
> > Greetings!
> >
> > El mié., 12 dic. 2018 a las 11:09, Armon, Rony ()
> > escribió:
> >
> > > Hello, When launching solr (Ubuntu 16.04) I'm getting:
> > > *   [WARN] *** Your open file limit is currently 1024.
> > >It should be set to 65000 to avoid operational
> > > disruption.
> > >If you no longer wish to see this warning, set
> > > SOLR_ULIMIT_CHECKS to false in your profile or
> > https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwIFa
> > Q=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0K
> > Fw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfLL8
> > I599seXXPaNeEmSZuUSmKTEo=
> > > *   [WARN] ***  Your Max Processes Limit is currently 15058.
> > >  It should be set to 65000 to avoid operational disruption.
> > >  If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS
> > > to false in your profile or
> > > https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwI
> > > Fa
> > > Q=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP
> > > 0K
> > > Fw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfL
> > > L8
> > > I599seXXPaNeEmSZuUSmKTEo=
> > >
> > > This appears to be related to a known bug in Ubuntu<
> > > https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.o
> > > rg
> > > _jira_browse_SOLR-2D13063=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvO
> > > zv
> > > Vop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3a
> > > J7 7UmLZcV0SIg=cCjwFj8bcVvmcTIcm0Rj2GNiW7wWaOeKufHSSdPd_uY=>
> > > https://urldefense.proofpoint.com/v2/url?u=https-3A__blog.jayway.com
> > > _2
> > > 012_02_11_how-2Dto-2Dreally-2Dfix-2Dthe-2Dtoo-2Dmany-2Dopen-2Dfiles-
> > > 2D
> > > problem-2Dfor-2Dtomcat-2Din-2Dubuntu_=DwIFaQ=0TzQCy9lgR5hSW-bDg5
> > > HA
> > > 76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kO
> > > n8
> > > BXuwT2pq3aJ77UmLZcV0SIg=o2V4fE_fqaIbrjbmNR6_fsyMlwSXxnUqLeXJ9BJOtq
> > > 8& e= I was wondering if you have some workaround. I followed the
> > > solutions in the following threads:
> > > https://urldefense.proofpoint.com/v2/url?u=https-3A__vufind.org_jira
> > > _b
> > > rowse_VUFIND-2D1290=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5
> > > GM
> > > 3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmL
> > > Zc V0SIg=eKl-Z7ZRS2z4218azygyFqhq9frW1NoxP1tQPyNsSNA=
> > >
> > > https://urldefense.proofpoint.com/v2/url?u=https-3A__underyx.me_2015
> > > _0
> > > 5_18_raising-2Dthe-2Dmaximum-2Dnumber-2Dof-2Dfile-2Ddescriptors=Dw
> > > IF
> > > aQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37t
> > > P0
> > > 

Re: Open file limit warning when starting solr

2018-12-12 Thread Daniel Carrasco
Hello,

What output you get with this commands?:

> root@solr-temp01:/# ulimit -n
> 1024
> root@solr-temp01:/# su solr
> solr@solr-temp01:/$ ulimit -n
> 65000


Greetings!

El mié., 12 dic. 2018 a las 12:53, Armon, Rony () escribió:

> Hi Daniel and thanks for the prompt reply. I tried that but I'm still
> getting the file limit warning.
>
> -Original Message-----
> From: Daniel Carrasco [mailto:d.carra...@i2tic.com]
> Sent: Wednesday, December 12, 2018 12:14 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Open file limit warning when starting solr
>
> Hello,
>
> Try creating a file in /etc/security/limits.d/solr.conf with this:
> solr softnofile  65000
> solr hardnofile  65000
> solr softnproc   65000
> solr hardnproc   65000
>
> This worked for me on Debian 9.
>
> Greetings!
>
> El mié., 12 dic. 2018 a las 11:09, Armon, Rony ()
> escribió:
>
> > Hello, When launching solr (Ubuntu 16.04) I'm getting:
> > *   [WARN] *** Your open file limit is currently 1024.
> >It should be set to 65000 to avoid operational
> > disruption.
> >If you no longer wish to see this warning, set
> > SOLR_ULIMIT_CHECKS to false in your profile or
> https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfLL8I599seXXPaNeEmSZuUSmKTEo=
> > *   [WARN] ***  Your Max Processes Limit is currently 15058.
> >  It should be set to 65000 to avoid operational disruption.
> >  If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS to
> > false in your profile or
> > https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwIFa
> > Q=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0K
> > Fw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfLL8
> > I599seXXPaNeEmSZuUSmKTEo=
> >
> > This appears to be related to a known bug in Ubuntu<
> > https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org
> > _jira_browse_SOLR-2D13063=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzv
> > Vop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ7
> > 7UmLZcV0SIg=cCjwFj8bcVvmcTIcm0Rj2GNiW7wWaOeKufHSSdPd_uY=>
> > https://urldefense.proofpoint.com/v2/url?u=https-3A__blog.jayway.com_2
> > 012_02_11_how-2Dto-2Dreally-2Dfix-2Dthe-2Dtoo-2Dmany-2Dopen-2Dfiles-2D
> > problem-2Dfor-2Dtomcat-2Din-2Dubuntu_=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA
> > 76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8
> > BXuwT2pq3aJ77UmLZcV0SIg=o2V4fE_fqaIbrjbmNR6_fsyMlwSXxnUqLeXJ9BJOtq8&
> > e= I was wondering if you have some workaround. I followed the
> > solutions in the following threads:
> > https://urldefense.proofpoint.com/v2/url?u=https-3A__vufind.org_jira_b
> > rowse_VUFIND-2D1290=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM
> > 3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZc
> > V0SIg=eKl-Z7ZRS2z4218azygyFqhq9frW1NoxP1tQPyNsSNA=
> >
> > https://urldefense.proofpoint.com/v2/url?u=https-3A__underyx.me_2015_0
> > 5_18_raising-2Dthe-2Dmaximum-2Dnumber-2Dof-2Dfile-2Ddescriptors=DwIF
> > aQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0
> > KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=vYN5GI8ow39gUfRse7
> > au0f5AMd53tNfns8fibFtBLJA= and was able to resolve Max Processes
> > Limit but not File limit:
> > *   [WARN] *** Your open file limit is currently 1024.
> >It should be set to 65000 to avoid operational
> > disruption.
> >If you no longer wish to see this warning, set
> > SOLR_ULIMIT_CHECKS to false in your profile or
> https://urldefense.proofpoint.com/v2/url?u=http-3A__solr.in.sh=DwIFaQ=0TzQCy9lgR5hSW-bDg5HA76y7nf4lvOzvVop5GM3Y80=pklH2GQ2S8JTF_37tP0KFw=5cbIisJU0sjIn9-_kOn8BXuwT2pq3aJ77UmLZcV0SIg=RlMA2272ZBR8WCRfLL8I599seXXPaNeEmSZuUSmKTEo=
> >   Waiting up to 180 seconds to see Solr running on
> > port
> > 8983 []
> >   Started Solr server on port 8983 (pid=2843). Happy
> > searching!
> >
> > cd proc# cat 2843/limits:
> > Max processes 6500065000
> > processes
> > Max open files4096 4096 files
> >
> > The problem persisted after upgrade to Ubuntu 18.10 Any other solution
> > would be appreciated.
> > Otherwise can you please tell me what are the likely consequences of
> > the open file limit?

Re: Open file limit warning when starting solr

2018-12-12 Thread Daniel Carrasco
Hello,

Try creating a file in /etc/security/limits.d/solr.conf with this:
solr softnofile  65000
solr hardnofile  65000
solr softnproc   65000
solr hardnproc   65000

This worked for me on Debian 9.

Greetings!

El mié., 12 dic. 2018 a las 11:09, Armon, Rony () escribió:

> Hello, When launching solr (Ubuntu 16.04) I'm getting:
> *   [WARN] *** Your open file limit is currently 1024.
>It should be set to 65000 to avoid operational
> disruption.
>If you no longer wish to see this warning, set
> SOLR_ULIMIT_CHECKS to false in your profile or solr.in.sh
> *   [WARN] ***  Your Max Processes Limit is currently 15058.
>  It should be set to 65000 to avoid operational disruption.
>  If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS to
> false in your profile or solr.in.sh
>
> This appears to be related to a known bug in Ubuntu<
> https://issues.apache.org/jira/browse/SOLR-13063>
> https://blog.jayway.com/2012/02/11/how-to-really-fix-the-too-many-open-files-problem-for-tomcat-in-ubuntu/
> I was wondering if you have some workaround. I followed the solutions in
> the following threads:
> https://vufind.org/jira/browse/VUFIND-1290
>
> https://underyx.me/2015/05/18/raising-the-maximum-number-of-file-descriptors
> and was able to resolve Max Processes Limit but not File limit:
> *   [WARN] *** Your open file limit is currently 1024.
>It should be set to 65000 to avoid operational
> disruption.
>If you no longer wish to see this warning, set
> SOLR_ULIMIT_CHECKS to false in your profile or solr.in.sh
>   Waiting up to 180 seconds to see Solr running on port
> 8983 []
>   Started Solr server on port 8983 (pid=2843). Happy
> searching!
>
> cd proc# cat 2843/limits:
> Max processes 6500065000
> processes
> Max open files4096 4096 files
>
> The problem persisted after upgrade to Ubuntu 18.10
> Any other solution would be appreciated.
> Otherwise can you please tell me what are the likely consequences of the
> open file limit?
>
>
>
>
>
>
> **
> The information in this email is confidential and may be legally
> privileged. It is intended solely for the addressee. Access to this email
> by anyone else is unauthorized. If you are not the intended recipient, any
> disclosure, copying, distribution or any action taken or omitted to be
> taken in reliance on it, is prohibited and may be unlawful. When addressed
> to our clients any opinions or advice contained in this email are subject
> to the terms and conditions expressed in the governing KPMG client
> engagement letter.
> ***
>


-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Solr Setup using NRT and PULL replicas

2018-12-02 Thread Daniel Carrasco
Thanks for all the info,

We're trying some things to find the way to make it more stable, becuase
we're getting into troubles with the cluster. This weekend we've got three
downtimes in a few hours because some nodes of the cluster loose their
connection a bit during an import (we don't know why), and then the entire
cluster goes down (all replicas are NRT).

Looks like was related with hugepages of that nodes and for now we're
investigateing why.

Greetings!


El dom., 2 dic. 2018 a las 18:48, Edward Ribeiro ()
escribió:

> To mix NRT and TLOG/PULL replicas is not recommended. It is all NRT nodes
> or TLOG nodes mixed (or not) with PULL replicas. As you know, all PULL
> replicas is not possible.
>
> According to the talk below, one of the reasons is that if you have NRT
> mixed with TLOG and PULL replicas then a leadership change could make all
> tlog/pull replicas to download the entire index segments from the new
> leader:
>
> https://youtu.be/XIb8X3MwVKc at minute 23:50
>
> Also see this Activate 2018 talk https://youtu.be/dkWy2ykzAv0 at minute
> 14:45.
>
> And this on is nice too: https://youtu.be/XqfTjd9KDWU
>
> Best,
> Edward
>
>
> Em sex, 30 de nov de 2018 08:56, Daniel Carrasco  escreveu:
>
> > Hello,
> >
> > We've a cluster consisting in 7 to 10 NRT nodes serving data to a webpage
> > (products, categories,...), but every time a leader node fails importing
> > data (connection lost, broken pipe...), the entire cluster goes to
> > recovering mode and then is not working for about 15-30 minutes. That's a
> > lot for the webpage so we're trying to reduce that problem to minimal and
> > we're thinking about new PULL nodes.
> >
> > We've for now Solr 7.2.1, but we're planning to migrate to Solr 7.5, and
> > I've read on Solr guide that recommended setups are:
> >
> >- All NRT
> >- All TLOG
> >- Some TLOG with PULL replicas
> >
> > We're not fully convinced about TLOG replicas because we've read
> something
> > about index problems if a node goes down suddenly, or using kill -9 (just
> > what the init.d script does if takes long to stop/restart), and is the
> > leader, or about the increase in recovering times respecting to NRT, so
> > we're thinking about to keep some NRT instead TLOG.
> >
> > To have two NRT nodes and the rest TLOG is a good setup, or better to
> think
> > in TOG nodes?
> >
> > Thanks!
> >
> > --
> > _
> >
> >   Daniel Carrasco Marín
> >   Ingeniería para la Innovación i2TIC, S.L.
> >   Tlf:  +34 911 12 32 84 Ext: 223
> >   www.i2tic.com
> > _
> >
>


-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Solr Setup using NRT and PULL replicas

2018-11-30 Thread Daniel Carrasco
Hello,

We've a cluster consisting in 7 to 10 NRT nodes serving data to a webpage
(products, categories,...), but every time a leader node fails importing
data (connection lost, broken pipe...), the entire cluster goes to
recovering mode and then is not working for about 15-30 minutes. That's a
lot for the webpage so we're trying to reduce that problem to minimal and
we're thinking about new PULL nodes.

We've for now Solr 7.2.1, but we're planning to migrate to Solr 7.5, and
I've read on Solr guide that recommended setups are:

   - All NRT
   - All TLOG
   - Some TLOG with PULL replicas

We're not fully convinced about TLOG replicas because we've read something
about index problems if a node goes down suddenly, or using kill -9 (just
what the init.d script does if takes long to stop/restart), and is the
leader, or about the increase in recovering times respecting to NRT, so
we're thinking about to keep some NRT instead TLOG.

To have two NRT nodes and the rest TLOG is a good setup, or better to think
in TOG nodes?

Thanks!

-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Manage new nodes types limit

2018-11-27 Thread Daniel Carrasco
Hello,

I'm working on a new cluster and I want to make it with two TLOG nodes and
N PULL nodes.
For now I've found the way to scale the cluster adding replicas when a node
joins the cluster and deleting the replicas when is unreachable (like for
example after a shutdown), but all new replicas are NRT.


   - Is there any way to keep two TLOG replicas always?, if one fails for
   example and is removed from cluster, then create a new TLOG replica or
   promote an existing replica.
   - Is possible to configure the cluster to create the other new replicas
   as PULL?

Thanks!!

-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Autoscaling using triggers to create new replicas

2018-11-27 Thread Daniel Carrasco
Hello,

Finally I've found the way to do it. Just limiting the replicas number by
node using policies is the trick:
curl -X POST -H 'Content-Type: application/json' '
http://localhost:8983/api/cluster/autoscaling' --data-binary '{
"set-cluster-policy": [{"replica": "<2", "shard": "#EACH", "node": "#ANY"}]
}'

Greetings!

El lun., 26 nov. 2018 a las 18:24, Daniel Carrasco ()
escribió:

> Hello,
>
> I'm trying to create an autoscaling cluster with node_added_trigger
> and node_lost_trigger triggers to allow to grow and srink depending of
> load, but I've not found too much info about this options and then only
> I've archieve is a collection that creates a lot of réplicas on same node
> when it starts, and another collection that just keeps with the same
> replica number.
>
> I've created three ZK nodes, and I've joined a Solr node to that ZK
> cluster.  After that, I've created two collections with this simple
> commands:
> curl "
> http://127.0.0.1:8983/solr/admin/collections?action=CREATE=test=1=1=1
> "
> curl "
> http://127.0.0.1:8983/solr/admin/collections?action=CREATE=test2=1=1=1
> "
>
> I've added the two triggers using the example of the Solr manual:
> curl -X POST -H 'Content-Type: application/json' '
> http://localhost:8983/solr/admin/autoscaling' --data-binary '{
> "set-trigger": {"name": "node_added_trigger","event": "nodeAdded",
>   "waitFor": "5s","preferredOperation": "ADDREPLICA" } }'
> curl -X POST -H 'Content-Type: application/json' '
> http://localhost:8983/solr/admin/autoscaling' --data-binary '{
>  "set-trigger": {"name": "node_lost_trigger","event": "nodeLost",
>   "waitFor": "120s","preferredOperation": "DELETENODE"  }}'
>
>
> And after all this, I've added a node to test.
> When that node starts, the test collection starts to add replicas without
> control (it added even 12 replicas on the new node), and the test2
> collection keeps on just one replica. If I delete the test collection and
> repeat the process of add the new node, then the test2 have the same
> behavior  and creates a lot of replicas on new node.
>
> Delete trigger works just fine, and when the node is down for about 120s
> it is deleted from collection without problem.
>
> Is there any way to create just one replica when a node joins the cluster?.
>
> Thanks!
> --
> _
>
>   Daniel Carrasco Marín
>   Ingeniería para la Innovación i2TIC, S.L.
>   Tlf:  +34 911 12 32 84 Ext: 223
>   www.i2tic.com
> _
>


-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Autoscaling using triggers to create new replicas

2018-11-26 Thread Daniel Carrasco
Hello,

I'm trying to create an autoscaling cluster with node_added_trigger
and node_lost_trigger triggers to allow to grow and srink depending of
load, but I've not found too much info about this options and then only
I've archieve is a collection that creates a lot of réplicas on same node
when it starts, and another collection that just keeps with the same
replica number.

I've created three ZK nodes, and I've joined a Solr node to that ZK
cluster.  After that, I've created two collections with this simple
commands:
curl "
http://127.0.0.1:8983/solr/admin/collections?action=CREATE=test=1=1=1
"
curl "
http://127.0.0.1:8983/solr/admin/collections?action=CREATE=test2=1=1=1
"

I've added the two triggers using the example of the Solr manual:
curl -X POST -H 'Content-Type: application/json' '
http://localhost:8983/solr/admin/autoscaling' --data-binary '{
"set-trigger": {"name": "node_added_trigger","event": "nodeAdded",
  "waitFor": "5s","preferredOperation": "ADDREPLICA" } }'
curl -X POST -H 'Content-Type: application/json' '
http://localhost:8983/solr/admin/autoscaling' --data-binary '{
 "set-trigger": {"name": "node_lost_trigger","event": "nodeLost",
  "waitFor": "120s","preferredOperation": "DELETENODE"  }}'


And after all this, I've added a node to test.
When that node starts, the test collection starts to add replicas without
control (it added even 12 replicas on the new node), and the test2
collection keeps on just one replica. If I delete the test collection and
repeat the process of add the new node, then the test2 have the same
behavior  and creates a lot of replicas on new node.

Delete trigger works just fine, and when the node is down for about 120s it
is deleted from collection without problem.

Is there any way to create just one replica when a node joins the cluster?.

Thanks!
-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Slow import from MsSQL and down cluster during process

2018-10-23 Thread Daniel Carrasco
Thanks for all, I'll try later ;)

Greetings!!.

El mié., 24 oct. 2018 a las 7:13, Walter Underwood ()
escribió:

> We handle request rates at a few thousand requests/minute with an 8 GB
> heap. 95th percentile response time is 200 ms. Median (cached) is 4 ms.
>
> An oversized heap will hurt your query performance because everything
> stops for the huge GC.
>
> RAM is still a thousand times faster than SSD, so you want a lot of RAM
> available for file system buffers managed by the OS.
>
> I recommend trying an 8 GB heap with the latest version of Java 8 and the
> G1 collector.
>
> We have this in our solr.in.sh:
>
> SOLR_HEAP=8g
> # Use G1 GC  -- wunder 2017-01-23
> # Settings from https://wiki.apache.org/solr/ShawnHeisey
> GC_TUNE=" \
> -XX:+UseG1GC \
> -XX:+ParallelRefProcEnabled \
> -XX:G1HeapRegionSize=8m \
> -XX:MaxGCPauseMillis=200 \
> -XX:+UseLargePages \
> -XX:+AggressiveOpts \
> "
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
>
> > On Oct 23, 2018, at 9:51 PM, Daniel Carrasco 
> wrote:
> >
> > Hello,
> >
> > I've set that heap size because the solr receives a lot of queries every
> > second and I want to cache as much as possible. Also I'm not sure about
> the
> > number of documents in the collection, but the webpage have a lot of
> > products.
> >
> > About store the index data in RAM is just an expression. The data is
> stored
> > on SSD disks with XFS (faster than EXT4).
> >
> > I'll take a look to the links tomorrow at work.
> >
> > Thanks!!
> > Greetings!!
> >
> >
> > El mar., 23 oct. 2018 23:48, Shawn Heisey 
> escribió:
> >
> >> On 10/23/2018 7:15 AM, Daniel Carrasco wrote:
> >>> Hello,
> >>>
> >>> Thanks for your response.
> >>>
> >>> We've already thought about that and doubled the instances. Just now
> for
> >>> every Solr instance we've 60GB of RAM (40GB configured on Solr), and a
> 16
> >>> Cores CPU. The entire Data can be stored on RAM and will not fill the
> RAM
> >>> (of course talking about raw data, not procesed data).
> >>
> >> Why are you making the heap so large?  I've set up servers that can
> >> handle hundreds of millions of Solr documents in a much smaller heap.  A
> >> 40GB heap would be something you might do if you're handling billions of
> >> documents on one server.
> >>
> >> When you say the entire data can be stored in RAM ... are you counting
> >> that 40GB you gave to Solr?  Because you can't count that -- that's for
> >> Solr, NOT the index data.
> >>
> >> The heap size should never be dictated by the amount of memory in the
> >> server.  It should be made as large as it needs to be for the job, and
> >> no larger.
> >>
> >> https://wiki.apache.org/solr/SolrPerformanceProblems#RAM
> >>
> >>> About the usage, I've checked the RAM and CPU usage and are not fully
> >> used.
> >>
> >> What exactly are you looking at?  I've had people swear that they can't
> >> see a problem with their systems when Solr is REALLY struggling to keep
> >> up with what it has been asked to do.
> >>
> >> Further down on the page I linked above is a section about asking for
> >> help.  If you can provide the screenshot it mentions there, that would
> >> be helpful.  Here's a direct link to that section:
> >>
> >>
> >>
> https://wiki.apache.org/solr/SolrPerformanceProblems#Asking_for_help_on_a_memory.2Fperformance_issue
> >>
> >> Thanks,
> >> Shawn
> >>
> >>
>
>

-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Slow import from MsSQL and down cluster during process

2018-10-23 Thread Daniel Carrasco
Hello,

I've set that heap size because the solr receives a lot of queries every
second and I want to cache as much as possible. Also I'm not sure about the
number of documents in the collection, but the webpage have a lot of
products.

About store the index data in RAM is just an expression. The data is stored
on SSD disks with XFS (faster than EXT4).

I'll take a look to the links tomorrow at work.

Thanks!!
Greetings!!


El mar., 23 oct. 2018 23:48, Shawn Heisey  escribió:

> On 10/23/2018 7:15 AM, Daniel Carrasco wrote:
> > Hello,
> >
> > Thanks for your response.
> >
> > We've already thought about that and doubled the instances. Just now for
> > every Solr instance we've 60GB of RAM (40GB configured on Solr), and a 16
> > Cores CPU. The entire Data can be stored on RAM and will not fill the RAM
> > (of course talking about raw data, not procesed data).
>
> Why are you making the heap so large?  I've set up servers that can
> handle hundreds of millions of Solr documents in a much smaller heap.  A
> 40GB heap would be something you might do if you're handling billions of
> documents on one server.
>
> When you say the entire data can be stored in RAM ... are you counting
> that 40GB you gave to Solr?  Because you can't count that -- that's for
> Solr, NOT the index data.
>
> The heap size should never be dictated by the amount of memory in the
> server.  It should be made as large as it needs to be for the job, and
> no larger.
>
> https://wiki.apache.org/solr/SolrPerformanceProblems#RAM
>
> > About the usage, I've checked the RAM and CPU usage and are not fully
> used.
>
> What exactly are you looking at?  I've had people swear that they can't
> see a problem with their systems when Solr is REALLY struggling to keep
> up with what it has been asked to do.
>
> Further down on the page I linked above is a section about asking for
> help.  If you can provide the screenshot it mentions there, that would
> be helpful.  Here's a direct link to that section:
>
>
> https://wiki.apache.org/solr/SolrPerformanceProblems#Asking_for_help_on_a_memory.2Fperformance_issue
>
> Thanks,
> Shawn
>
>


Re: Slow import from MsSQL and down cluster during process

2018-10-23 Thread Daniel Carrasco
Hello,

Thanks for your response.

We've already thought about that and doubled the instances. Just now for
every Solr instance we've 60GB of RAM (40GB configured on Solr), and a 16
Cores CPU. The entire Data can be stored on RAM and will not fill the RAM
(of course talking about raw data, not procesed data).

About the usage, I've checked the RAM and CPU usage and are not fully used.

Greetings!

El mar., 23 oct. 2018 a las 14:02, Chris Ulicny ()
escribió:

> Dan,
>
> Do you have any idea on the resource usage for the hosts when Solr starts
> to become unresponsive? It could be that you need more resources or better
> AWS instances for the hosts.
>
> We had what sounds like a similar scenario when attempting to move one of
> our solrcloud instances to a cloud computing platform. During periods of
> heaving indexing, segment merging, and searches, the cluster would become
> unresponsive due to solr waiting for numerous I/O operations which we being
> throttled. Solr can be very I/O intensive, especially when you can't cache
> the entire index in memory.
>
> Thanks,
> Chris
>
>
> On Tue, Oct 23, 2018 at 5:40 AM Daniel Carrasco 
> wrote:
>
> > Hi,
> > El mar., 23 oct. 2018 a las 10:18, Charlie Hull ()
> > escribió:
> >
> > > On 23/10/2018 02:57, Daniel Carrasco wrote:
> > > > annoyingHello,
> > > >
> > > > I've a Solr Cluster that is created with 7 machines on AWS instances.
> > The
> > > > Solr version is 7.2.1 (b2b6438b37073bee1fca40374e85bf91aa457c0b) and
> > all
> > > > nodes are running on NTR mode and I've a replica by node (7
> replicas).
> > > One
> > > > node is used to import, and the rest are just for serve data.
> > > >
> > > > My problem is that I'm having problems from about two weeks with a
> > MsSQL
> > > > import on my Solr Cluster: when the process becomes slow or takes too
> > > long,
> > > > the entire cluster goes down.
> > >
> > > How exactly are you importing from MsSQL to Solr? Are you using the
> Data
> > > Import Handler (DIH) and if so, how?
> >
> >
> > yeah, we're using import handler with jdbc connector:
> >
> > 
> >> driver="com.microsoft.sqlserver.jdbc.SQLServerDriver"
> > url="jdbc:sqlserver://..." user="..." password="..."/>
> >  > query="A_Long_Query" />
> > 
> > ... A lot of fields configuration ...
> > 
> > ... some entities similar to above ...
> >   
> > 
> >
> >
> >
> > > What evidence do you have that  this is slow or takes too long?
> > >
> >
> > Well, the process normally takes less than 20 minutes and doesn't affect
> at
> > all to cluster (normally near 15m). I've a monit system that notice when
> > this process takes more than 25 minutes, and just a bit later after that
> > alert, the entire collection goes to recovery mode and then we're unable
> to
> > continue to serve the requests made by the webpage. We've to stop all the
> > requests until the collection is OK again. The rest of time the cluster
> > works perfect without downtime, but lately the problem is happen more
> often
> > (I'd to recover the cluster two times in less than an hour this night,
> and
> > it didn't fail again because we've stopped the import cron).
> > This is the soft problem, because sometimes the entire cluster becomes
> > unstable and affects to other collections. Sometimes even the node that
> is
> > Leader fails and we're unable to release that Leadership (even shutting
> > down the Leader server, running the FORCELEADER API command), and that
> make
> > hard to recovery the cluster. If we're lucky, the cluster recovers itself
> > even with recovering leader (taking so long, of course), but sometimes
> > we've no luck and we've to reboot all the machines to force a full
> recover.
> >
> >
> > >
> > > Charlie
> > > >
> > > > I'm confused, because the main reason to have a cluster is HA, and
> > every
> > > > time the import node "fails" (is not really failing, just taking more
> > > time
> > > > to finish), the entire cluster fails and I've to stop the webpage
> until
> > > > nodes are green again.
> > > >
> > > > I don't know if maybe I've to change something in configuration to
> > allow
> > > > the cluster to keep working even when the import freezes or the
> import
> > > node
> > > > dies, but

Re: Slow import from MsSQL and down cluster during process

2018-10-23 Thread Daniel Carrasco
Hi,
El mar., 23 oct. 2018 a las 10:18, Charlie Hull ()
escribió:

> On 23/10/2018 02:57, Daniel Carrasco wrote:
> > annoyingHello,
> >
> > I've a Solr Cluster that is created with 7 machines on AWS instances. The
> > Solr version is 7.2.1 (b2b6438b37073bee1fca40374e85bf91aa457c0b) and all
> > nodes are running on NTR mode and I've a replica by node (7 replicas).
> One
> > node is used to import, and the rest are just for serve data.
> >
> > My problem is that I'm having problems from about two weeks with a MsSQL
> > import on my Solr Cluster: when the process becomes slow or takes too
> long,
> > the entire cluster goes down.
>
> How exactly are you importing from MsSQL to Solr? Are you using the Data
> Import Handler (DIH) and if so, how?


yeah, we're using import handler with jdbc connector:


  


... A lot of fields configuration ...

... some entities similar to above ...
  




> What evidence do you have that  this is slow or takes too long?
>

Well, the process normally takes less than 20 minutes and doesn't affect at
all to cluster (normally near 15m). I've a monit system that notice when
this process takes more than 25 minutes, and just a bit later after that
alert, the entire collection goes to recovery mode and then we're unable to
continue to serve the requests made by the webpage. We've to stop all the
requests until the collection is OK again. The rest of time the cluster
works perfect without downtime, but lately the problem is happen more often
(I'd to recover the cluster two times in less than an hour this night, and
it didn't fail again because we've stopped the import cron).
This is the soft problem, because sometimes the entire cluster becomes
unstable and affects to other collections. Sometimes even the node that is
Leader fails and we're unable to release that Leadership (even shutting
down the Leader server, running the FORCELEADER API command), and that make
hard to recovery the cluster. If we're lucky, the cluster recovers itself
even with recovering leader (taking so long, of course), but sometimes
we've no luck and we've to reboot all the machines to force a full recover.


>
> Charlie
> >
> > I'm confused, because the main reason to have a cluster is HA, and every
> > time the import node "fails" (is not really failing, just taking more
> time
> > to finish), the entire cluster fails and I've to stop the webpage until
> > nodes are green again.
> >
> > I don't know if maybe I've to change something in configuration to allow
> > the cluster to keep working even when the import freezes or the import
> node
> > dies, but is very annoying to wake up at 3AM to fix the cluster.
> >
> > Is there any way to avoid this?, maybe keeping the import node as NTR and
> > convert the rest to TLOG?
> >
> > I'm a bit noob in Solr, so I don't know if I've to sent something to help
> > to find the problem, and the cluster was created just creating a
> Zookeeper
> > cluster, connecting the Solr nodes to that Zk cluster, importing the
> > collections and adding réplicas manually to every collection.
> > Also I've upgraded that cluster from Solr 6 to Solr 7.1 and later to Solr
> > 7.2.1.
> >
> > Thanks and greetings!
> >
>
>
> --
> Charlie Hull
> Flax - Open Source Enterprise Search
>
> tel/fax: +44 (0)8700 118334
> mobile:  +44 (0)7767 825828
> web: www.flax.co.uk
>


Thanks, and greetings!!

-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Slow import from MsSQL and down cluster during process

2018-10-22 Thread Daniel Carrasco
annoyingHello,

I've a Solr Cluster that is created with 7 machines on AWS instances. The
Solr version is 7.2.1 (b2b6438b37073bee1fca40374e85bf91aa457c0b) and all
nodes are running on NTR mode and I've a replica by node (7 replicas). One
node is used to import, and the rest are just for serve data.

My problem is that I'm having problems from about two weeks with a MsSQL
import on my Solr Cluster: when the process becomes slow or takes too long,
the entire cluster goes down.

I'm confused, because the main reason to have a cluster is HA, and every
time the import node "fails" (is not really failing, just taking more time
to finish), the entire cluster fails and I've to stop the webpage until
nodes are green again.

I don't know if maybe I've to change something in configuration to allow
the cluster to keep working even when the import freezes or the import node
dies, but is very annoying to wake up at 3AM to fix the cluster.

Is there any way to avoid this?, maybe keeping the import node as NTR and
convert the rest to TLOG?

I'm a bit noob in Solr, so I don't know if I've to sent something to help
to find the problem, and the cluster was created just creating a Zookeeper
cluster, connecting the Solr nodes to that Zk cluster, importing the
collections and adding réplicas manually to every collection.
Also I've upgraded that cluster from Solr 6 to Solr 7.1 and later to Solr
7.2.1.

Thanks and greetings!
-- 
_____

      Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Inconsistent leader between ZK and Solr and a lot of downtime

2018-10-18 Thread Daniel Carrasco
Hello,

I'm investigating an 8 nodes Solr 7.2.1 cluster because we've a lot of
problems, like when a node fails to import from a DB (maybe it freeze), the
entire cluster goes down, and other like the leader wont change even when
is down (all nodes detects that is down but no leader election is
triggered), and similar problems. Every few days we've to recover the
cluster because becomes inestable and goes down.

The last problem that I've got, is three collections that have nodes on
"recovery" state from a lot of hours, and the log shows an error telling
that "leader node is not the leader" so I'm trying to change the leader.
After shutting down the "leader" (detected by the other nodes as down and
waiting about 20 minutes), trying REBALANCELEADER and FORCELEADER, I'm
unable to change the leader on the cluster, and that's why started to see
on ZooKeeper. The problem I've seen on Zookeeper is that Leaders are
different than Solr admin cluster info, so Maybe that's why the nodes are
unable to connect to real leader and cannot end the recovery.

The entire cluster and ZK has the traffic open to avoid problems (the VPC
is private), so is not a connection problem.

Is there any way to sync the leader info between solr and ZK?, also I want
to know if exists a way to force to change the leader (FORCELEADER don't
work when the solr denies to change the leader, because it say that a
leader exists).

Thanks!
-- 
_________

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Secure way to backup Solrcloud

2018-03-07 Thread Daniel Carrasco
Hello,

My question is if there is any way to backup a Solr cluster even when all
replicas are "not synced"...

I'm using the api to create the dumps:
http://localhost:8983/solr/admin/collections?action=BACKUP=myBackupName=myCollectionName=/path/to/my/shared/drive

But is a lottery where the most of time a collection fails and is
impossible to backup until you restart the node that looks out of sync.

Error: {
  "responseHeader":{
"status":500,
"QTime":4057},
  "failure":{

"192.168.3.14:8983_solr":"org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException:Error
from server at http://192.168.3.14:8983/solr: Failed to backup
core=myproducts_shard1_replica_n91 because
java.nio.file.NoSuchFileException:
/server/solr/solr-data/data/myproducts_shard1_replica_n91/data/index.20180212231626422/segments_5kjp"},
  "Operation backup caused
exception:":"org.apache.solr.common.SolrException:org.apache.solr.common.SolrException:
Could not backup all replicas",
  "exception":{
"msg":"Could not backup all replicas",
"rspCode":500},
  "error":{
"metadata":[
  "error-class","org.apache.solr.common.SolrException",
  "root-error-class","org.apache.solr.common.SolrException"],
"msg":"Could not backup all replicas",
"trace":"org.apache.solr.common.SolrException: Could not backup all
replicas\n\tat
org.apache.solr.handler.admin.CollectionsHandler.handleResponse(CollectionsHandler.java:306)\n\tat
org.apache.solr.handler.admin.CollectionsHandler.invokeAction(CollectionsHandler.java:243)\n\tat
org.apache.solr.handler.admin.CollectionsHandler.handleRequestBody(CollectionsHandler.java:221)\n\tat
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:177)\n\tat
org.apache.solr.servlet.HttpSolrCall.handleAdmin(HttpSolrCall.java:745)\n\tat
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:726)\n\tat
org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:507)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:382)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:326)\n\tat
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1751)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:582)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)\n\tat
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)\n\tat
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)\n\tat
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)\n\tat
org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(RewriteHandler.java:335)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)\n\tat
org.eclipse.jetty.server.Server.handle(Server.java:534)\n\tat
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:320)\n\tat
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:251)\n\tat
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283)\n\tat
org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:108)\n\tat
org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)\n\tat
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303)\n\tat
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148)\n\tat
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589)\n\tat
java.lang.Thread.run(Thread.java:748)\n",
"code":500}}

After restart the node that give the error then is able to dump the backup
without my problem, but i can't restart all nodes that fail and every day
to dump the data.

is there any way to do it?

Thanks!


-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: [poll] which loadbalancer are you using for SolrCloud

2018-03-02 Thread Daniel Carrasco
I use HAProxy, because is much more configurable than Nginx and I can send
commands to solr collection and search for text to check if the node is
healthy.

Nginx is very fast too, but health check are worst than HAProxy.

Greetings!!

2018-03-02 17:11 GMT+01:00 David Hastings <hastings.recurs...@gmail.com>:

> Ill have to take a look at HAProxy.  How much faster than nginx is it?
>
> To answer the question, I personally use nginx for load balancing/failovers
> and its been good, use the same nginx servers to load balance a Galera
> cluster as well.
>
> On Fri, Mar 2, 2018 at 11:09 AM, Shawn Heisey <elyog...@elyograg.org>
> wrote:
>
> > On 3/2/2018 6:13 AM, Bernd Fehling wrote:
> >
> >> I would like to poll for the loadbalancer you are using for SolrCloud.
> >>
> >> Are you using a loadbalancer for SolrCloud?
> >>
> >> If yes, which one (SolrJ, HAProxy, Varnish, Nginx,...) and why?
> >>
> >
> > I use haproxy for Solr -- not SolrCloud.  It is an amazing and FAST piece
> > of software, without the overhead of a full webserver (apache, nginx).
> It
> > also has zero cost, which is far more attractive than hardware load
> > balancers, and can do anything I've seen a hardware load balancer do.
> With
> > the presence of another piece of software (such as pacemaker) you can
> even
> > have hardware redundancy for the load balancer.
> >
> > Most of my clients talking to Solr are Java, so they use
> > HttpSolrClient/HttpSolrServer from SolrJ, connecting to the load
> balancer.
> >
> > For SolrCloud, if your clients are Java, you don't need a load balancer,
> > because the client (CloudSolrClient in SolrJ) talks to the entire cluster
> > and dynamically adjusts to changes in clusterstate.
> >
> > Thanks,
> > Shawn
> >
> >
>



-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Solr node is out of sync (looks Healthy)

2018-02-13 Thread Daniel Carrasco
Hello,

I answer inline ;)

2018-02-12 23:56 GMT+01:00 Emir Arnautović <emir.arnauto...@sematext.com>:

> Hi Daniel,
> Please see inline comments.
> --
> Monitoring - Log Management - Alerting - Anomaly Detection
> Solr & Elasticsearch Consulting Support Training - http://sematext.com/
>
>
>
> > On 12 Feb 2018, at 13:13, Daniel Carrasco <d.carra...@i2tic.com> wrote:
> >
> > Hello,
> >
> > 2018-02-12 12:32 GMT+01:00 Emir Arnautović <emir.arnauto...@sematext.com
> <mailto:emir.arnauto...@sematext.com>>:
> >
> >> Hi Daniel,
> >> Maybe it is Monday and I am still not warmed up, but your details seems
> a
> >> bit imprecise to me. Maybe not directly related to your problem, but
> just
> >> to exclude that you are having some strange Solr setup, here is my
> >> understanding: You are running a single SolrCloud cluster with 8 nodes.
> It
> >> has a single collection with X shards and Y replicas. You use DIH to
> index
> >> data and you use curl to interact with Solr and start DIH process. You
> see
> >> some of replicas for some of shards having less data and after node
> restart
> >> it ends up being ok.
> >
> >
> >> Is this right? If it is, what is X and Y?
> >
> >
> > Near to reality:
> >
> >   - I've a SolrCloud cluster with 8 nodes but has multiple collections.
> >   - Every collection has only one shard for performance purpose (I did
> >   some test splitting shards and queries were slower).
> Distributed request comes with an overhead and if collection is small,
> that overhead can be larger than benefit of parallelising search.
>
> >   - Every collection has 8 replicas (one by node)
> I would compare all shards on all nodes (64 Solr cores) v.s. having just
> one replica (16 Solr cores)
>

We've all shards on all nodes because we want to avoid the overhead of send
data between nodes (latency, network traffic). The page has a lot of
petitions per second and we want the fastest response posible with HA if
some nodes fails.

Also we've 8 collections: five are small (less than 15Mb), and three of
then has some GB (the bigger is about 10Gb).
Is the first SolrCloud I've created and I'd decided this architecture to
avoid what I say above, the overhead of sent the data between nodes when
the client ask for data to another node. Maybe is better to have a
replication factor of 3-4 for example and create shards for big collections?


>
> >   - After restart the node it start to recover the collections. I don't
> >   know if Solr serve data directly on that state or get the data from
> other
> >   nodes before serve it, but even while is recovering, the data looks OK.
> Recovery can be from transaction logs (logs can tell) and that would mean
> that there was no hard commit after some updates.
>
> >
> >
> >
> >> Do you have autocommit set up or you commit explicitly?
> >
> >
> > I'm not sure about that. How I can check it?
> It is part of solr.xml
>

I've checked the file and looks like there's no configuration about
autocommit. I'll search a bit about how works to see if can help.


>
> >
> > On curl command is not specified, but will be true by default, right?
> I think it is for full import.
>
> >
> >
> >
> >> Did you check logs on node with less data and did you see any
> >> errors/warnings?
> >
> >
> > I'm not sure when it failed and the cluster has a lot warnings and error
> > every time (maybe related with queries from shop), so is hard to
> determine
> > if import error exists and what's the error related to the import. Is
> like
> > search a needle on a haystack
> Not with some logging solution - one such is Sematext’s Logsene:
> https://sematext.com/logsene/ <https://sematext.com/logsene/>
>
> >
> >
> >> Do you do full imports or incremental imports?
> >>
> >
> > I've checked the curl command and looks like is doing full imports
> without
> > clean data:
> > http://'  . $solr_ip .
> > ':8983/solr/descriptions/dataimport?command=full-
> import=false=description_’.$idm[$j].'_lastupdate
> This is not a good strategy since Solr does not have real updates - it is
> delete + insert and deleted documents are purged on segment merges. Also
> this will not eliminate documents that are deleted in the primary storage.
> It is much better to index it in new collection and use aliases to point to
> used collection. This way you can even roll back if not happy with new
> index.
>
>
But if you update three products for example and you create a new
collection 

Re: Solr node is out of sync (looks Healthy)

2018-02-12 Thread Daniel Carrasco
Hello,

2018-02-12 12:32 GMT+01:00 Emir Arnautović <emir.arnauto...@sematext.com>:

> Hi Daniel,
> Maybe it is Monday and I am still not warmed up, but your details seems a
> bit imprecise to me. Maybe not directly related to your problem, but just
> to exclude that you are having some strange Solr setup, here is my
> understanding: You are running a single SolrCloud cluster with 8 nodes. It
> has a single collection with X shards and Y replicas. You use DIH to index
> data and you use curl to interact with Solr and start DIH process. You see
> some of replicas for some of shards having less data and after node restart
> it ends up being ok.


> Is this right? If it is, what is X and Y?


Near to reality:

   - I've a SolrCloud cluster with 8 nodes but has multiple collections.
   - Every collection has only one shard for performance purpose (I did
   some test splitting shards and queries were slower).
   - Every collection has 8 replicas (one by node)
   - After restart the node it start to recover the collections. I don't
   know if Solr serve data directly on that state or get the data from other
   nodes before serve it, but even while is recovering, the data looks OK.



> Do you have autocommit set up or you commit explicitly?


I'm not sure about that. How I can check it?

On curl command is not specified, but will be true by default, right?



> Did you check logs on node with less data and did you see any
> errors/warnings?


I'm not sure when it failed and the cluster has a lot warnings and error
every time (maybe related with queries from shop), so is hard to determine
if import error exists and what's the error related to the import. Is like
search a needle on a haystack


> Do you do full imports or incremental imports?
>

I've checked the curl command and looks like is doing full imports without
clean data:
http://' . $solr_ip .
':8983/solr/descriptions/dataimport?command=full-import=false=description_'.$idm[$j].'_lastupdate


>
> Not related to issue, but just a note that Solr does not guaranty
> consistency at any time - it has something called “eventual consistency” -
> once updates stop all replicas will (should) end up in the same state.
> Having said that, using Solr results directly in your UI would either
> require you to cache used documents on UI/middle layer or implement some
> sort of stickiness or retrieve only ID from Solr and load data from primary
> storage. If you have static data, and you update index once a day, you can
> use aliases and switch between new and old index and you will suffer from
> this issue only at the time when doing switches.
>

But is normal that data will be inconsistent for a very long time?, because
looks like the data is inconsistent from about a week...

Another question: With HDFS, data will be consistent?. With HDFS the data
will be shared between nodes and then updates will be avaible on all nodes
at same time, right?

Thanks!!


>
> Regards,
> Emir
> --
> Monitoring - Log Management - Alerting - Anomaly Detection
> Solr & Elasticsearch Consulting Support Training - http://sematext.com/
>
>
>
> > On 12 Feb 2018, at 12:00, Daniel Carrasco <d.carra...@i2tic.com> wrote:
> >
> > Hello, thanks for your help.
> >
> > I answer bellow.
> >
> > Greetings!!
> >
> > 2018-02-12 11:31 GMT+01:00 Emir Arnautović <emir.arnauto...@sematext.com
> <mailto:emir.arnauto...@sematext.com>>:
> >
> >> Hi Daniel,
> >> Can you tell us more about your document update process. How do you
> commit
> >> changes? Since it got fixed after restart, it seems to me that on that
> one
> >> node index searcher was not reopened after updates. Do you see any
> >> errors/warnings on that node?
> >>
> >
> > i've asked to the programmers and looks like they are using the
> collections
> > dataimport using curl. I think the data is imported from a Microsoft SQL
> > server using a solr plugin.
> >
> >
> >> Also, what do you mean by “All nodes are standalone”?
> >>
> >
> > I mean that nodes don't share filesystem (I'm planning to use Hadoop, but
> > I've to learn to create and maintain the cluster first). All nodes has
> its
> > own data drive and are connected to the cluster using zookeeper.
> >
> >
> >>
> >> Regards,
> >> Emir
> >> --
> >> Monitoring - Log Management - Alerting - Anomaly Detection
> >> Solr & Elasticsearch Consulting Support Training - http://sematext.com/
> >>
> >>
> >>
> >>> On 12 Feb 2018, at 11:16, Daniel Carrasco <d.carra...@i2tic.com>
> wrote:
> >>>
> >>> Hello,
> >>>
>

Re: Solr node is out of sync (looks Healthy)

2018-02-12 Thread Daniel Carrasco
Hello, thanks for your help.

I answer bellow.

Greetings!!

2018-02-12 11:31 GMT+01:00 Emir Arnautović <emir.arnauto...@sematext.com>:

> Hi Daniel,
> Can you tell us more about your document update process. How do you commit
> changes? Since it got fixed after restart, it seems to me that on that one
> node index searcher was not reopened after updates. Do you see any
> errors/warnings on that node?
>

i've asked to the programmers and looks like they are using the collections
dataimport using curl. I think the data is imported from a Microsoft SQL
server using a solr plugin.


> Also, what do you mean by “All nodes are standalone”?
>

I mean that nodes don't share filesystem (I'm planning to use Hadoop, but
I've to learn to create and maintain the cluster first). All nodes has its
own data drive and are connected to the cluster using zookeeper.


>
> Regards,
> Emir
> --
> Monitoring - Log Management - Alerting - Anomaly Detection
> Solr & Elasticsearch Consulting Support Training - http://sematext.com/
>
>
>
> > On 12 Feb 2018, at 11:16, Daniel Carrasco <d.carra...@i2tic.com> wrote:
> >
> > Hello,
> >
> > We're using Solr to manage products data on our shop and the last week
> some
> > customers called us telling that price between shop and shopping basket
> > differs. After research a bit I've noticed that it happens sometimes on
> > page refresh.
> > After disabling all cache I've queried all solr instances to see if data
> is
> > correct and I've seen that one of them give a different price for the
> > product, so looks like the instance has not got the updated data.
> >
> >   - How is possible that a node on a cluster have different data?
> >   - How i can check if data is in sync?, because the cluster looks al
> >   healthy on admin, and the node is active and OK.
> >   - Is there any way to detect this error? and How I can force resyncs?
> >
> > After restart the node it got synced, so the data now is OK, but I can't
> > restart the nodes every time to see if data is right (it tooks a lot of
> > time to be synced again).
> >
> > My configuration is: 8 Solr nodes using v7.1.0 and zookeeper v3.4.11. All
> > nodes are standalone (I'm not using hadoop).
> >
> > Thanks and greetings!
> > --
> > _
> >
> >  Daniel Carrasco Marín
> >  Ingeniería para la Innovación i2TIC, S.L.
> >  Tlf:  +34 911 12 32 84 Ext: 223 <+34%20911%2012%2032%2084>
> >  www.i2tic.com
> > _
>
>


-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Solr node is out of sync (looks Healthy)

2018-02-12 Thread Daniel Carrasco
Hello,

We're using Solr to manage products data on our shop and the last week some
customers called us telling that price between shop and shopping basket
differs. After research a bit I've noticed that it happens sometimes on
page refresh.
After disabling all cache I've queried all solr instances to see if data is
correct and I've seen that one of them give a different price for the
product, so looks like the instance has not got the updated data.

   - How is possible that a node on a cluster have different data?
   - How i can check if data is in sync?, because the cluster looks al
   healthy on admin, and the node is active and OK.
   - Is there any way to detect this error? and How I can force resyncs?

After restart the node it got synced, so the data now is OK, but I can't
restart the nodes every time to see if data is right (it tooks a lot of
time to be synced again).

My configuration is: 8 Solr nodes using v7.1.0 and zookeeper v3.4.11. All
nodes are standalone (I'm not using hadoop).

Thanks and greetings!
-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Request node status independently

2018-02-01 Thread Daniel Carrasco
Thanks to both.

Finally I've found a way to do it with haproxy. I do what wunder said
sending the command for every collection and see if it's able to answer.
Even in recovering answer, so looks like it takes the data from other nodes
of use the data that have.

Greetings!!


El 1 feb. 2018 17:57, "Walter Underwood" <wun...@wunderwood.org> escribió:

Also, “recovering” is a status for a particular core in a collection. A
Solr process might have some cores that are healthy and some that are not.

Even if you only have one collection, you can still have multiple cores
(with different status) from the same collection on one node.

Personally, I do a search to see if a collection is ready. If a search for
“q=*:*=0” returns OK, then I’ll send traffic to that node.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)

> On Feb 1, 2018, at 8:35 AM, Erick Erickson <erickerick...@gmail.com>
wrote:
>
> The Collections API CLUSTERSTATUS essentially gives you back the ZK
> state.json for individual collections (or your cluster, see the
> params). One note: Just because the state.json reports a replica as
> "active" isn't definitive. If the node died unexpectedly its replicas
> can't set the state when shutting down. So you also have to check
> whether the replica's node is in the "live_nodes" znode.
>
> Best,
> Erick
>
> On Thu, Feb 1, 2018 at 4:34 AM, Daniel Carrasco <d.carra...@i2tic.com>
wrote:
>> Hello,
>>
>> I'm trying to create a load balancer using HAProxy to detect nodes that
are
>> down or recovering, but I'm not able to find the way to detect if the
node
>> is healthy (the only commands i've seen check the entire cluster).
>> Is there any way to check the node status using http responses and get
only
>> if is healthy or recovering?. Of course if is dead I've got no response,
so
>> that's easy.
>>
>> Thanks and greetings!!
>>
>> --
>> _
>>
>>  Daniel Carrasco Marín
>>  Ingeniería para la Innovación i2TIC, S.L.
>>  Tlf:  +34 911 12 32 84 Ext: 223
>>  www.i2tic.com
>> _


Request node status independently

2018-02-01 Thread Daniel Carrasco
Hello,

I'm trying to create a load balancer using HAProxy to detect nodes that are
down or recovering, but I'm not able to find the way to detect if the node
is healthy (the only commands i've seen check the entire cluster).
Is there any way to check the node status using http responses and get only
if is healthy or recovering?. Of course if is dead I've got no response, so
that's easy.

Thanks and greetings!!

-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Re: Uncheck dataimport checkboxes by default

2018-01-15 Thread Daniel Carrasco
Thanks Erick, I'll take a look into the js.


Greetings!!

2018-01-15 17:46 GMT+01:00 Erick Erickson <erickerick...@gmail.com>:

> Daniel:
>
> There's no preferences section in the admin UI. That said, it's
> all angular js and the source is there wherever you unpacked
> the package you could just change it. There's no need to
> rebuild Solr etc
>
> BTW, the mail server is pretty aggressive about stripping attachments,
> your (presumed) screenshot is blank
>
> Best,
> Erick
>
> On Mon, Jan 15, 2018 at 2:30 AM, Daniel Carrasco <d.carra...@i2tic.com>
> wrote:
>
> > Hello,
> >
> > My question is just what I've summarized on the subject: Is there any way
> > to change the default state of the checkboxes on dataimport admin page?
> >
> > I want to change the default state of the "clean" checkbox to uncheck
> > because sometimes I import incremental data and I forgot to uncheck that
> > box, then all data is cleared and I've to import all again.
> >
> >
> >
> > Thanks!!​
> >
> > --
> > _
> >
> >   Daniel Carrasco Marín
> >   Ingeniería para la Innovación i2TIC, S.L.
> >   Tlf:  +34 911 12 32 84 Ext: 223
> >   www.i2tic.com
> > _
> >
>



-- 
_

  Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_


Uncheck dataimport checkboxes by default

2018-01-15 Thread Daniel Carrasco
Hello,

My question is just what I've summarized on the subject: Is there any way
to change the default state of the checkboxes on dataimport admin page?

I want to change the default state of the "clean" checkbox to uncheck
because sometimes I import incremental data and I forgot to uncheck that
box, then all data is cleared and I've to import all again.



Thanks!!​

-- 
_____

      Daniel Carrasco Marín
  Ingeniería para la Innovación i2TIC, S.L.
  Tlf:  +34 911 12 32 84 Ext: 223
  www.i2tic.com
_