It takes more time after I stopped the indexing. The load firstly was with the first node and after I restarted the indexing process the load with changed to the second node the first node worked properly.
Thanks, Mahmoud On Mon, Oct 16, 2017 at 5:29 PM, Emir Arnautović < emir.arnauto...@sematext.com> wrote: > Does the load stops when you stop indexing or it last for some more time? > Is it always one node that behaves like this and it starts as soon as you > start indexing? Is load different between nodes when you are doing lighter > indexing? > > -- > Monitoring - Log Management - Alerting - Anomaly Detection > Solr & Elasticsearch Consulting Support Training - http://sematext.com/ > > > > > On 16 Oct 2017, at 13:35, Mahmoud Almokadem <prog.mahm...@gmail.com> > wrote: > > > > The transition of the load happened after I restarted the bulk insert > > process. > > > > The size of the index on each server about 500GB. > > > > There are about 8 warnings on each server for "Not found segment file" > like > > that > > > > Error getting file length for [segments_2s4] > > > > java.nio.file.NoSuchFileException: > > /media/ssd_losedata/solr-home/data/documents_online_shard16_ > replica_n1/data/index/segments_2s4 > > at > > java.base/sun.nio.fs.UnixException.translateToIOException( > UnixException.java:92) > > at > > java.base/sun.nio.fs.UnixException.rethrowAsIOException( > UnixException.java:111) > > at > > java.base/sun.nio.fs.UnixException.rethrowAsIOException( > UnixException.java:116) > > at > > java.base/sun.nio.fs.UnixFileAttributeViews$Basic.readAttributes( > UnixFileAttributeViews.java:55) > > at > > java.base/sun.nio.fs.UnixFileSystemProvider.readAttributes( > UnixFileSystemProvider.java:145) > > at > > java.base/sun.nio.fs.LinuxFileSystemProvider.readAttributes( > LinuxFileSystemProvider.java:99) > > at java.base/java.nio.file.Files.readAttributes(Files.java:1755) > > at java.base/java.nio.file.Files.size(Files.java:2369) > > at org.apache.lucene.store.FSDirectory.fileLength(FSDirectory.java:243) > > at > > org.apache.lucene.store.NRTCachingDirectory.fileLength( > NRTCachingDirectory.java:128) > > at > > org.apache.solr.handler.admin.LukeRequestHandler.getFileLength( > LukeRequestHandler.java:611) > > at > > org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo( > LukeRequestHandler.java:584) > > at > > org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody( > LukeRequestHandler.java:136) > > at > > org.apache.solr.handler.RequestHandlerBase.handleRequest( > RequestHandlerBase.java:177) > > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2474) > > at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:720) > > at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:526) > > at > > org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDispatchFilter.java:378) > > at > > org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDispatchFilter.java:322) > > at > > org.eclipse.jetty.servlet.ServletHandler$CachedChain. > doFilter(ServletHandler.java:1691) > > at > > org.eclipse.jetty.servlet.ServletHandler.doHandle( > ServletHandler.java:582) > > at > > org.eclipse.jetty.server.handler.ScopedHandler.handle( > ScopedHandler.java:143) > > at > > org.eclipse.jetty.security.SecurityHandler.handle( > SecurityHandler.java:548) > > at > > org.eclipse.jetty.server.session.SessionHandler. > doHandle(SessionHandler.java:226) > > at > > org.eclipse.jetty.server.handler.ContextHandler. > doHandle(ContextHandler.java:1180) > > at org.eclipse.jetty.servlet.ServletHandler.doScope( > ServletHandler.java:512) > > at > > org.eclipse.jetty.server.session.SessionHandler. > doScope(SessionHandler.java:185) > > at > > org.eclipse.jetty.server.handler.ContextHandler. > doScope(ContextHandler.java:1112) > > at > > org.eclipse.jetty.server.handler.ScopedHandler.handle( > ScopedHandler.java:141) > > at > > org.eclipse.jetty.server.handler.ContextHandlerCollection.handle( > ContextHandlerCollection.java:213) > > at > > org.eclipse.jetty.server.handler.HandlerCollection. > handle(HandlerCollection.java:119) > > at > > org.eclipse.jetty.server.handler.HandlerWrapper.handle( > HandlerWrapper.java:134) > > at > > org.eclipse.jetty.rewrite.handler.RewriteHandler.handle( > RewriteHandler.java:335) > > at > > org.eclipse.jetty.server.handler.HandlerWrapper.handle( > HandlerWrapper.java:134) > > at org.eclipse.jetty.server.Server.handle(Server.java:534) > > at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:320) > > at > > org.eclipse.jetty.server.HttpConnection.onFillable( > HttpConnection.java:251) > > at > > org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded( > AbstractConnection.java:273) > > at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95) > > at > > org.eclipse.jetty.io.SelectChannelEndPoint$2.run( > SelectChannelEndPoint.java:93) > > at > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume. > executeProduceConsume(ExecuteProduceConsume.java:303) > > at > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume. > produceConsume(ExecuteProduceConsume.java:148) > > at > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run( > ExecuteProduceConsume.java:136) > > at > > org.eclipse.jetty.util.thread.QueuedThreadPool.runJob( > QueuedThreadPool.java:671) > > at > > org.eclipse.jetty.util.thread.QueuedThreadPool$2.run( > QueuedThreadPool.java:589) > > at java.base/java.lang.Thread.run(Thread.java:844) > > > > On Mon, Oct 16, 2017 at 1:08 PM, Emir Arnautović < > > emir.arnauto...@sematext.com> wrote: > > > >> I did not look at graph details - now I see that it is over 3h time > span. > >> It seems that there was a load on the other server before this one and > >> ended with 14GB read spike and 10GB write spike, just before load > started > >> on this server. Do you see any errors or suspicious logs lines? > >> How big is your index? > >> > >> Emir > >> -- > >> Monitoring - Log Management - Alerting - Anomaly Detection > >> Solr & Elasticsearch Consulting Support Training - http://sematext.com/ > >> > >> > >> > >>> On 16 Oct 2017, at 12:39, Mahmoud Almokadem <prog.mahm...@gmail.com> > >> wrote: > >>> > >>> Yes, it's constantly since I started this bulk indexing process. > >>> As you see the write operations on the loaded server are 3x the normal > >>> server despite Disk writes not 3x times. > >>> > >>> Mahmoud > >>> > >>> > >>> On Mon, Oct 16, 2017 at 12:32 PM, Emir Arnautović < > >>> emir.arnauto...@sematext.com> wrote: > >>> > >>>> Hi Mahmoud, > >>>> Is this something that you see constantly? Network charts suggests > that > >>>> your servers are loaded equally and as you said - you are not using > >> routing > >>>> so expected. Disk read/write and CPU are not equal and it is expected > to > >>>> not be equal during heavy indexing since it also triggers segment > merges > >>>> which require those resources. Even if host same documents (e.g. > leader > >> and > >>>> replica) merges are not likely to happen at the same time and you can > >>>> expect to see such cases. > >>>> > >>>> Thanks, > >>>> Emir > >>>> -- > >>>> Monitoring - Log Management - Alerting - Anomaly Detection > >>>> Solr & Elasticsearch Consulting Support Training - > http://sematext.com/ > >>>> > >>>> > >>>> > >>>>> On 16 Oct 2017, at 11:58, Mahmoud Almokadem <prog.mahm...@gmail.com> > >>>> wrote: > >>>>> > >>>>> Here are the screen shots for the two server metrics on Amazon > >>>>> > >>>>> https://ibb.co/kxBQam > >>>>> https://ibb.co/fn0Jvm > >>>>> https://ibb.co/kUpYT6 > >>>>> > >>>>> > >>>>> > >>>>> On Mon, Oct 16, 2017 at 11:37 AM, Mahmoud Almokadem < > >>>> prog.mahm...@gmail.com> > >>>>> wrote: > >>>>> > >>>>>> Hi Emir, > >>>>>> > >>>>>> We doesn't use routing. > >>>>>> > >>>>>> Servers is already balanced and the number of documents on each > shard > >>>> are > >>>>>> approximately the same. > >>>>>> > >>>>>> Nothing running on the servers except Solr and ZooKeeper. > >>>>>> > >>>>>> I initialized the client as > >>>>>> > >>>>>> String zkHost = "192.168.1.89:2181,192.168.1.99:2181"; > >>>>>> > >>>>>> CloudSolrClient solrCloud = new CloudSolrClient.Builder() > >>>>>> .withZkHost(zkHost) > >>>>>> .build(); > >>>>>> > >>>>>> solrCloud.setIdField("document_id"); > >>>>>> solrCloud.setDefaultCollection(collection); > >>>>>> solrCloud.setRequestWriter(new BinaryRequestWriter()); > >>>>>> > >>>>>> > >>>>>> And the documents are approximately the same size. > >>>>>> > >>>>>> I Used 10 threads with 10 SolrClients to send data to solr and every > >>>>>> thread send a batch of 1000 documents every time. > >>>>>> > >>>>>> Thanks, > >>>>>> Mahmoud > >>>>>> > >>>>>> > >>>>>> > >>>>>> On Mon, Oct 16, 2017 at 11:01 AM, Emir Arnautović < > >>>>>> emir.arnauto...@sematext.com> wrote: > >>>>>> > >>>>>>> Hi Mahmoud, > >>>>>>> Do you use routing? Are your servers equally balanced - do you end > up > >>>>>>> having approximately the same number of documents hosted on both > >>>> servers > >>>>>>> (counted all shards)? > >>>>>>> Do you have anything else running on those servers? > >>>>>>> How do you initialise your SolrJ client? > >>>>>>> Are documents of similar size? > >>>>>>> > >>>>>>> Thanks, > >>>>>>> Emir > >>>>>>> -- > >>>>>>> Monitoring - Log Management - Alerting - Anomaly Detection > >>>>>>> Solr & Elasticsearch Consulting Support Training - > >>>> http://sematext.com/ > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>>> On 16 Oct 2017, at 10:46, Mahmoud Almokadem < > prog.mahm...@gmail.com > >>> > >>>>>>> wrote: > >>>>>>>> > >>>>>>>> We've installed SolrCloud 7.0.1 with two nodes and 8 shards per > >> node. > >>>>>>>> > >>>>>>>> The configurations and the specs of the two servers are identical. > >>>>>>>> > >>>>>>>> When running bulk indexing using SolrJ we see one of the servers > is > >>>>>>> fully > >>>>>>>> loaded as you see on the images and the other is normal. > >>>>>>>> > >>>>>>>> Images URLs: > >>>>>>>> > >>>>>>>> https://ibb.co/jkE6gR > >>>>>>>> https://ibb.co/hyzvam > >>>>>>>> https://ibb.co/mUpvam > >>>>>>>> https://ibb.co/e4bxo6 > >>>>>>>> > >>>>>>>> How can I figure this issue? > >>>>>>>> > >>>>>>>> Thanks, > >>>>>>>> Mahmoud > >>>>>>> > >>>>>>> > >>>>>> > >>>> > >>>> > >> > >> > >