[prev in list] [next in list] [prev in thread] [next in thread]
List: solr-dev
Subject: [jira] [Commented] (SOLR-10277) On 'downnode', lots of wasteful mutations are done to ZK
From: "Varun Thacker (JIRA)" <jira () apache ! org>
Date: 2017-03-31 18:04:41
Message-ID: JIRA.13055739.1489442149000.178552.1490983481643 () Atlassian ! JIRA
[Download RAW message or body]
[ https://issues.apache.org/jira/browse/SOLR-10277?page=com.atlassian.jira.plugin. \
system.issuetabpanels:comment-tabpanel&focusedCommentId=15951407#comment-15951407 ]
Varun Thacker commented on SOLR-10277:
--------------------------------------
Hi Scott,
Any comments on the last patch apart from addressing the nocommit ?
Joshua curious if you deployed this on your cluster and how much improvements did you \
see?
Regardless this should be a good candidate for Solr 6.5.1 as the fix is pretty safe. \
[~joel.bernstein] let me know if you're okay to commit this to the 6_5 branch as well
> On 'downnode', lots of wasteful mutations are done to ZK
> --------------------------------------------------------
>
> Key: SOLR-10277
> URL: https://issues.apache.org/jira/browse/SOLR-10277
> Project: Solr
> Issue Type: Bug
> Security Level: Public(Default Security Level. Issues are Public)
> Components: SolrCloud
> Affects Versions: 5.5.3, 5.5.4, 6.0.1, 6.2.1, 6.3, 6.4.2
> Reporter: Joshua Humphries
> Assignee: Scott Blum
> Labels: leader, zookeeper
> Attachments: SOLR-10277-5.5.3.patch, SOLR-10277.patch
>
>
> When a node restarts, it submits a single 'downnode' message to the overseer's \
> state update queue. When the overseer processes the message, it does way more \
> writes to ZK than necessary. In our cluster of 48 hosts, the majority of \
> collections have only 1 shard and 1 replica. So a single node restarting should \
> only result in ~1/40th of the collections being updated with new replica states (to \
> indicate the node that is no longer active). However, the current logic in \
> NodeMutator#downNode always updates *every* collection. So we end up having to do \
> rolling restarts very slowly to avoid having a severe outage due to the overseer \
> having to do way too much work for each host that is restarted. And subsequent \
> shards becoming leader can't get processed until the `downnode` message is fully \
> processed. So a fast rolling restart can result in the overseer queue growing \
> incredibly large and nearly all shards winding up in a leader-less state until that \
> backlog is processed. The fix is a trivial logic change to only add a \
> ZkWriteCommand for collections that actually have an impacted replica.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@lucene.apache.org
For additional commands, e-mail: dev-help@lucene.apache.org
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic