[ 
https://issues.apache.org/jira/browse/PHOENIX-3654?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15894985#comment-15894985
 ] 

Rahul Shrivastava commented on PHOENIX-3654:
--------------------------------------------


--Do we need to pull in curator for this? Is Curator buying enough over the 
standard Java ZK API to make it worth it? -- Curator has service 
discovery/service registration mechanism. We can use the Zk client to achieve 
the same. Will discuss with [~jamestaylor] on this. 

--How will you know how many connections each PQS instance has from ZooKeeper? 
You said the znode contents were just going to be a JDBC url, right? - I think 
[~apurtell] suggested that store a Json in zknode.

--How are multiple clients of the load balancer aware of each others' 
connections? Does this imply that the LB is actually a daemon running outside 
of the clients' JVM? - Clients will only know the load on PQS servers via the 
zookeeper via ZKnode watcher. It will not be aware of each other load. 

--Consider the following case:
--PQS is running, maintaining its znode in ZK
--Client begins check to PQS
--PQS swaps-out/pauses
--Client times out and thinks PQS node is bad
--PQS recovers without losing ZK session
--This would result in the client never using this PQS instance again.

                             Good point. The client will keep its location of 
PQS updated via Zookeeper. There could be a mechanism to update the full PQS 
load data via by going through all znodes data. I need to think more on this 
issue.  

-- And, on that note, how how PQS failure be handled (or not) by this design? - 
If PQS fails, their is no failover of the sessions. The client has to retry. 

-- A concern is that there is an upper limit on the number of watchers that a 
ZooKeeper server can support. Considering HBase and whatever else is already 
setting/using watchers of its own, we should try to make sure to be "good 
citizens". ZOOKEEPER-1177 for context. As you've described, this would result 
in each client keeping a zookeeper connection (which, itself, is overhead) open 
as well as the watchers set server-side. Architecturally, this is what makes ZK 
a bad choice for a load balancer – once you find a PQS to talk to, you don't 
need to maintain the ZK connection anymore. You only need the ZK connection 
when the PQS instance you were talking to goes away.

                 I will discuss more with [~jamestaylor] [~apurtell] on this. I 
was not aware of the connection limit issue and surely this will limit ability 
of single ZK cluster to support large set of PQS nodes. 


> Load Balancer for thin client
> -----------------------------
>
>                 Key: PHOENIX-3654
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-3654
>             Project: Phoenix
>          Issue Type: New Feature
>    Affects Versions: 4.8.0
>         Environment: Linux 3.13.0-107-generic kernel, v4.9.0-HBase-0.98
>            Reporter: Rahul Shrivastava
>             Fix For: 4.9.0
>
>         Attachments: LoadBalancerDesign.pdf
>
>   Original Estimate: 240h
>  Remaining Estimate: 240h
>
> We have been having internal discussion on load balancer for thin client for 
> PQS. The general consensus we have is to have an embedded load balancer with 
> the thin client instead of using external load balancer such as haproxy. The 
> idea is to not to have another layer between client and PQS. This reduces 
> operational cost for system, which currently leads to delay in executing 
> projects.
> But this also comes with challenge of having an embedded load balancer which 
> can maintain sticky sessions, do fair load balancing knowing the load 
> downstream of PQS server. In addition, load balancer needs to know location 
> of multiple PQS server. Now, the thin client needs to keep track of PQS 
> servers via zookeeper ( or other means). 
> In the new design, the client ( PQS client) , it is proposed to  have an 
> embedded load balancer.
> Where will the load Balancer sit ?
> The load load balancer will embedded within the app server client.  
> How will the load balancer work ? 
> Load balancer will contact zookeeper to get location of PQS. In this case, 
> PQS needs to register to ZK itself once it comes online. Zookeeper location 
> is in hbase-site.xml. It will maintain a small cache of connection to the 
> PQS. When a request comes in, it will check for an open connection from the 
> cache. 
> How will load balancer know load on PQS ?
> To start with, it will pick a random open connection to PQS. This means that 
> load balancer does not know PQS load. Later , we can augment the code so that 
> thin client can receive load info from PQS and make intelligent decisions.  
> How will load balancer maintain sticky sessions ?
> While we still need to investigate how to implement sticky sessions. We can 
> look for some open source implementation for the same.
> How will PQS register itself to service locator ?
> PQS will have location of zookeeper in hbase-site.xml and it would register 
> itself to the zookeeper. Thin client will find out PQS location using 
> zookeeper.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to