[jira] Updated: (ZOOKEEPER-483) ZK fataled on me, and ugly

2009-08-10 Thread Benjamin Reed (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benjamin Reed updated ZOOKEEPER-483:


Status: Open  (was: Patch Available)

 ZK fataled on me, and ugly
 --

 Key: ZOOKEEPER-483
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-483
 Project: Zookeeper
  Issue Type: Bug
Affects Versions: 3.1.1
Reporter: ryan rawson
Assignee: Benjamin Reed
 Fix For: 3.2.1, 3.3.0

 Attachments: zklogs.tar.gz, ZOOKEEPER-483.patch, ZOOKEEPER-483.patch, 
 ZOOKEEPER-483.patch


 here are the part of the log whereby my zookeeper instance crashed, taking 3 
 out of 5 down, and thus ruining the quorum for all clients:
 2009-07-23 12:29:06,769 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5161350 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:00,756 WARN org.apache.zookeeper.server.quorum.Follower: 
 Exception when following the leader
 java.io.EOFException
 at java.io.DataInputStream.readInt(DataInputStream.java:375)
 at 
 org.apache.jute.BinaryInputArchive.readInt(BinaryInputArchive.java:63)
 at 
 org.apache.zookeeper.server.quorum.QuorumPacket.deserialize(QuorumPacket.java:65)
 at 
 org.apache.jute.BinaryInputArchive.readRecord(BinaryInputArchive.java:108)
 at 
 org.apache.zookeeper.server.quorum.Follower.readPacket(Follower.java:114)
 at 
 org.apache.zookeeper.server.quorum.Follower.followLeader(Follower.java:243)
 at 
 org.apache.zookeeper.server.quorum.QuorumPeer.run(QuorumPeer.java:494)
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.168:39489]
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0578 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46797]
 2009-07-23 12:29:06,771 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa013e NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:33998]
 2009-07-23 12:29:06,771 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5160593 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:06,808 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e02bb NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.158:53758]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13e4 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.154:58681]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691382 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59967]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb1354 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.163:49957]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13cd NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.150:34212]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691383 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46813]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59956]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e139b NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.156:55138]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e1398 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.167:41257]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161355 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:34032]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d516011c NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected 

[jira] Updated: (ZOOKEEPER-483) ZK fataled on me, and ugly

2009-08-10 Thread Benjamin Reed (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benjamin Reed updated ZOOKEEPER-483:


Status: Patch Available  (was: Open)

 ZK fataled on me, and ugly
 --

 Key: ZOOKEEPER-483
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-483
 Project: Zookeeper
  Issue Type: Bug
Affects Versions: 3.1.1
Reporter: ryan rawson
Assignee: Benjamin Reed
 Fix For: 3.2.1, 3.3.0

 Attachments: zklogs.tar.gz, ZOOKEEPER-483.patch, ZOOKEEPER-483.patch, 
 ZOOKEEPER-483.patch


 here are the part of the log whereby my zookeeper instance crashed, taking 3 
 out of 5 down, and thus ruining the quorum for all clients:
 2009-07-23 12:29:06,769 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5161350 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:00,756 WARN org.apache.zookeeper.server.quorum.Follower: 
 Exception when following the leader
 java.io.EOFException
 at java.io.DataInputStream.readInt(DataInputStream.java:375)
 at 
 org.apache.jute.BinaryInputArchive.readInt(BinaryInputArchive.java:63)
 at 
 org.apache.zookeeper.server.quorum.QuorumPacket.deserialize(QuorumPacket.java:65)
 at 
 org.apache.jute.BinaryInputArchive.readRecord(BinaryInputArchive.java:108)
 at 
 org.apache.zookeeper.server.quorum.Follower.readPacket(Follower.java:114)
 at 
 org.apache.zookeeper.server.quorum.Follower.followLeader(Follower.java:243)
 at 
 org.apache.zookeeper.server.quorum.QuorumPeer.run(QuorumPeer.java:494)
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.168:39489]
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0578 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46797]
 2009-07-23 12:29:06,771 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa013e NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:33998]
 2009-07-23 12:29:06,771 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5160593 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:06,808 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e02bb NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.158:53758]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13e4 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.154:58681]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691382 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59967]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb1354 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.163:49957]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13cd NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.150:34212]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691383 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46813]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59956]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e139b NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.156:55138]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e1398 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.167:41257]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161355 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:34032]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d516011c NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected 

[jira] Updated: (ZOOKEEPER-504) ClassCastException in LedgerManagementProcessor

2009-08-10 Thread Mahadev konar (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahadev konar updated ZOOKEEPER-504:


  Description: 

java.lang.ClassCastException: 
org.apache.bookkeeper.client.LedgerManagementProcessor$OpenLedgerOp cannot be 
cast to org.apache.bookkeeper.client.LedgerManagementProcessor$CloseLedgerOp
at 
org.apache.bookkeeper.client.LedgerManagementProcessor.processResult(LedgerManagementProcessor.java:1083)

This seems to be happening because its a nested switch case statement. And the 
OPEN: case, doesn't ever call a break. It only calls a break from the inner 
switch-case and hence falls through into the CLOSE: case.


  was:


java.lang.ClassCastException: 
org.apache.bookkeeper.client.LedgerManagementProcessor$OpenLedgerOp cannot be 
cast to org.apache.bookkeeper.client.LedgerManagementProcessor$CloseLedgerOp
at 
org.apache.bookkeeper.client.LedgerManagementProcessor.processResult(LedgerManagementProcessor.java:1083)

This seems to be happening because its a nested switch case statement. And the 
OPEN: case, doesn't ever call a break. It only calls a break from the inner 
switch-case and hence falls through into the CLOSE: case.


Fix Version/s: 3.3.0
 Assignee: Utkarsh Srivastava

 ClassCastException in LedgerManagementProcessor
 ---

 Key: ZOOKEEPER-504
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-504
 Project: Zookeeper
  Issue Type: Bug
  Components: contrib-bookkeeper
Affects Versions: 3.2.0
Reporter: Utkarsh Srivastava
Assignee: Utkarsh Srivastava
 Fix For: 3.3.0

 Attachments: ZOOKEEPER-504.1.patch, ZOOKEEPER-504.patch, 
 ZOOKEEPER-504.patch


 java.lang.ClassCastException: 
 org.apache.bookkeeper.client.LedgerManagementProcessor$OpenLedgerOp cannot be 
 cast to org.apache.bookkeeper.client.LedgerManagementProcessor$CloseLedgerOp
   at 
 org.apache.bookkeeper.client.LedgerManagementProcessor.processResult(LedgerManagementProcessor.java:1083)
 This seems to be happening because its a nested switch case statement. And 
 the OPEN: case, doesn't ever call a break. It only calls a break from the 
 inner switch-case and hence falls through into the CLOSE: case.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-499) electionAlg should default to FLE (3) - regression

2009-08-10 Thread Patrick Hunt (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Patrick Hunt updated ZOOKEEPER-499:
---

Attachment: ZOOKEEPER-499_br3.2.patch
ZOOKEEPER-499.patch

updated patch based on ben's comments:

1) using the name of the package is fine, this is actually taking advantage of 
a feature of log4j
http://logging.apache.org/log4j/1.2/manual.html
Each enabled logging request for a given logger will be forwarded to all the 
appenders in that logger as well as the appenders higher in the hierarchy.

2) I was removing the appender from the wrong logger, I've updated the patch to 
correctly remove.

 electionAlg should default to FLE (3) - regression
 --

 Key: ZOOKEEPER-499
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-499
 Project: Zookeeper
  Issue Type: Bug
  Components: server, tests
Affects Versions: 3.2.0
Reporter: Patrick Hunt
Assignee: Patrick Hunt
Priority: Blocker
 Fix For: 3.2.1, 3.3.0

 Attachments: ZOOKEEPER-499.patch, ZOOKEEPER-499.patch, 
 ZOOKEEPER-499_br3.2.patch, ZOOKEEPER-499_br3.2.patch


 there's a regression in 3.2 - electionAlg is no longer defaulting to 3 
 (incorrectly defaults to 0)
 also - need to have tests to validate this

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-499) electionAlg should default to FLE (3) - regression

2009-08-10 Thread Patrick Hunt (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Patrick Hunt updated ZOOKEEPER-499:
---

Status: Patch Available  (was: Open)

 electionAlg should default to FLE (3) - regression
 --

 Key: ZOOKEEPER-499
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-499
 Project: Zookeeper
  Issue Type: Bug
  Components: server, tests
Affects Versions: 3.2.0
Reporter: Patrick Hunt
Assignee: Patrick Hunt
Priority: Blocker
 Fix For: 3.2.1, 3.3.0

 Attachments: ZOOKEEPER-499.patch, ZOOKEEPER-499.patch, 
 ZOOKEEPER-499_br3.2.patch, ZOOKEEPER-499_br3.2.patch


 there's a regression in 3.2 - electionAlg is no longer defaulting to 3 
 (incorrectly defaults to 0)
 also - need to have tests to validate this

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-496) zookeeper-tree utility for export, import and incremental updates

2009-08-10 Thread Anirban Roy (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anirban Roy updated ZOOKEEPER-496:
--

 Tags: zookeeper tree trree-data utility zktree
Fix Version/s: 3.3.0
Affects Version/s: (was: 3.1.1)
   3.3.0
 Release Note: 
==
zktreeutil - Zookeeper Tree Data Utility
Author: Anirban Roy
Organization: Yahoo Inc.
==

zktreeutil program is intendened to manage and manipulate zk-tree data quickly, 
effi-
ciently and with ease.The utility operates on free-form ZK-tree and hence can 
be used
for any cluster managed by Zookeeper. Here are the basic functionalities -

EXPORT: The whole ZK-tree is exported into a XML file. This helps in capturing 
a cur-
rent snapshot of the data for backup/analysis. In addition, the data can be 
dumped
as tree-like text format which is easily readable.

IMPORT: The ZK-tree can be imported from XML into ZK cluster. This helps in 
priming
the new ZK cluster with static configuration. The import can be non-intrusive by
making only the additions in the existing data.

DIFF: Creates a diff between live ZK data vs data saved in XML file. Diff can 
ignore
some ZK-tree branches (possibly dynamic data) on reading the flag from XML file.

UPDATE: Make the incremental changes into the live ZK-tree from saved XML, 
essentia-
illy after running the diff.

DUMP: Dumps the ZK-tree on the standard output device reading either from live 
ZK
server or XML file.

The exported ZK data into XML file can be shortened by only keeping the static 
ZK
nodes which are required to prime a cluster. The dynamic zk nodes (created 
on-the-
fly) can be ignored by setting a 'ignore' attribute at the root node of the 
dynamic
subtree (see tests/zk_sample.xml), possibly deleting all inner ZK nodes under 
that.
Once ignored, the whole subtree is ignored during DIFF, UPDATE and WRITE.

Pre-requisites
--
1. Linux system with 2.6.X kernel.
2. Zookeeper C client library = 3.X.X
3. Development build libraries:
  a. libxml2 = 2.7.3
  b. log4cxx = 0.10.0

Build instructions
--
1. cd into this directory
2. autoreconf -if
3. ./configure
4. make
5. 'zktreeutil' binary created under src directory

Testing zktreeutil
--
1.  Run Zookeeper server locally on port 2181
2.  export LD_LIBRARY_PATH=../../c/.libs/:/usr/local/lib/
3.  ./src/zktreeutil --help # show help
4.  ./src/zktreeutil --zookeeper=localhost:2181 --import 
--xmlfile=tests/zk_sample.xml 2/dev/null   # import sample ZK tree
5.  ./src/zktreeutil --zookeeper=localhost:2181 --dump 2/dev/null  
 # dump on the terminal
6.  ./src/zktreeutil --xmlfile=zk_sample.xml -D 2/dev/null 
 # dump the xml data
7.  Change zk_sample.xml with adding/deleting/chaging some nodes
8.  ./src/zktreeutil --zookeeper=localhost:2181 --diff --xmlfile=zk_sample.xml 
2/dev/null   # take a diff of changes
9.  ./src/zktreeutil --zookeeper=localhost:2181 --export 2/dev/null  
zk_sample2.xml# export the mofied ZK tree
10. ./src/zktreeutil --zookeeper=localhost:2181 --update 
--xmlfile=zk_sample.xml 2/dev/null # update with incr. changes
11. ./src/zktreeutil --zookeeper=localhost:2181 --import --force 
--xmlfile=zk_sample2.xml 2/dev/null# re-prime the ZK tree

   Status: Patch Available  (was: Open)

New feature

 zookeeper-tree utility for export, import and incremental updates
 -

 Key: ZOOKEEPER-496
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-496
 Project: Zookeeper
  Issue Type: New Feature
  Components: contrib
Affects Versions: 3.3.0
 Environment: RHEL 4.6, libxml2
Reporter: Anirban Roy
 Fix For: 3.3.0

   Original Estimate: 168h
  Remaining Estimate: 168h

 I propose and plan to develop a ZK-tree utility for managing cluster data (ZK 
 tree) quickly and efficiently. The utility operates on free-form ZK-tree and 
 hence can be used for any cluster managed by Zookeeper. Here are the basic 
 functionalities -
 Export ZK-tree: The whole ZK-tree is exported into a XML file. This helps in 
 capturing a current snapshot of the data for backup/analysis. In addition, 
 the data can be dumped as tree-like text format which is easily readable.
 Import ZK-tree: The ZK-tree can be imported from XML into ZK cluster. This 
 helps in priming the new ZK cluster with static configuration. The import can 
 be non-intrusive by making only the additions in the existing data.
 Diff ZK-tree: Creates a diff between live ZK data vs data saved in XML file. 
 Diff can ignore some ZK-tree branches (possibly dynamic data) on reading 

[jira] Updated: (ZOOKEEPER-496) zookeeper-tree utility for export, import and incremental updates

2009-08-10 Thread Anirban Roy (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anirban Roy updated ZOOKEEPER-496:
--

Attachment: zktreeutil.diff

zktreeutil diff with the latest code in svn trunk.

 zookeeper-tree utility for export, import and incremental updates
 -

 Key: ZOOKEEPER-496
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-496
 Project: Zookeeper
  Issue Type: New Feature
  Components: contrib
Affects Versions: 3.3.0
 Environment: RHEL 4.6, libxml2
Reporter: Anirban Roy
 Fix For: 3.3.0

 Attachments: zktreeutil.diff

   Original Estimate: 168h
  Remaining Estimate: 168h

 I propose and plan to develop a ZK-tree utility for managing cluster data (ZK 
 tree) quickly and efficiently. The utility operates on free-form ZK-tree and 
 hence can be used for any cluster managed by Zookeeper. Here are the basic 
 functionalities -
 Export ZK-tree: The whole ZK-tree is exported into a XML file. This helps in 
 capturing a current snapshot of the data for backup/analysis. In addition, 
 the data can be dumped as tree-like text format which is easily readable.
 Import ZK-tree: The ZK-tree can be imported from XML into ZK cluster. This 
 helps in priming the new ZK cluster with static configuration. The import can 
 be non-intrusive by making only the additions in the existing data.
 Diff ZK-tree: Creates a diff between live ZK data vs data saved in XML file. 
 Diff can ignore some ZK-tree branches (possibly dynamic data) on reading the 
 flag from XML file.
 Update ZK-tree: Make the incremental changes into the live ZK-tree from saved 
 XML, essentially after running the diff.
 Thanks
 Anirban

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-496) zookeeper-tree utility for export, import and incremental updates

2009-08-10 Thread Mahadev konar (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahadev konar updated ZOOKEEPER-496:


Affects Version/s: (was: 3.3.0)
 Assignee: Anirban Roy

 zookeeper-tree utility for export, import and incremental updates
 -

 Key: ZOOKEEPER-496
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-496
 Project: Zookeeper
  Issue Type: New Feature
  Components: contrib
 Environment: RHEL 4.6, libxml2
Reporter: Anirban Roy
Assignee: Anirban Roy
 Fix For: 3.3.0

 Attachments: zktreeutil.diff

   Original Estimate: 168h
  Remaining Estimate: 168h

 I propose and plan to develop a ZK-tree utility for managing cluster data (ZK 
 tree) quickly and efficiently. The utility operates on free-form ZK-tree and 
 hence can be used for any cluster managed by Zookeeper. Here are the basic 
 functionalities -
 Export ZK-tree: The whole ZK-tree is exported into a XML file. This helps in 
 capturing a current snapshot of the data for backup/analysis. In addition, 
 the data can be dumped as tree-like text format which is easily readable.
 Import ZK-tree: The ZK-tree can be imported from XML into ZK cluster. This 
 helps in priming the new ZK cluster with static configuration. The import can 
 be non-intrusive by making only the additions in the existing data.
 Diff ZK-tree: Creates a diff between live ZK data vs data saved in XML file. 
 Diff can ignore some ZK-tree branches (possibly dynamic data) on reading the 
 flag from XML file.
 Update ZK-tree: Make the incremental changes into the live ZK-tree from saved 
 XML, essentially after running the diff.
 Thanks
 Anirban

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-474) add compile, test, and improved package targets to zkperl build.xml

2009-08-10 Thread Chris Darroch (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-474?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741503#action_12741503
 ] 

Chris Darroch commented on ZOOKEEPER-474:
-

Nothing yet, sorry.  Might be a couple of weeks before I get a chance to look; 
it's extra busy at work at the moment.

 add compile, test, and improved package targets to zkperl build.xml
 ---

 Key: ZOOKEEPER-474
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-474
 Project: Zookeeper
  Issue Type: Improvement
  Components: contrib
Affects Versions: 3.2.0
Reporter: Chris Darroch
Assignee: Chris Darroch
 Fix For: 3.2.1, 3.3.0

 Attachments: zk474_testout.txt, ZOOKEEPER-474.patch


 This patch adds compile and test targets to the zkperl build.xml, and tweaks 
 the package target a little to use the manifest file.
 For me, ant compile, ant test, and ant clean all work (from scratch, in 
 each case) when using Ant in the local src/contrib/zkperl directory.
 Further, ant package in the top-level directory seems to continue to build 
 zkperl along with everything else, and leaves out the build.xml and 
 t/zkServer.sh files, which is appropriate.
 From what I can see, the top-level build.xml doesn't actually invoke the 
 test-contrib target, so I'm not sure if there's a way to integrate the zkperl 
 tests into the main hudson automated test process, but that would be ideal, 
 if at all possible.  I feel like I've seen comments to the effect that the 
 zkpython tests are run automatically, but I'm not sure if that's actually 
 true or not.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-474) add compile, test, and improved package targets to zkperl build.xml

2009-08-10 Thread Patrick Hunt (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-474?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Patrick Hunt updated ZOOKEEPER-474:
---

Fix Version/s: (was: 3.2.1)

moving out of 3.2.1, won't be ready for release timeframe, can go into 3.3


 add compile, test, and improved package targets to zkperl build.xml
 ---

 Key: ZOOKEEPER-474
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-474
 Project: Zookeeper
  Issue Type: Improvement
  Components: contrib
Affects Versions: 3.2.0
Reporter: Chris Darroch
Assignee: Chris Darroch
 Fix For: 3.3.0

 Attachments: zk474_testout.txt, ZOOKEEPER-474.patch


 This patch adds compile and test targets to the zkperl build.xml, and tweaks 
 the package target a little to use the manifest file.
 For me, ant compile, ant test, and ant clean all work (from scratch, in 
 each case) when using Ant in the local src/contrib/zkperl directory.
 Further, ant package in the top-level directory seems to continue to build 
 zkperl along with everything else, and leaves out the build.xml and 
 t/zkServer.sh files, which is appropriate.
 From what I can see, the top-level build.xml doesn't actually invoke the 
 test-contrib target, so I'm not sure if there's a way to integrate the zkperl 
 tests into the main hudson automated test process, but that would be ideal, 
 if at all possible.  I feel like I've seen comments to the effect that the 
 zkpython tests are run automatically, but I'm not sure if that's actually 
 true or not.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-498) Unending Leader Elections : WAN configuration

2009-08-10 Thread Patrick Hunt (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741526#action_12741526
 ] 

Patrick Hunt commented on ZOOKEEPER-498:


Looks good to me, I tested out a number of q configs (2/3/4/5/7/9) with various 
weights/groups (granted, not exhaustive), and the quorum always formed. Was 
able to connect client, also tried stopping/starting servers to ensure rejoin 
of quorum. Looks good to me.

Also - the logging seems better, no longer errors for things that are not 
errors.


 Unending Leader Elections : WAN configuration
 -

 Key: ZOOKEEPER-498
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-498
 Project: Zookeeper
  Issue Type: Bug
  Components: leaderElection
Affects Versions: 3.2.0
 Environment: Each machine:
 CentOS 5.2 64-bit
 2GB ram
 java version 1.6.0_13
 Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed 
 Network Topology:
 DC : central data center
 POD(N): remote data center
 Zookeeper Topology:
 Leaders may be elected only in DC (weight = 1)
 Only followers are elected in PODS (weight = 0)
Reporter: Todd Greenwood-Geer
Assignee: Flavio Paiva Junqueira
Priority: Critical
 Fix For: 3.2.1, 3.3.0

 Attachments: dc-zook-logs-01.tar.gz, pod-zook-logs-01.tar.gz, 
 zk498-test.tar.gz, zoo.cfg, ZOOKEEPER-498.patch, ZOOKEEPER-498.patch, 
 ZOOKEEPER-498.patch, ZOOKEEPER-498.patch


 In a WAN configuration, ZooKeeper is endlessly electing, terminating, and 
 re-electing a ZooKeeper leader. The WAN configuration involves two groups, a 
 central DC group of ZK servers that have a voting weight = 1, and a group of 
 servers in remote pods with a voting weight of 0.
 What we expect to see is leaders elected only in the DC, and the pods to 
 contain only followers. What we are seeing is a continuous cycling of 
 leaders. We have seen this consistently with 3.2.0, 3.2.0 + recommended 
 patches (473, 479, 481, 491), and now release 3.2.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-498) Unending Leader Elections : WAN configuration

2009-08-10 Thread Patrick Hunt (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741529#action_12741529
 ] 

Patrick Hunt commented on ZOOKEEPER-498:


btw, my testing was on trunk. I'll give branch a try too and report I find any 
issues (otw assume it's ok)


 Unending Leader Elections : WAN configuration
 -

 Key: ZOOKEEPER-498
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-498
 Project: Zookeeper
  Issue Type: Bug
  Components: leaderElection
Affects Versions: 3.2.0
 Environment: Each machine:
 CentOS 5.2 64-bit
 2GB ram
 java version 1.6.0_13
 Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed 
 Network Topology:
 DC : central data center
 POD(N): remote data center
 Zookeeper Topology:
 Leaders may be elected only in DC (weight = 1)
 Only followers are elected in PODS (weight = 0)
Reporter: Todd Greenwood-Geer
Assignee: Flavio Paiva Junqueira
Priority: Critical
 Fix For: 3.2.1, 3.3.0

 Attachments: dc-zook-logs-01.tar.gz, pod-zook-logs-01.tar.gz, 
 zk498-test.tar.gz, zoo.cfg, ZOOKEEPER-498.patch, ZOOKEEPER-498.patch, 
 ZOOKEEPER-498.patch, ZOOKEEPER-498.patch


 In a WAN configuration, ZooKeeper is endlessly electing, terminating, and 
 re-electing a ZooKeeper leader. The WAN configuration involves two groups, a 
 central DC group of ZK servers that have a voting weight = 1, and a group of 
 servers in remote pods with a voting weight of 0.
 What we expect to see is leaders elected only in the DC, and the pods to 
 contain only followers. What we are seeing is a continuous cycling of 
 leaders. We have seen this consistently with 3.2.0, 3.2.0 + recommended 
 patches (473, 479, 481, 491), and now release 3.2.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-461) Condense ledger configuration in ZooKeeper

2009-08-10 Thread Flavio Paiva Junqueira (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flavio Paiva Junqueira updated ZOOKEEPER-461:
-

Attachment: ZOOKEEPER-461.patch

Attaching preliminary patch. It is not passing unit tests yes.

 Condense ledger configuration in ZooKeeper
 --

 Key: ZOOKEEPER-461
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-461
 Project: Zookeeper
  Issue Type: Improvement
  Components: contrib-bookkeeper
Reporter: Flavio Paiva Junqueira
 Attachments: ZOOKEEPER-461.patch


 We currently use several znodes to represent the configuration and state of a 
 ledger in ZooKeeper. Although this is good for readability, it makes 
 operations such as create, close, and open more complex, in particular the 
 asynchronous versions of these operations. The main idea of this jira is to 
 store as much as possible of the configuration of a ledger as data of a 
 single znode.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-498) Unending Leader Elections : WAN configuration

2009-08-10 Thread Mahadev konar (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741565#action_12741565
 ] 

Mahadev konar commented on ZOOKEEPER-498:
-

flavio, can you upload a patch for 3.2 as well?

 Unending Leader Elections : WAN configuration
 -

 Key: ZOOKEEPER-498
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-498
 Project: Zookeeper
  Issue Type: Bug
  Components: leaderElection
Affects Versions: 3.2.0
 Environment: Each machine:
 CentOS 5.2 64-bit
 2GB ram
 java version 1.6.0_13
 Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed 
 Network Topology:
 DC : central data center
 POD(N): remote data center
 Zookeeper Topology:
 Leaders may be elected only in DC (weight = 1)
 Only followers are elected in PODS (weight = 0)
Reporter: Todd Greenwood-Geer
Assignee: Flavio Paiva Junqueira
Priority: Critical
 Fix For: 3.2.1, 3.3.0

 Attachments: dc-zook-logs-01.tar.gz, pod-zook-logs-01.tar.gz, 
 zk498-test.tar.gz, zoo.cfg, ZOOKEEPER-498.patch, ZOOKEEPER-498.patch, 
 ZOOKEEPER-498.patch, ZOOKEEPER-498.patch


 In a WAN configuration, ZooKeeper is endlessly electing, terminating, and 
 re-electing a ZooKeeper leader. The WAN configuration involves two groups, a 
 central DC group of ZK servers that have a voting weight = 1, and a group of 
 servers in remote pods with a voting weight of 0.
 What we expect to see is leaders elected only in the DC, and the pods to 
 contain only followers. What we are seeing is a continuous cycling of 
 leaders. We have seen this consistently with 3.2.0, 3.2.0 + recommended 
 patches (473, 479, 481, 491), and now release 3.2.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-483) ZK fataled on me, and ugly

2009-08-10 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741572#action_12741572
 ] 

Hadoop QA commented on ZOOKEEPER-483:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12415974/ZOOKEEPER-483.patch
  against trunk revision 802188.

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 3 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed core unit tests.

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
http://hudson.zones.apache.org/hudson/job/Zookeeper-Patch-vesta.apache.org/182/testReport/
Findbugs warnings: 
http://hudson.zones.apache.org/hudson/job/Zookeeper-Patch-vesta.apache.org/182/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Console output: 
http://hudson.zones.apache.org/hudson/job/Zookeeper-Patch-vesta.apache.org/182/console

This message is automatically generated.

 ZK fataled on me, and ugly
 --

 Key: ZOOKEEPER-483
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-483
 Project: Zookeeper
  Issue Type: Bug
Affects Versions: 3.1.1
Reporter: ryan rawson
Assignee: Benjamin Reed
 Fix For: 3.2.1, 3.3.0

 Attachments: zklogs.tar.gz, ZOOKEEPER-483.patch, ZOOKEEPER-483.patch, 
 ZOOKEEPER-483.patch


 here are the part of the log whereby my zookeeper instance crashed, taking 3 
 out of 5 down, and thus ruining the quorum for all clients:
 2009-07-23 12:29:06,769 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5161350 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:00,756 WARN org.apache.zookeeper.server.quorum.Follower: 
 Exception when following the leader
 java.io.EOFException
 at java.io.DataInputStream.readInt(DataInputStream.java:375)
 at 
 org.apache.jute.BinaryInputArchive.readInt(BinaryInputArchive.java:63)
 at 
 org.apache.zookeeper.server.quorum.QuorumPacket.deserialize(QuorumPacket.java:65)
 at 
 org.apache.jute.BinaryInputArchive.readRecord(BinaryInputArchive.java:108)
 at 
 org.apache.zookeeper.server.quorum.Follower.readPacket(Follower.java:114)
 at 
 org.apache.zookeeper.server.quorum.Follower.followLeader(Follower.java:243)
 at 
 org.apache.zookeeper.server.quorum.QuorumPeer.run(QuorumPeer.java:494)
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.168:39489]
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0578 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46797]
 2009-07-23 12:29:06,771 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa013e NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:33998]
 2009-07-23 12:29:06,771 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5160593 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:06,808 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e02bb NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.158:53758]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13e4 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.154:58681]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691382 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59967]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb1354 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.163:49957]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13cd NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.150:34212]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 

[jira] Commented: (ZOOKEEPER-499) electionAlg should default to FLE (3) - regression

2009-08-10 Thread Mahadev konar (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741592#action_12741592
 ] 

Mahadev konar commented on ZOOKEEPER-499:
-

+1 this looks good. .. 

 electionAlg should default to FLE (3) - regression
 --

 Key: ZOOKEEPER-499
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-499
 Project: Zookeeper
  Issue Type: Bug
  Components: server, tests
Affects Versions: 3.2.0
Reporter: Patrick Hunt
Assignee: Patrick Hunt
Priority: Blocker
 Fix For: 3.2.1, 3.3.0

 Attachments: ZOOKEEPER-499.patch, ZOOKEEPER-499.patch, 
 ZOOKEEPER-499_br3.2.patch, ZOOKEEPER-499_br3.2.patch


 there's a regression in 3.2 - electionAlg is no longer defaulting to 3 
 (incorrectly defaults to 0)
 also - need to have tests to validate this

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-499) electionAlg should default to FLE (3) - regression

2009-08-10 Thread Mahadev konar (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahadev konar updated ZOOKEEPER-499:


  Resolution: Fixed
Hadoop Flags: [Reviewed]
  Status: Resolved  (was: Patch Available)

I just committed this. thanks pat.

 electionAlg should default to FLE (3) - regression
 --

 Key: ZOOKEEPER-499
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-499
 Project: Zookeeper
  Issue Type: Bug
  Components: server, tests
Affects Versions: 3.2.0
Reporter: Patrick Hunt
Assignee: Patrick Hunt
Priority: Blocker
 Fix For: 3.2.1, 3.3.0

 Attachments: ZOOKEEPER-499.patch, ZOOKEEPER-499.patch, 
 ZOOKEEPER-499_br3.2.patch, ZOOKEEPER-499_br3.2.patch


 there's a regression in 3.2 - electionAlg is no longer defaulting to 3 
 (incorrectly defaults to 0)
 also - need to have tests to validate this

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (ZOOKEEPER-483) ZK fataled on me, and ugly

2009-08-10 Thread Mahadev konar (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741599#action_12741599
 ] 

Mahadev konar commented on ZOOKEEPER-483:
-

the test added with the patch failed on hudson.. ben can you take a look?

 ZK fataled on me, and ugly
 --

 Key: ZOOKEEPER-483
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-483
 Project: Zookeeper
  Issue Type: Bug
Affects Versions: 3.1.1
Reporter: ryan rawson
Assignee: Benjamin Reed
 Fix For: 3.2.1, 3.3.0

 Attachments: zklogs.tar.gz, ZOOKEEPER-483.patch, ZOOKEEPER-483.patch, 
 ZOOKEEPER-483.patch


 here are the part of the log whereby my zookeeper instance crashed, taking 3 
 out of 5 down, and thus ruining the quorum for all clients:
 2009-07-23 12:29:06,769 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5161350 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:00,756 WARN org.apache.zookeeper.server.quorum.Follower: 
 Exception when following the leader
 java.io.EOFException
 at java.io.DataInputStream.readInt(DataInputStream.java:375)
 at 
 org.apache.jute.BinaryInputArchive.readInt(BinaryInputArchive.java:63)
 at 
 org.apache.zookeeper.server.quorum.QuorumPacket.deserialize(QuorumPacket.java:65)
 at 
 org.apache.jute.BinaryInputArchive.readRecord(BinaryInputArchive.java:108)
 at 
 org.apache.zookeeper.server.quorum.Follower.readPacket(Follower.java:114)
 at 
 org.apache.zookeeper.server.quorum.Follower.followLeader(Follower.java:243)
 at 
 org.apache.zookeeper.server.quorum.QuorumPeer.run(QuorumPeer.java:494)
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.168:39489]
 2009-07-23 12:29:06,770 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0578 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46797]
 2009-07-23 12:29:06,771 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa013e NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:33998]
 2009-07-23 12:29:06,771 WARN org.apache.zookeeper.server.NIOServerCnxn: 
 Exception causing close of session 0x52276d1d5160593 due to 
 java.io.IOException: Read error
 2009-07-23 12:29:06,808 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e02bb NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.158:53758]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13e4 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.154:58681]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691382 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59967]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb1354 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.163:49957]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x42276d1d3fa13cd NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.150:34212]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x22276d15e691383 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.159:46813]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x12276d15dfb0350 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.162:59956]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e139b NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.156:55138]
 2009-07-23 12:29:06,809 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x32276d15d2e1398 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.167:41257]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing session:0x52276d1d5161355 NIOServerCnxn: 
 java.nio.channels.SocketChannel[connected local=/10.20.20.151:2181 
 remote=/10.20.20.153:34032]
 2009-07-23 12:29:06,810 INFO org.apache.zookeeper.server.NIOServerCnxn: 
 closing 

[jira] Commented: (ZOOKEEPER-498) Unending Leader Elections : WAN configuration

2009-08-10 Thread Benjamin Reed (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741605#action_12741605
 ] 

Benjamin Reed commented on ZOOKEEPER-498:
-

+1 looks good. when setting the stop flags, you should really do an interrupt 
to wake up the wait, but that will cause a message to be printed to stdout. 
i'll open another jira to fix that.

 Unending Leader Elections : WAN configuration
 -

 Key: ZOOKEEPER-498
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-498
 Project: Zookeeper
  Issue Type: Bug
  Components: leaderElection
Affects Versions: 3.2.0
 Environment: Each machine:
 CentOS 5.2 64-bit
 2GB ram
 java version 1.6.0_13
 Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed 
 Network Topology:
 DC : central data center
 POD(N): remote data center
 Zookeeper Topology:
 Leaders may be elected only in DC (weight = 1)
 Only followers are elected in PODS (weight = 0)
Reporter: Todd Greenwood-Geer
Assignee: Flavio Paiva Junqueira
Priority: Critical
 Fix For: 3.2.1, 3.3.0

 Attachments: dc-zook-logs-01.tar.gz, pod-zook-logs-01.tar.gz, 
 zk498-test.tar.gz, zoo.cfg, ZOOKEEPER-498.patch, ZOOKEEPER-498.patch, 
 ZOOKEEPER-498.patch, ZOOKEEPER-498.patch


 In a WAN configuration, ZooKeeper is endlessly electing, terminating, and 
 re-electing a ZooKeeper leader. The WAN configuration involves two groups, a 
 central DC group of ZK servers that have a voting weight = 1, and a group of 
 servers in remote pods with a voting weight of 0.
 What we expect to see is leaders elected only in the DC, and the pods to 
 contain only followers. What we are seeing is a continuous cycling of 
 leaders. We have seen this consistently with 3.2.0, 3.2.0 + recommended 
 patches (473, 479, 481, 491), and now release 3.2.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-498) Unending Leader Elections : WAN configuration

2009-08-10 Thread Benjamin Reed (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benjamin Reed updated ZOOKEEPER-498:


Hadoop Flags: [Reviewed]

 Unending Leader Elections : WAN configuration
 -

 Key: ZOOKEEPER-498
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-498
 Project: Zookeeper
  Issue Type: Bug
  Components: leaderElection
Affects Versions: 3.2.0
 Environment: Each machine:
 CentOS 5.2 64-bit
 2GB ram
 java version 1.6.0_13
 Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 11.3-b02, mixed 
 Network Topology:
 DC : central data center
 POD(N): remote data center
 Zookeeper Topology:
 Leaders may be elected only in DC (weight = 1)
 Only followers are elected in PODS (weight = 0)
Reporter: Todd Greenwood-Geer
Assignee: Flavio Paiva Junqueira
Priority: Critical
 Fix For: 3.2.1, 3.3.0

 Attachments: dc-zook-logs-01.tar.gz, pod-zook-logs-01.tar.gz, 
 zk498-test.tar.gz, zoo.cfg, ZOOKEEPER-498.patch, ZOOKEEPER-498.patch, 
 ZOOKEEPER-498.patch, ZOOKEEPER-498.patch


 In a WAN configuration, ZooKeeper is endlessly electing, terminating, and 
 re-electing a ZooKeeper leader. The WAN configuration involves two groups, a 
 central DC group of ZK servers that have a voting weight = 1, and a group of 
 servers in remote pods with a voting weight of 0.
 What we expect to see is leaders elected only in the DC, and the pods to 
 contain only followers. What we are seeing is a continuous cycling of 
 leaders. We have seen this consistently with 3.2.0, 3.2.0 + recommended 
 patches (473, 479, 481, 491), and now release 3.2.1.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-477) zkCleanup.sh is flaky

2009-08-10 Thread Mahadev konar (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahadev konar updated ZOOKEEPER-477:


Attachment: ZOOKEEPER-477.patch

patch against trunk the patch is the same but just generated against trunks 
top level directory.

 zkCleanup.sh is flaky
 -

 Key: ZOOKEEPER-477
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-477
 Project: Zookeeper
  Issue Type: Bug
  Components: scripts
Affects Versions: 3.2.0
Reporter: Fernando
Assignee: Fernando
 Fix For: 3.2.1, 3.3.0

 Attachments: ppp, ZOOKEEPER-477.patch


 the zkCleanup.sh script is buggy in two ways:
 1) it doesn't actually pass through the snapshot count, so it doesn't work
 2) it assumes that there is only dataDir, it doesn't support dataLogDir
 And it can use cleanup, so that it doesn't blindly call eval from the config 
 file..

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-477) zkCleanup.sh is flaky

2009-08-10 Thread Mahadev konar (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahadev konar updated ZOOKEEPER-477:


  Resolution: Fixed
Hadoop Flags: [Reviewed]
  Status: Resolved  (was: Patch Available)

I just committed this. thanks fernando!

 zkCleanup.sh is flaky
 -

 Key: ZOOKEEPER-477
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-477
 Project: Zookeeper
  Issue Type: Bug
  Components: scripts
Affects Versions: 3.2.0
Reporter: Fernando
Assignee: Fernando
 Fix For: 3.2.1, 3.3.0

 Attachments: ppp, ZOOKEEPER-477.patch


 the zkCleanup.sh script is buggy in two ways:
 1) it doesn't actually pass through the snapshot count, so it doesn't work
 2) it assumes that there is only dataDir, it doesn't support dataLogDir
 And it can use cleanup, so that it doesn't blindly call eval from the config 
 file..

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (ZOOKEEPER-496) zookeeper-tree utility for export, import and incremental updates

2009-08-10 Thread Anirban Roy (JIRA)

 [ 
https://issues.apache.org/jira/browse/ZOOKEEPER-496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anirban Roy updated ZOOKEEPER-496:
--

Release Note: 
==
zktreeutil - Zookeeper Tree Data Utility
Author: Anirban Roy
Organization: Yahoo Inc.
==

zktreeutil program is intended to manage and manipulate zk-tree data quickly, 
effi-
ciently and with ease.The utility operates on free-form ZK-tree and hence can 
be used
for any cluster managed by Zookeeper. Here are the basic functionalities -

EXPORT: The whole ZK-tree is exported into a XML file. This helps in capturing 
a cur-
rent snapshot of the data for backup/analysis. In addition, the data can be 
dumped
as tree-like text format which is easily readable.

IMPORT: The ZK-tree can be imported from XML into ZK cluster. This helps in 
priming
the new ZK cluster with static configuration. The import can be non-intrusive by
making only the additions in the existing data.

DIFF: Creates a diff between live ZK data vs data saved in XML file. Diff can 
ignore
some ZK-tree branches (possibly dynamic data) on reading the flag from XML file.

UPDATE: Make the incremental changes into the live ZK-tree from saved XML, 
essentia-
lly after running the diff.

DUMP: Dumps the ZK-tree on the standard output device reading either from live 
ZK
server or XML file.

The exported ZK data into XML file can be shortened by only keeping the static 
ZK
nodes which are required to prime a cluster. The dynamic zk nodes (created 
on-the-
fly) can be ignored by setting a 'ignore' attribute at the root node of the 
dynamic
subtree (see tests/zk_sample.xml), possibly deleting all inner ZK nodes under 
that.
Once ignored, the whole subtree is ignored during DIFF, UPDATE and WRITE.

Pre-requisites
--
1. Linux system with 2.6.X kernel.
2. Zookeeper C client library = 3.X.X
3. Development build libraries:
  a. libxml2 = 2.7.3
  b. log4cxx = 0.10.0

Build instructions
--
1. cd into this directory
2. autoreconf -if
3. ./configure
4. make
5. 'zktreeutil' binary created under src directory

Testing zktreeutil
--
1.  Run Zookeeper server locally on port 2181
2.  export LD_LIBRARY_PATH=../../c/.libs/:/usr/local/lib/
3.  ./src/zktreeutil --help # show help
4.  ./src/zktreeutil --zookeeper=localhost:2181 --import 
--xmlfile=tests/zk_sample.xml 2/dev/null   # import sample ZK tree
5.  ./src/zktreeutil --zookeeper=localhost:2181 --dump 2/dev/null  
 # dump on the terminal
6.  ./src/zktreeutil --xmlfile=zk_sample.xml -D 2/dev/null 
 # dump the xml data
7.  Change zk_sample.xml with adding/deleting/chaging some nodes
8.  ./src/zktreeutil --zookeeper=localhost:2181 --diff --xmlfile=zk_sample.xml 
2/dev/null   # take a diff of changes
9.  ./src/zktreeutil --zookeeper=localhost:2181 --export 2/dev/null  
zk_sample2.xml# export the mofied ZK tree
10. ./src/zktreeutil --zookeeper=localhost:2181 --update 
--xmlfile=zk_sample.xml 2/dev/null # update with incr. changes
11. ./src/zktreeutil --zookeeper=localhost:2181 --import --force 
--xmlfile=zk_sample2.xml 2/dev/null# re-prime the ZK tree


  was:
==
zktreeutil - Zookeeper Tree Data Utility
Author: Anirban Roy
Organization: Yahoo Inc.
==

zktreeutil program is intendened to manage and manipulate zk-tree data quickly, 
effi-
ciently and with ease.The utility operates on free-form ZK-tree and hence can 
be used
for any cluster managed by Zookeeper. Here are the basic functionalities -

EXPORT: The whole ZK-tree is exported into a XML file. This helps in capturing 
a cur-
rent snapshot of the data for backup/analysis. In addition, the data can be 
dumped
as tree-like text format which is easily readable.

IMPORT: The ZK-tree can be imported from XML into ZK cluster. This helps in 
priming
the new ZK cluster with static configuration. The import can be non-intrusive by
making only the additions in the existing data.

DIFF: Creates a diff between live ZK data vs data saved in XML file. Diff can 
ignore
some ZK-tree branches (possibly dynamic data) on reading the flag from XML file.

UPDATE: Make the incremental changes into the live ZK-tree from saved XML, 
essentia-
illy after running the diff.

DUMP: Dumps the ZK-tree on the standard output device reading either from live 
ZK
server or XML file.

The exported ZK data into XML file can be shortened by only keeping the static 
ZK
nodes which are required to prime a cluster. The dynamic zk nodes (created 
on-the-
fly) can be ignored by setting a 'ignore' attribute at the root node of the 
dynamic
subtree (see tests/zk_sample.xml), possibly deleting all inner ZK nodes under 
that.
Once ignored, the whole subtree is 

[jira] Commented: (ZOOKEEPER-472) Making DataNode not instantiate a HashMap when the node is ephmeral

2009-08-10 Thread Erik Holstad (JIRA)

[ 
https://issues.apache.org/jira/browse/ZOOKEEPER-472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12741689#action_12741689
 ] 

Erik Holstad commented on ZOOKEEPER-472:


After the HBase meeting this weekend we might have a use case for a new node 
type which keeps the children sorted, so including that we would have 3 
different types of nodes, if this new node would be interesting for ZooKeeper 
in general. Tying that to this issue I came up with the following:
1. Make a Node interface that only have the following methods
copyStat(...)
deserialize(...)
serialize(...)
2. For the ephemeral nodes this would be it, they wouldn't have to have any 
notion about children.
3, For the regular nodes and the sorted ones would also have:
setChildren(...)
addChild(...)
getChildren(...)
where getChildren(...) could have different return types depending on the node.

4. For the nodes that can have children we also keep the structure 
uninstantiated until children are actually set or added to the parent. This 
part I have already done, jsut need to do some more testing on it.

Erik


 Making DataNode not instantiate a HashMap when the node is ephmeral
 ---

 Key: ZOOKEEPER-472
 URL: https://issues.apache.org/jira/browse/ZOOKEEPER-472
 Project: Zookeeper
  Issue Type: Improvement
  Components: server
Affects Versions: 3.1.1, 3.2.0
Reporter: Erik Holstad
Assignee: Erik Holstad
Priority: Minor
 Fix For: 3.3.0


 Looking at the code, there is an overhead of a HashSet object for that nodes 
 children, even though the node might be an ephmeral node and cannot have 
 children.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.