if you're not specifying initialtoken, every time you wipe your installation it will generate new tokens. for a small number of machines you'll definitely see some random tokens better balanced than others.
On Tue, Nov 10, 2009 at 9:38 AM, Richard grossman <[email protected]> wrote: > Hi > > I've build the 0.4.2 from the tags in SVN. > I've made exactly the same cluster with same configuration as 0.4.1 > I've delete all the data in all server > > Now I send the data to first server and the data is not more distributed > across the other server as previously. > I've configured replica to 1 > > here is my storage-conf.xml > > <Storage> > > <ClusterName>BeeCluster</ClusterName> > > > <Keyspaces> > <Keyspace Name="Keyspace1"> > > <ColumnFamily CompareWith="BytesType" > Name="channelShowLink" > FlushPeriodInMinutes="15"/> > > <ColumnFamily CompareWith="BytesType" > Name="channelShow" > FlushPeriodInMinutes="15"/> > > <ColumnFamily CompareWith="BytesType" > Name="userAction" > FlushPeriodInMinutes="15"/> > > <ColumnFamily CompareWith="BytesType" > Name="headends" > FlushPeriodInMinutes="15"/> > > <ColumnFamily CompareWith="BytesType" > Name="similarity" > FlushPeriodInMinutes="500"/> > </Keyspace> > </Keyspaces> > > > > <Partitioner>org.apache.cassandra.dht.OrderPreservingPartitioner</Partitioner> > > <InitialToken></InitialToken> > > > > <EndPointSnitch>org.apache.cassandra.locator.EndPointSnitch</EndPointSnitch> > > > > <ReplicaPlacementStrategy>org.apache.cassandra.locator.RackUnawareStrategy</ReplicaPlacementStrategy> > > <ReplicationFactor>1</ReplicationFactor> > > > > <CommitLogDirectory>/home/beecloud/cassandrapart/commitlog</CommitLogDirectory> > <DataFileDirectories> > > <DataFileDirectory>/home/beecloud/cassandrapart/data</DataFileDirectory> > </DataFileDirectories> > <CalloutLocation>/home/beecloud/cassandrapart/callouts</CalloutLocation> > > <BootstrapFileDirectory>/home/beecloud/cassandrapart/bootstrap</BootstrapFileDirectory> > > <StagingFileDirectory>/home/beecloud/cassandrapart/staging</StagingFileDirectory> > > > > <Seeds> > <Seed>192.168.249.200</Seed> > <Seed>192.168.249.222</Seed> > <Seed>192.168.249.95</Seed> > </Seeds> > > > <!-- Miscellaneous --> > > <!-- Time to wait for a reply from other nodes before failing the command > --> > <RpcTimeoutInMillis>50000</RpcTimeoutInMillis> > <!-- Size to allow commitlog to grow to before creating a new segment --> > <CommitLogRotationThresholdInMB>128</CommitLogRotationThresholdInMB> > > > > <ListenAddress>192.168.249.200</ListenAddress> > <!-- TCP port, for commands and data --> > <StoragePort>7000</StoragePort> > <!-- UDP port, for membership communications (gossip) --> > <ControlPort>7001</ControlPort> > > > <ThriftAddress>0.0.0.0</ThriftAddress> > <!-- Thrift RPC port (the port clients connect to). --> > <ThriftPort>9160</ThriftPort> > <!-- > ~ Whether or not to use a framed transport for Thrift. If this option > ~ is set to true then you must also use a framed transport on the > ~ client-side, (framed and non-framed transports are not compatible). > --> > <ThriftFramedTransport>false</ThriftFramedTransport> > > > > <SlicedBufferSizeInKB>64</SlicedBufferSizeInKB> > > > <FlushDataBufferSizeInMB>32</FlushDataBufferSizeInMB> > <FlushIndexBufferSizeInMB>8</FlushIndexBufferSizeInMB> > > > <ColumnIndexSizeInKB>64</ColumnIndexSizeInKB> > > > <MemtableSizeInMB>64</MemtableSizeInMB> > > <MemtableObjectCountInMillions>0.1</MemtableObjectCountInMillions> > > > <ConcurrentReads>8</ConcurrentReads> > <ConcurrentWrites>32</ConcurrentWrites> > > > <CommitLogSync>periodic</CommitLogSync> > > <CommitLogSyncPeriodInMS>1000</CommitLogSyncPeriodInMS> > > <!-- <CommitLogSyncBatchWindowInMS>1</CommitLogSyncBatchWindowInMS> --> > > > <GCGraceSeconds>864000</GCGraceSeconds> > > > <FlushMinThreads>1</FlushMinThreads> > <FlushMaxThreads>1</FlushMaxThreads> > > > <BinaryMemtableSizeInMB>256</BinaryMemtableSizeInMB> > > </Storage> > > Is there anyone with the same problem ?? > > Thanks >
