1. THE STORE IS *LESS* EFFECTIVE THAN THE CACHE!
------------------------------------------------

Please could people post their store statistics? Cache hits, store hits,
cached keys, stored keys.

So far:
[23:11] <nextgens> # Cached keys: 6,389 (199 MiB)
[23:11] <nextgens> # Stored keys: 24,550 (767 MiB)
[23:09] <nextgens> # Cache hits: 217 / 12,738 (1%)
[23:09] <nextgens> # Store hits: 14 / 10,818 (0%)

(Cached hits / cached keys) / (Stored hits / stored keys) = 59.56

[23:12] <cyberdo> # Cached keys: 17,930 (560 MiB)
[23:12] <cyberdo> # Stored keys: 24,895 (777 MiB)
[23:14] <cyberdo> # Cache hits: 178 / 3,767 (4%)
[23:14] <cyberdo> # Store hits: 11 / 2,970 (0%)

(Cached hits / cached keys) / (Stored hits / stored keys) = 22.47

[23:14] <sandos> # Cached keys: 45,148 (1.37 GiB)
[23:14] <sandos> # Stored keys: 16,238 (507 MiB)
[23:11] <sandos> # Cache hits: 41 / 861 (4%)
[23:11] <sandos> # Store hits: 5 / 677 (0%)

(Cached hits / cached keys) / (Stored hits / stored keys) = 2.95

Thus, in practice, the cache is far more efficient than the store.

The cache caches every key fetched or inserted through this node.

The store stores only keys inserted, and of those, only those for which
there is no closer node to the key amongst our peers.


The cache being more effective than the store (and note that the above
is for CHKs only) implies either:
1. Routing is broken.
2. There is more location churn than the store can cope with.
3. There is more data churn than the store can cope with.


2. SUSPICIONS OF EXCESSIVE LOCATION CHURN
-----------------------------------------

ljn1981 said that his node would often do a swap and then reverse it.
However several people say their location is more or less what it was.
It is necessary to make a log of a node's location changes over time...


3. PROBE REQUESTS NOT WORKING
-----------------------------

"Probe requests" are a new class of requests which simply take a
location, and try to find the next location - the lowest location
greater than the one they started with. Here's a recent trace (these can
be triggered by telneting to 2323 and typing PROBEALL:, then watching
wrapper.log):

LOCATION 1: 0.00917056526893234
LOCATION 2: 0.009450590423585203
LOCATION 3: 0.009507800765948482
LOCATION 4: 0.03378227720218496
[ delays ]
LOCATION 5: 0.033884263580090224
[ delays ]
LOCATION 6: 0.03557139211207139
LOCATION 7: 0.04136594238104219
LOCATION 8: 0.06804731119243879
LOCATION 9: 0.06938071503433951
LOCATION 10: 0.11468659860500963
[ big delays ]
LOCATION 11: 0.11498938134581993
LOCATION 12: 0.11800179518614218
LOCATION 13: 0.1180104005154885
LOCATION 14: 0.11907112718505641
LOCATION 15: 0.3332896508938398
[ biggish delays ]
LOCATION 16: 0.6963082287578662
LOCATION 17: 0.7003642648424434
LOCATION 18: 0.7516363167204175
LOCATION 19: 0.7840227104081505
LOCATION 20: 0.8238921670991454
LOCATION 21: 0.8551853934902863
LOCATION 22: 0.8636946791670825
LOCATION 23: 0.8755575572906827
LOCATION 24: 0.883042607673485
LOCATION 25: 0.8910451777595195
LOCATION 26: 0.8930966991557874
LOCATION 27: 0.8939968594038799
LOCATION 28: 0.8940798222254085
LOCATION 29: 0.8941104802690825
LOCATION 30: 0.9103443172876444
LOCATION 31: 0.9103717579924239
LOCATION 32: 0.9107237145701387
LOCATION 33: 0.9108357699627044
LOCATION 34: 0.9130496893125409
LOCATION 35: 0.9153056056305631
[ delays ]
LOCATION 36: 0.9180229911856111
LOCATION 37: 0.9184676396364483
LOCATION 38: 0.9198162081803294
LOCATION 39: 0.9232383399833453
[ big delays ]
LOCATION 40: 0.9232484869765467
LOCATION 41: 0.9398827726484242
LOCATION 42: 0.9420672052844097
LOCATION 43: 0.9442367949642505
LOCATION 44: 0.9521296958111133
[ big delays ]
LOCATION 45: 0.9521866483104723
LOCATION 46: 0.9562645053030697
LOCATION 47: 0.9715290823566148
LOCATION 48: 0.9722492845296398
LOCATION 49: 0.974283274258849
[ big delays ... ]

Clearly there are more than around 50 nodes on freenet at any given
time, and the above includes some really big jumps, as well as some
really small ones. This may be a problem with probe requests, but it
is suspicious...
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 189 bytes
Desc: Digital signature
URL: 
<https://emu.freenetproject.org/pipermail/devl/attachments/20061007/45437771/attachment.pgp>

Reply via email to