Hi Peter,
A filter callback based on blockdata records would more frequently
inform the filter
of progress in the stream and the amount of data being consumed is
already available
from streamBytes. That could improve detection of excessive data
without complicating
the interface and it would not be all that useful to pass the size of
individual block data records.
And it only helps in the case of a class with exclusively primitive data
and poor
checks on the data stream. As far as I can see, it increases the number
of filter callbacks
quite a bit and adds little value.
The current thresholds depth, number of refs, array size, and stream
size can identify
out of bounds conditions. I don't want argue that it can't happen but
we're definitely
down the slippery slope beyond what is needed if it tries to expose the
shape of the graph or
what bytes go where.
Roger
On 7/22/2016 3:00 AM, Peter Levart wrote:
Hi Roger,
On 07/21/2016 08:19 PM, Roger Riggs wrote:
- The call-back is invoked after the type of the object and
possible array length is read from stream but before the object's
state is read. Suppose that the object that is about to be read is
either Externalizable object or an object with a readObject()
method(s) that consume block data from the stream. This block data
can be large. Should there be a call-back to "announce" the block
data too? (for example, when the 'clazz' is null and the 'size' is
0, the call-back reports a back-reference to a previously read
object, but when the 'clazz' is null and the 'size' > 0, it
announces the 'size' bytes of block data. Does this make sense?)
Interesting case, I'll take another look at that. Since block data
records are <= 1024, a filter might not
have enough information to make an informed decision. Those bytes
would show up in
the stream bytes but not until the next object is read.
...which could be to late. If the filter is to be also used as a
defense against forged streams that try to provoke DOS by triggering
frequent GCs and OutOfMemoryError(s), then such call-back that
announces each block data record could help achieve that.
Individual block data lengths are not very informative since block
data can be segmented but
a cumulative (for the whole stream) block data size suitable for a
callback from the
start of each block data segment might be useful.
Is it possible to identify which block data records "belong" to a
particular object? If yes, then perhaps cumulative sum(s) of block
data sizes for a particular object could be passed to the call-back
together with the Class of the object the data belongs to (similar to
how array is reported, the size would be cumulative block data size
read so-far and the filter could distinguish such callback from array
callback by inspecting clazz.isArray()). In conjunction with
cumulative size of the whole stream which is already passed now, I
think this is enough to implement all kinds of safety-belts.
If you think that such addition starts to become complex from
combinatorial standpoint then what about passing an additional enum
argument identifying a particular "event"? It would then be easy do
document the rest of parameters for each event type...
Regards, Peter