> On July 23, 2015, 6:01 p.m., Jie Yu wrote:
> > 3rdparty/libprocess/3rdparty/stout/include/stout/recordio.hpp, line 130
> > <https://reviews.apache.org/r/36677/diff/3/?file=1019201#file1019201line130>
> >
> >     Not needed?

Yeah, I've restructured the logic to be more clear (now we always transition to 
RECORD, but for 0 length we parse immediately and transition back to HEADER). 
Thanks jie!


- Ben


-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/36677/#review92627
-----------------------------------------------------------


On July 22, 2015, 11:24 p.m., Ben Mahler wrote:
> 
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/36677/
> -----------------------------------------------------------
> 
> (Updated July 22, 2015, 11:24 p.m.)
> 
> 
> Review request for mesos, Anand Mazumdar, Benjamin Hindman, and Jie Yu.
> 
> 
> Bugs: MESOS-3067
>     https://issues.apache.org/jira/browse/MESOS-3067
> 
> 
> Repository: mesos
> 
> 
> Description
> -------
> 
> Note that most "Record-IO" encodings are used for file I/O
> and consequently use a fixed-size header to encode the record
> length. However, decoding a base-10 integer is more
> straightforward to implement in most languages, and so this
> was chosen instead. (Note that the Twitter streaming API
> uses the same technique for portability).
> 
> 
> Diffs
> -----
> 
>   3rdparty/libprocess/3rdparty/stout/include/Makefile.am 
> 2394b95462182273464f0847f416ad83c3b64485 
>   3rdparty/libprocess/3rdparty/stout/include/stout/protobuf.hpp 
> 8c75f6b28c18596018eaefe427b238424aae2fd9 
>   3rdparty/libprocess/3rdparty/stout/include/stout/recordio.hpp PRE-CREATION 
>   3rdparty/libprocess/3rdparty/stout/tests/recordio_tests.cpp PRE-CREATION 
> 
> Diff: https://reviews.apache.org/r/36677/diff/
> 
> 
> Testing
> -------
> 
> Added tests.
> 
> 
> Thanks,
> 
> Ben Mahler
> 
>

Reply via email to