On Fri, May 15, 2015 at 02:28:37PM -0400, Konstantin Ryabitsev wrote:
> On 15/05/15 02:22 PM, Junio C Hamano wrote:
> > Also, is it worth allocating small and then growing up to the maximum?
> > I think this only relays one request at a time anyway, and I suspect
> > that a single 1MB allocation at the first call kept getting reused
> > may be sufficient (and much simpler).
>
> Does it make sense to make that configurable via an env variable at all?
> I suspect the vast majority of people would not hit this bug unless they
> are dealing with repositories polluted with hundreds of refs created by
> automation (like the codeaurora chromium repo).
>
> E.g. can be set via an Apache directive like
>
> SetEnv FOO_MAX_SIZE 2048
>
> The backend can then be configured to emit an error message when the
> spool size is exhausted saying "foo exhausted, increase FOO_MAX_SIZE to
> allow for moar foo."
Yeah, that was the same conclusion I came to elsewhere in the thread.
Here's a re-roll:
[1/3]: http-backend: fix die recursion with custom handler
[2/3]: t5551: factor out tag creation
[3/3]: http-backend: spool ref negotiation requests to buffer
It makes the size configurable (either through the environment, which is
convenient for setting via Apache; or through the config, which is
convenient if you have one absurdly-sized repo). It mentions the
variable name when it barfs into the Apache log. I also bumped the
default to 10MB, which I think should be enough to handle even
ridiculous cases.
I also adapted Dennis's test into the third patch. Beware that it's
quite slow to run (and is protected by the "EXPENSIVE" prerequisite).
Patch 2 is new, and just refactors the script to make adding the new
test easier.
I really wanted to add a test like:
diff --git a/t/t5551-http-fetch-smart.sh b/t/t5551-http-fetch-smart.sh
index e2a2fa1..1fff812 100755
--- a/t/t5551-http-fetch-smart.sh
+++ b/t/t5551-http-fetch-smart.sh
@@ -273,6 +273,16 @@ test_expect_success 'large fetch-pack requests can be
split across POSTs' '
test_line_count = 2 posts
'
+test_expect_success 'http-backend does not buffer arbitrarily large requests' '
+ test_when_finished "(
+ cd \"$HTTPD_DOCUMENT_ROOT_PATH/repo.git\" &&
+ test_unconfig http.maxrequestbuffer
+ )" &&
+ git -C "$HTTPD_DOCUMENT_ROOT_PATH/repo.git" \
+ config http.maxrequestbuffer 100 &&
+ test_must_fail git clone $HTTPD_URL/smart/repo.git foo.git
+'
+
test_expect_success EXPENSIVE 'http can handle enormous ref negotiation' '
(
cd "$HTTPD_DOCUMENT_ROOT_PATH/repo.git" &&
to test that the maxRequestBuffer code does indeed work. Unfortunately,
even though the server behaves reasonably in this case, the client ends
up hanging forever. I'm not sure there is a simple solution to that; I
think it is a protocol issue where remote-http is waiting for fetch-pack
to speak, but fetch-pack is waiting for more data from the remote.
Personally, I think I'd be much more interested in pursuing a saner,
full duplex http solution like git-over-websockets than trying to iron
out all of the corner cases in the stateless-rpc protocol.
-Peff
--
To unsubscribe from this list: send the line "unsubscribe git" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html