----- Original Message ----- From: "Andrew Gaul" <[email protected]> To: "Yehuda Sadeh-Weinraub" <[email protected]> Cc: [email protected], "Alfredo Deza" <[email protected]> Sent: Friday, February 27, 2015 1:26:23 PM Subject: Re: s3-tests development
On Fri, Feb 27, 2015 at 12:11:45PM -0500, Yehuda Sadeh-Weinraub wrote: > Andrew Gaul wrote: > > While s3-tests has good functionality as it > > exists, the project has not progressed much over the last six months. I > > have submitted over 20 pull requests to fix incorrect tests and for > > additional test coverage but most remain unreviewed[2]. > > Right. We do need to be more responsive. The main reason we took our time was > that these tests are used for the ceph nightly qa tests, and any changes in > these that exposes new incompatibility will fail these. We'd rather first > have the issue fixed, then merge the change. An alternative way to doing it > is to open a ceph tracker issue about the incompatibility, mark the test as > 'fails_on_rgw', in which case we could merge it immediately. Instead of testing against master, perhaps you can tag a s3-tests 1.0.0 release and Ceph can use that? Alternatively you can run your tests against a specific commit hash. In addition to running s3-tests against Ceph, it would be good to run it against AWS, although we are blocked as discussed below. > > s3-tests remains biased towards Ceph and not the AWS de facto standard, > > failing to run against the latter due to TooManyBuckets failures[3]. > > Not sure what's the appropriate way to attack this one. Maybe the > create_bucket function could keep a list of created buckets and then remove > them if encountering this error using lru. jclouds uses an LRU scheme to recycle buckets between tests which works well. Alternatively s3-tests could remove buckets immediately after each test although this has some tooling challenges discussed in the issue. > > Finally some features like V4 signature support[4] will require more > > extensive changes. We are at-risk of diverging; how can we best move > > forward together? > > We'd be happy with more tests, and with more features tested even if these > weren't implemented yet. It can later help with the development of these > features. E.g., I looked a few weeks back at v4 signatures, and having such a > test would have helped. The only thing we need though would be a way to > easily disable such tests. So adding 'fails_on_rgw', or some other way to > detect would help a lot. Annotating every failing test will not scale with additional providers; I have almost 100 annotations for S3Proxy presently[1]. Could we implement an external profile mechanism which tracks which tests should or should not run? Perhaps nosetests has a way to do this already? Like I mentioned, we are working to port these to py.test with a re-organization included that should allow to configure what tests are run and how failures are treated. [1] https://github.com/andrewgaul/s3-tests/commit/7891cb4d9a1cd6e6f4d8119f33c4f0bdb35348d3 -- Andrew Gaul http://gaul.org/ -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to [email protected] More majordomo info at http://vger.kernel.org/majordomo-info.html
