Tracking this here: https://issues.apache.org/jira/browse/FLINK-16122

On Sat, Feb 15, 2020 at 8:12 PM Robert Metzger <rmetz...@apache.org> wrote:

> I agree that we need to fix this.
>
> We could either misuse the "build artifact" feature of azure pipelines to
> publish the logs, or we set up something simple for Flink (like running an
> instance of https://github.com/lachs0r/0x0 or
> https://github.com/dutchcoders/transfer.sh :) )
>
> On Fri, Feb 14, 2020 at 8:44 PM Chesnay Schepler <ches...@apache.org>
> wrote:
>
>> The S3 setup only works in the apache repo though; not on contributor
>> branches or PR builds.
>>
>> We can tighten the timeouts (already talked to Robert about that), at
>> which point it doesn't hurt.
>>
>> On 14/02/2020 18:28, Stephan Ewen wrote:
>> > Hi all!
>> >
>> > I propose to remove the log upload via transfer.sh and rely on the S3
>> > upload instead.
>> >
>> > The reason is that transfer.sh seems to be very unreliable (times out in
>> > many profiles recently) and it seems that we also often don't get
>> access to
>> > uploaded logs (errors out on the transfer.sh website).
>> >
>> > Best,
>> > Stephan
>> >
>>
>>

Reply via email to