Thanks for the update!

Just to be clear, you're proposing going forward with the "simple fix" of 
reverting FLINK-7294?

> On 1. Dec 2017, at 18:39, Eron Wright <eronwri...@gmail.com> wrote:
> 
> Update on reported Mesos issue (FLINK-8174):
> 
> TLDR; a PR will be ready within 24 hours that will undo reservation support.
> 
> A couple of months ago, a fix (FLINK-7294) was merged related to how Flink
> accepts Mesos resource offers.  The intention was to allow Flink to make
> use of so-called +reserved+ resources, a Mesos feature which makes it
> possible to reserve hosts for use by a specific framework/role.  The fix
> inadvertently regressed the ability to use +unreserved+ resources.  This is
> a serious regression because unreserved resources are the common case.
> 
> The simple solution is to revert the earlier fix, deferring support for
> reservations to another release.   We are spending some time to find a fix
> that works for all scenarios, but seems unlikely at this time.   I am
> reaching out to the original contributor to get their feedback.
> 
> In the course of the investigation, a related flaw was discovered in Fenzo
> that causes Flink to misinterpret offers that contain a mix of reserved and
> unreserved resources.   I believe that a small fix is possible purely
> within Flink; an update to Fenzo does not appear necessary.
> 
> Going forward, we will contribute an improved integration test suite with
> which to test Flink under diverse Mesos conditions (e.g. reservations).
> 
> Thanks,
> Eron
> 
> On Thu, Nov 30, 2017 at 9:47 PM, Tzu-Li (Gordon) Tai <tzuli...@apache.org>
> wrote:
> 
>> Hi,
>> 
>> I’ve noticed a behavioral regression in the Kafka producer, that should
>> also be considered a blocker: https://issues.apache.org/
>> jira/browse/FLINK-8181
>> There’s already a PR for the issue here: https://github.com/
>> apache/flink/pull/5108
>> 
>> Best,
>> Gordon
>> 
>> On 30 November 2017 at 5:27:22 PM, Fabian Hueske (fhue...@gmail.com)
>> wrote:
>> 
>> I've created a JIRA issue for the the Hadoop 2.9.0 build problem [1].
>> 
>> Best, Fabian
>> 
>> [1] https://issues.apache.org/jira/browse/FLINK-8177
>> 
>> 2017-11-30 4:35 GMT+01:00 Eron Wright <eronwri...@gmail.com>:
>> 
>>> Unfortunately we've identified a blocker bug for Flink on Mesos -
>>> FLINK-8174. We'll have a patch ready on Thursday.
>>> 
>>> Thanks,
>>> Eron
>>> 
>>> On Wed, Nov 29, 2017 at 3:40 PM, Eron Wright <eronwri...@gmail.com>
>> wrote:
>>> 
>>>> On Dell EMC side, we're testing the RC2 on DCOS 1.10.0. Seeing a
>>>> potential issue with offer acceptance and we'll update the thread with
>> a
>>> +1
>>>> or with a more concrete issue within 24 hours.
>>>> 
>>>> Thanks,
>>>> Eron
>>>> 
>>>> On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <ches...@apache.org>
>>>> wrote:
>>>> 
>>>>> I don't think anyone has taken a look yet, nor was there a discussion
>> as
>>>>> to postponing it.
>>>>> 
>>>>> It just slipped through the cracks i guess...
>>>>> 
>>>>> 
>>>>> On 29.11.2017 15:47, Gyula Fóra wrote:
>>>>> 
>>>>>> Hi guys,
>>>>>> I ran into this again while playing with savepoint/restore
>> parallelism:
>>>>>> 
>>>>>> https://issues.apache.org/jira/browse/FLINK-7595
>>>>>> https://github.com/apache/flink/pull/4651
>>>>>> 
>>>>>> Anyone has some idea about the status of this PR or were we planning
>> to
>>>>>> postpone this to 1.5?
>>>>>> 
>>>>>> Thanks,
>>>>>> Gyula
>>>>>> 
>>>>>> 
>>>>>> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov. 29.,
>>>>>> Sze,
>>>>>> 13:10):
>>>>>> 
>>>>>> OK, the situation is the following:
>>>>>>> 
>>>>>>> The test class (org.apache.flink.yarn.UtilsTest) implements a
>> Hadoop
>>>>>>> interface (Container) that was extended in Hadoop 2.9.0 by a getter
>>> and
>>>>>>> setter.
>>>>>>> By adding the methods, we can compile Flink for Hadoop 2.9.0.
>> However,
>>>>>>> the
>>>>>>> getter/setter add a dependency on a class that was also added in
>>> Hadoop
>>>>>>> 2.9.0.
>>>>>>> Therefore, the implementation is not backwards compatible with
>> Hadoop
>>>>>>> versions < 2.9.0.
>>>>>>> 
>>>>>>> Not sure how we can fix the problem. We would need two version of
>> the
>>>>>>> class
>>>>>>> that are chosen based on the Hadoop version. Do we have something
>> like
>>>>>>> that
>>>>>>> somewhere else?
>>>>>>> 
>>>>>>> Since this is only a problem in a test class, Flink 1.4.0 might
>> still
>>>>>>> work
>>>>>>> very well with Hadoop 2.9.0.
>>>>>>> However, this has not been tested AFAIK.
>>>>>>> 
>>>>>>> Cheers, Fabian
>>>>>>> 
>>>>>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>:
>>>>>>> 
>>>>>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0
>>> (released
>>>>>>>> a
>>>>>>>> few days ago) and got a compilation failure in a test class.
>>>>>>>> 
>>>>>>>> Right now, I'm assessing how much we need to fix to support Hadoop
>>>>>>>> 2.9.0.
>>>>>>>> I'll report later.
>>>>>>>> 
>>>>>>>> Best, Fabian
>>>>>>>> 
>>>>>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org>:
>>>>>>>> 
>>>>>>>> Agreed, this is a regression compared to the previous
>> functionality.
>>> I
>>>>>>>>> updated the issue to "Blocker".
>>>>>>>>> 
>>>>>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com>
>> wrote:
>>>>>>>>>> 
>>>>>>>>>> Hi all,
>>>>>>>>>> 
>>>>>>>>>> I have found the following issue:
>>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
>>>>>>>>>> 
>>>>>>>>>> I would say this is a blocker (I personally pass the
>> ParameterTool
>>>>>>>>>> all
>>>>>>>>>> 
>>>>>>>>> over
>>>>>>>>> 
>>>>>>>>>> the place in my production apps), but a pretty trivial issue to
>>> fix,
>>>>>>>>>> 
>>>>>>>>> we
>>>>>>> 
>>>>>>>> can
>>>>>>>>> 
>>>>>>>>>> wait a little to find other potential problems.
>>>>>>>>>> 
>>>>>>>>>> I can submit a fix in a little bit.
>>>>>>>>>> 
>>>>>>>>>> Cheers,
>>>>>>>>>> Gyula
>>>>>>>>>> 
>>>>>>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont:
>> 2017.
>>>>>>>>>> 
>>>>>>>>> nov.
>>>>>>> 
>>>>>>>> 29., Sze, 9:23):
>>>>>>>>>> 
>>>>>>>>>> +1
>>>>>>>>>>> 
>>>>>>>>>>> Verified:
>>>>>>>>>>> - No missing release Maven artifacts
>>>>>>>>>>> - Staged Apache source & binary convenience releases looks good
>>>>>>>>>>> - NOTICE / LICENSE is correct, README is sane
>>>>>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop
>> 2.8)
>>>>>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
>>>>>>>>>>> 
>>>>>>>>>> configuration
>>>>>>> 
>>>>>>>> details)
>>>>>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency
>> clashes
>>> on
>>>>>>>>>>> cluster execution, works locally in IDE)
>>>>>>>>>>> 
>>>>>>>>>>> Thanks a lot for managing the release Aljoscha!
>>>>>>>>>>> 
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Gordon
>>>>>>>>>>> 
>>>>>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
>>>>>>>>>>> s.rich...@data-artisans.com) wrote:
>>>>>>>>>>> 
>>>>>>>>>>> +1 (non-binding)
>>>>>>>>>>> 
>>>>>>>>>>> I tested Flink in a cluster setup on Google Cloud, YARN-per-job,
>>>>>>>>>>> 
>>>>>>>>>> checked
>>>>>>>>> 
>>>>>>>>>> that for all backends that HA, recovery, at-least-once,
>> end-to-end
>>>>>>>>>>> 
>>>>>>>>>> exactly
>>>>>>>>> 
>>>>>>>>>> once (with Kafka11 Producer), savepoints, externalized
>> checkpoints,
>>>>>>>>>>> 
>>>>>>>>>> and
>>>>>>> 
>>>>>>>> rescaling work correctly.
>>>>>>>>>>> 
>>>>>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
>>>>>>>>>>>> 
>>>>>>>>>>> aljos...@apache.org
>>>>>>> 
>>>>>>>> :
>>>>>>>>>> 
>>>>>>>>>>> +1
>>>>>>>>>>>> 
>>>>>>>>>>>> Verified:
>>>>>>>>>>>> - NOTICE and LICENSE are correct
>>>>>>>>>>>> - source doesn't contain binaries
>>>>>>>>>>>> - verified signatures
>>>>>>>>>>>> - verified hashes
>>>>>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
>>>>>>>>>>>> 
>>>>>>>>>>> release-testing doc)
>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <
>> aljos...@apache.org
>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>> wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Phew, thanks for the update!
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org>
>>> wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was
>>> corrupted
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> (makes
>>>>>>>>> 
>>>>>>>>>> total sense) :P
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017. nov.
>>>>>>>>>>>>>> 28.,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> K,
>>>>>>>>> 
>>>>>>>>>> 11:13):
>>>>>>>>>>> 
>>>>>>>>>>>> Hi Aljoscha,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> building
>>>>>>> 
>>>>>>>> the rc,
>>>>>>>>>>> 
>>>>>>>>>>>> I seem to get this error no matter what I do:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> (shade-hadoop) on
>>>>>>>>>>> 
>>>>>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> invalid
>>>>>>>>> 
>>>>>>>>>> LOC
>>>>>>>>>>> 
>>>>>>>>>>>> header (bad signature) -> [Help 1]
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> (Apache Maven 3.3.9)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Any idea what I am missing?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont:
>>> 2017.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> nov.
>>>>>>>>> 
>>>>>>>>>> 27.,
>>>>>>>>>>> 
>>>>>>>>>>>> H, 19:35):
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi everyone,
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Please review and vote on release candidate #2 for the
>>> version
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 1.4.0, as
>>>>>>>>>>> 
>>>>>>>>>>>> follows:
>>>>>>>>>>>>>>>> [ ] +1, Approve the release
>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide specific
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> comments)
>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>>>>>>>> The complete staging area is available for your review,
>> which
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> includes:
>>>>>>>>>>> 
>>>>>>>>>>>> * JIRA release notes [1],
>>>>>>>>>>>>>>>> * the official Apache source release and binary convenience
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> releases
>>>>>>>>> 
>>>>>>>>>> to
>>>>>>>>>>> 
>>>>>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> key
>>>>>>> 
>>>>>>>> with
>>>>>>>>>>> 
>>>>>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
>>>>>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central
>>> Repository
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> [4],
>>>>>>>>> 
>>>>>>>>>> * source code tag "release-1.4.0-rc1" [5],
>>>>>>>>>>>>>>>> * website pull request listing the new release [6].
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Please have a careful look at the website PR because I
>>> changed
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> some
>>>>>>>>> 
>>>>>>>>>> wording and we're now also releasing a binary without Hadoop
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> dependencies.
>>>>>>>>>>> 
>>>>>>>>>>>> Please use this document for coordinating testing efforts: [7]
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> The only change between RC1 and this RC2 is that the source
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> release
>>>>>>>>> 
>>>>>>>>>> package does not include the erroneously included binary Ruby
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> dependencies
>>>>>>>>>>> 
>>>>>>>>>>>> of the documentation anymore. Because of this I would like to
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> propose a
>>>>>>>>>>> 
>>>>>>>>>>>> shorter voting time and close the vote around the time that RC1
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> would have
>>>>>>>>>>> 
>>>>>>>>>>>> closed. This would mean closing by end of Wednesday. Please let
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> me
>>>>>>> 
>>>>>>>> know if
>>>>>>>>>>> 
>>>>>>>>>>>> you disagree with this. The vote is adopted by majority
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> approval,
>>>>>>> 
>>>>>>>> with at
>>>>>>>>>>> 
>>>>>>>>>>>> least 3 PMC affirmative votes.
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>> Your friendly Release Manager
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> [1]
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?
>> proje
>>>>>>>>>>> 
>>>>>>>>>> ctId=12315522&version=12340533
>>>>>>>>> 
>>>>>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
>>>>>>>>>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>>>>>>>>>>>>>>>> [4]
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://repository.apache.org/
>> content/repositories/orgapache
>>>>>>> flink-1140
>>>>>>> 
>>>>>>>> [5]
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://git-wip-us.apache.org/
>> repos/asf?p=flink.git;a=tag;h=
>>>>>>>>>>> 
>>>>>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
>>>>>>>>> 
>>>>>>>>>> [6] https://github.com/apache/flink-web/pull/95
>>>>>>>>>>>>>>>> [7]
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://docs.google.com/document/d/
>> 1HqYyrNoMSXwo8zBpZj7s39Uz
>>>>>>>>>>> 
>>>>>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
>>>>>>>>> 
>>>>>>>>>> Pro-tip: you can create a settings.xml file with these contents:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> <settings>
>>>>>>>>>>>>>>>> <activeProfiles>
>>>>>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
>>>>>>>>>>>>>>>> </activeProfiles>
>>>>>>>>>>>>>>>> <profiles>
>>>>>>>>>>>>>>>> <profile>
>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>> <repositories>
>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://repository.apache.org/
>> content/repositories/orgapache
>>>>>>>>>>> 
>>>>>>>>>> flink-1140/
>>>>>>>>> 
>>>>>>>>>> </url>
>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>> <id>archetype</id>
>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> https://repository.apache.org/
>> content/repositories/orgapache
>>>>>>>>>>> 
>>>>>>>>>> flink-1140/
>>>>>>>>> 
>>>>>>>>>> </url>
>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>> </repositories>
>>>>>>>>>>>>>>>> </profile>
>>>>>>>>>>>>>>>> </profiles>
>>>>>>>>>>>>>>>> </settings>
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> And reference that in you maven commands via --settings
>>>>>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a
>>> quickstart
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> based
>>>>>>>>> 
>>>>>>>>>> on the
>>>>>>>>>>> 
>>>>>>>>>>>> staged release and for building against the staged jars.
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>> 
>>>>> 
>>>> 
>>> 
>> 

Reply via email to