Hi,

could you tell me how exactly you started the cluster and with which
parameters (configured memory, maybe vcores, etc.)?

Cheers,
Till

On Thu, Jan 3, 2019 at 2:37 AM varuy322 <rui2.w...@intel.com> wrote:

> Hi, Till
> It's very kind of your reply. I got your point, I'm sorry to not make it
> clear about my issue.
> I generated data by streaming benchmark just as the link:
>
> https://github.com/dataArtisans/databricks-benchmark/blob/master/src/main/scala/com/databricks/benchmark/flink/EventGenerator.scala
> .
>
> What I wanna to say is that, let the parallelism is same assume to 96, just
> changes the tm and slots/tm. The first test to configure tm 3 with 32
> slots/tm, there does not occur data skew, three machine receive same data
> and each partition processed approximate data. Then second test to
> configure
> tm 6 with 16 slots/tm, I find each partition processed same data too, but
> one machine processed data more than the other two machine.
>
> I wonder whether the taskmanager(jvm) competes in one machine? What's more,
> how does the streaming benchmark do with backpressure? I test on cluster
> with 4 node, one for master and three for worker, each node with Intel Xeon
> E5-2699 v4 @ 2.20GHz/3.60GHz, 256G memory, 88 cores, 10Gbps network, I
> could
> not find the bottleneck. It confused me!
>
> Best Regards & Thanks
>
> Rui
>
>
>
> -----
> stay hungry, stay foolish.
> --
> Sent from:
> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/
>

Reply via email to