[
https://issues.apache.org/jira/browse/FLINK-36411?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17885989#comment-17885989
]
Gyula Fora commented on FLINK-36411:
------------------------------------
Sounds a bit complex, why not simply exclude these sinks from scaling and set a
manual parallelism if they are throttled (and you know it ahead of time)
> Allow configuring job vertex throughput limits for auto scaler
> --------------------------------------------------------------
>
> Key: FLINK-36411
> URL: https://issues.apache.org/jira/browse/FLINK-36411
> Project: Flink
> Issue Type: Improvement
> Components: Autoscaler
> Reporter: Sai Sharath Dandi
> Priority: Major
>
> *Problem Statement*
>
> Currently, auto scaler has the ability to detect ineffective scalings and
> prevent further scale ups. However, the ineffective scaling detection does
> not work when there is no scaling history. Moreover, the ineffective scaling
> detection does not prevent the job from making an ineffective scaling for the
> first time.
>
> This is particularly important for some of the sinks that could enforce
> throttling (For example, quota limit for Kafka sink). In these cases, we can
> avoid ineffective scalings by comparing the throughput limit and current
> processing rate.
> *Solution*
> Some high level ideas below
>
> # Allow user to specify static throughput limit at job vertex level
> # Allow sink to implement an interface for publishing the throughput limits.
> If the interface is implemented, we can dynamically fetch the throughput
> limits and make an intelligent scaling decision.
>
--
This message was sent by Atlassian Jira
(v8.20.10#820010)