[
https://issues.apache.org/jira/browse/HADOOP-15703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Thomas Marquardt updated HADOOP-15703:
--------------------------------------
Resolution: Fixed
Status: Resolved (was: Patch Available)
commit e5d50c734dd7160daeee789b0fd6ca614319bd1f
Author: Thomas Marquardt <[email protected]>
Date: Thu Sep 6 01:35:49 2018 +0000
Fixing findbugs and license issues related to:
HADOOP-15703. ABFS - Implement client-side throttling.
Contributed by Sneha Varma and Thomas Marquardt.
> ABFS - Implement client-side throttling
> ----------------------------------------
>
> Key: HADOOP-15703
> URL: https://issues.apache.org/jira/browse/HADOOP-15703
> Project: Hadoop Common
> Issue Type: Sub-task
> Reporter: Sneha Varma
> Assignee: Thomas Marquardt
> Priority: Major
> Attachments: HADOOP-15703-HADOOP-15407-001.patch,
> HADOOP-15703-HADOOP-15407-002.patch, HADOOP-15703-HADOOP-15407-003.patch,
> HADOOP-15703-HADOOP-15407-004.patch
>
>
> Big data workloads frequently exceed the AzureBlobFS max ingress and egress
> limits
> (https://docs.microsoft.com/en-us/azure/storage/common/storage-scalability-targets).
> For example, the max ingress limit for a GRS account in the United States is
> currently 10 Gbps. When the limit is exceeded, the AzureBlobFS service fails
> a percentage of incoming requests, and this causes the client to initiate the
> retry policy. The retry policy delays requests by sleeping, but the sleep
> duration is independent of the client throughput and account limit. This
> results in low throughput, due to the high number of failed requests and
> thrashing causes by the retry policy.
> To fix this, we introduce a client-side throttle which minimizes failed
> requests and maximizes throughput.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]