[jira] [Updated] (HDFS-16430) Validate maximum blocks in EC group when adding an EC policy

2022-07-12 Thread Masatake Iwasaki (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HDFS-16430:

Fix Version/s: 3.2.4
   (was: 3.2.3)

> Validate maximum blocks in EC group when adding an EC policy
> 
>
> Key: HDFS-16430
> URL: https://issues.apache.org/jira/browse/HDFS-16430
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ec, erasure-coding
>Affects Versions: 3.3.0, 3.3.1
>Reporter: daimin
>Assignee: daimin
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0, 3.2.4, 3.3.9
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> HDFS EC adopts the last 4 bits of block ID to store the block index in EC 
> block group. Therefore maximum blocks in EC block group is 2^4=16, and which 
> is defined here: HdfsServerConstants#MAX_BLOCKS_IN_GROUP.
> Currently there is no limitation or warning when adding a bad EC policy with 
> numDataUnits + numParityUnits > 16. It only results in read/write error on EC 
> file with bad EC policy. To users this is not very straightforward.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-16430) Validate maximum blocks in EC group when adding an EC policy

2022-01-24 Thread Wei-Chiu Chuang (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang updated HDFS-16430:
---
Fix Version/s: 3.2.3
   3.3.3

> Validate maximum blocks in EC group when adding an EC policy
> 
>
> Key: HDFS-16430
> URL: https://issues.apache.org/jira/browse/HDFS-16430
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ec, erasure-coding
>Affects Versions: 3.3.0, 3.3.1
>Reporter: daimin
>Assignee: daimin
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0, 3.2.3, 3.3.3
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> HDFS EC adopts the last 4 bits of block ID to store the block index in EC 
> block group. Therefore maximum blocks in EC block group is 2^4=16, and which 
> is defined here: HdfsServerConstants#MAX_BLOCKS_IN_GROUP.
> Currently there is no limitation or warning when adding a bad EC policy with 
> numDataUnits + numParityUnits > 16. It only results in read/write error on EC 
> file with bad EC policy. To users this is not very straightforward.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-16430) Validate maximum blocks in EC group when adding an EC policy

2022-01-24 Thread Ayush Saxena (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ayush Saxena updated HDFS-16430:

Fix Version/s: 3.4.0

> Validate maximum blocks in EC group when adding an EC policy
> 
>
> Key: HDFS-16430
> URL: https://issues.apache.org/jira/browse/HDFS-16430
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ec, erasure-coding
>Affects Versions: 3.3.0, 3.3.1
>Reporter: daimin
>Assignee: daimin
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> HDFS EC adopts the last 4 bits of block ID to store the block index in EC 
> block group. Therefore maximum blocks in EC block group is 2^4=16, and which 
> is defined here: HdfsServerConstants#MAX_BLOCKS_IN_GROUP.
> Currently there is no limitation or warning when adding a bad EC policy with 
> numDataUnits + numParityUnits > 16. It only results in read/write error on EC 
> file with bad EC policy. To users this is not very straightforward.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-16430) Validate maximum blocks in EC group when adding an EC policy

2022-01-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-16430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HDFS-16430:
--
Labels: pull-request-available  (was: )

> Validate maximum blocks in EC group when adding an EC policy
> 
>
> Key: HDFS-16430
> URL: https://issues.apache.org/jira/browse/HDFS-16430
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ec, erasure-coding
>Affects Versions: 3.3.0, 3.3.1
>Reporter: daimin
>Assignee: daimin
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> HDFS EC adopts the last 4 bits of block ID to store the block index in EC 
> block group. Therefore maximum blocks in EC block group is 2^4=16, and which 
> is defined here: HdfsServerConstants#MAX_BLOCKS_IN_GROUP.
> Currently there is no limitation or warning when adding a bad EC policy with 
> numDataUnits + numParityUnits > 16. It only results in read/write error on EC 
> file with bad EC policy. To users this is not very straightforward.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org