[ 
https://issues.apache.org/jira/browse/HDDS-3630?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

runzhiwang updated HDDS-3630:
-----------------------------
    Description: Currently, one rocksdb for one container. one container has 
5GB capacity. 10TB data need more than 2000 rocksdb in one datanode.  It's 
difficult to limit the memory of 2000 rocksdb. So maybe we should use one 
rocksdb for each disk.  (was: Currently, one rocksdb for one container. one 
container has 5GB capacity. 10TB data need more than 2000 rocksdb in one 
datanode.  It's difficult to limit the memory of 2000 rocksdb. So maybe we 
should only use one rocksdb for all containers.)

> Merge rocksdb into one in datanode
> ----------------------------------
>
>                 Key: HDDS-3630
>                 URL: https://issues.apache.org/jira/browse/HDDS-3630
>             Project: Hadoop Distributed Data Store
>          Issue Type: Sub-task
>            Reporter: runzhiwang
>            Assignee: runzhiwang
>            Priority: Major
>
> Currently, one rocksdb for one container. one container has 5GB capacity. 
> 10TB data need more than 2000 rocksdb in one datanode.  It's difficult to 
> limit the memory of 2000 rocksdb. So maybe we should use one rocksdb for each 
> disk.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to