[ 
https://issues.apache.org/jira/browse/HADOOP-15616?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16718572#comment-16718572
 ] 

YangY edited comment on HADOOP-15616 at 12/12/18 7:57 AM:
----------------------------------------------------------

Thanks [~xyao] for comment on this code.

Here are the answers to your comments:

1. Changes under hadoop-tools/hadoop-aliyun unrelated to this patch.
 This may be a misoperation when formatting my code, and the error has been 
corrected in the new path.

2. Should we put hadoop-cos under hadoop-tools project like s3a, adsl, etc. 
instead of hadoop-cloud-storage-project?
 At first, I also thought I should put it under the hadoop-tools project. 
However, as steve's comment above, using "hadoop-cloud-storage-project" seems 
more appropriate,isn't it?

3. More description to SecretKey -> SecretID.
 Thank you for your reminder, I will add a detailed description for getting 
SecretKey and SecretID in our document.

4. BufferPool.java: since it sets the disk buffer file delete on exist, does it 
support recovery if client restart?
 BufferPool is a shared buffer pool. It initially provides two buffer types: 
Memory and Disk. The latter uses the memory file mapping to construct a byte 
buffer that can be used by other classes uniformly.
 Therefore, it can not support recovery if client restart. After all, the disk 
buffer is mapped a temporal file, and it will be cleaned up automatically when 
the Java Virtual Machine exists.
 
 In the latest patch, I further optimize it by combining two buffer types: 
memory usage and buffer performance. For this reason, the type of buffers here 
will not be visible to the user.

Finally, I look forward to your more comments.


was (Author: yuyang733):
Thanks [~xyao] for comment on this code.

Here are the answers to your comments:

1. Changes under hadoop-tools/hadoop-aliyun unrelated to this patch.
    This may be a misoperation when formatting my code, and the error has been 
corrected in the new path.

2. Should we put hadoop-cos under hadoop-tools project like s3a, adsl, etc. 
instead of hadoop-cloud-storage-project?
    At first, I also thought I should put it under the hadoop-tools project. 
However, as steve's comment above, using "hadoop-cloud-storage-project" seems 
more appropriate,isn't it?

3. More description to SecretKey -> SecretID.
    Thank you for your reminder, I will add a detailed description for getting 
SecretKey and SecretID in our document.

4. BufferPool.java: since it sets the disk buffer file delete on exist, does it 
support recovery if client restart?
    BufferPool is a shared buffer pool. It initially provides two buffer types: 
 Memory and Disk. The latter uses the memory file mapping to construct a byte 
buffer that can be used by other classes uniformly.
    Therefore, it can not support recovery if client restart. After all, the 
disk buffer is mapped a temporal file, and it will be cleaned up automatically 
when the Java Virtual Machine exists.


Finally, I look forward to your more comments.


> Incorporate Tencent Cloud COS File System Implementation
> --------------------------------------------------------
>
>                 Key: HADOOP-15616
>                 URL: https://issues.apache.org/jira/browse/HADOOP-15616
>             Project: Hadoop Common
>          Issue Type: New Feature
>          Components: fs/cos
>            Reporter: Junping Du
>            Assignee: YangY
>            Priority: Major
>         Attachments: HADOOP-15616.001.patch, HADOOP-15616.002.patch, 
> HADOOP-15616.003.patch, HADOOP-15616.004.patch, HADOOP-15616.005.patch, 
> Tencent-COS-Integrated.pdf
>
>
> Tencent cloud is top 2 cloud vendors in China market and the object store COS 
> ([https://intl.cloud.tencent.com/product/cos]) is widely used among China’s 
> cloud users but now it is hard for hadoop user to access data laid on COS 
> storage as no native support for COS in Hadoop.
> This work aims to integrate Tencent cloud COS with Hadoop/Spark/Hive, just 
> like what we do before for S3, ADL, OSS, etc. With simple configuration, 
> Hadoop applications can read/write data from COS without any code change.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to