[ 
https://issues.apache.org/jira/browse/HDFS-10175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15236410#comment-15236410
 ] 

Colin Patrick McCabe edited comment on HDFS-10175 at 4/12/16 1:40 AM:
----------------------------------------------------------------------

Thanks, [~mingma].  I agree that if we are going with a map-based approach, we 
can simplify the implementation of HDFS-9579.  This may also save a small 
amount of space... for example, if FileSystem threads don't have any bytes read 
at distance 3 or 4, they will no longer need to store space for 
{{bytesReadDistanceOfThreeOrFour}}.

It also seems like a good idea to make the detailed statistics optional.  Then 
clients could opt-in to paying the overheads, rather than getting the overheads 
imposed whether they want them or not.

[~liuml07] wrote:
bq. 2) Move long getBytesReadByDistance(int distance) from Statistics to 
StatisticsData. If the user is getting bytes read for all distances, she can 
call getData() and then iterate the map/array, in which case the getData() will 
be called only once. For cases of 1K client threads, this may save the effort 
of aggregation. Colin Patrick McCabe may have different comments about this?

Hmm.  I'm not sure I see much of a benefit to this.  Users should not be 
iterating over internal data structures.  It exposes implementation details 
that we don't want to expose.  I also don't see how it's more efficient, since 
you have to iterate over it in either case.

bq. For newly supported APIs, adding an entry in the map and one line of 
increment in the new method will make the magic done. From the point of file 
system APIs, its public methods are not evolving rapidly.

I agree that adding new statistics is relatively easy with the current patch.  
My comment was more that currently, many of these statistics are HDFS-specific. 
 Since MapReduce needs to work with alternate filesystems, it will need to 
handle the case where these detailed statistics are missing or slightly 
different.

bq. Another dimension will be needed for cross-DC analysis, while based on the 
current use case, I don't think this dimension is heavily needed. One point is 
that, all the same kind of file system is sharing the statistic data among 
threads, regardless of the remote/local HDFS clusters.

(revised earlier comment about per-class stats) Yes, it does appear that stats 
are per-class.


was (Author: cmccabe):
Thanks, [~mingma].  I agree that if we are going with a map-based approach, we 
can simplify the implementation of HDFS-9579.  This may also save a small 
amount of space... for example, if FileSystem threads don't have any bytes read 
at distance 3 or 4, they will no longer need to store space for 
{{bytesReadDistanceOfThreeOrFour}}.

It also seems like a good idea to make the detailed statistics optional.  Then 
clients could opt-in to paying the overheads, rather than getting the overheads 
imposed whether they want them or not.

[~liuml07] wrote:
bq. 2) Move long getBytesReadByDistance(int distance) from Statistics to 
StatisticsData. If the user is getting bytes read for all distances, she can 
call getData() and then iterate the map/array, in which case the getData() will 
be called only once. For cases of 1K client threads, this may save the effort 
of aggregation. Colin Patrick McCabe may have different comments about this?

Hmm.  I'm not sure I see much of a benefit to this.  Users should not be 
iterating over internal data structures.  It exposes implementation details 
that we don't want to expose.  I also don't see how it's more efficient, since 
you have to iterate over it in either case.

bq. For newly supported APIs, adding an entry in the map and one line of 
increment in the new method will make the magic done. From the point of file 
system APIs, its public methods are not evolving rapidly.

I agree that adding new statistics is relatively easy with the current patch.  
My comment was more that currently, many of these statistics are HDFS-specific. 
 Since MapReduce needs to work with alternate filesystems, it will need to 
handle the case where these detailed statistics are missing or slightly 
different.

bq. Another dimension will be needed for cross-DC analysis, while based on the 
current use case, I don't think this dimension is heavily needed. One point is 
that, all the same kind of file system is sharing the statistic data among 
threads, regardless of the remote/local HDFS clusters.

It is not the case that "all the same kind of file system is sharing the 
statistic data among threads."  Each {{Filesystem}} instance has a separate 
Statistics object associated with it: 
{code}
  /** Recording statistics per a FileSystem class */
  private static final Map<Class<? extends FileSystem>, Statistics>
    statisticsTable =
      new IdentityHashMap<Class<? extends FileSystem>, Statistics>();
{code}

So as long as you use separate FS instances for the remote FS and local FS, it 
should work for this use-case.  This is also part of the point that I was 
making earlier that memory overheads add up quickly, since each new FS instance 
multiplies the overhead.

> add per-operation stats to FileSystem.Statistics
> ------------------------------------------------
>
>                 Key: HDFS-10175
>                 URL: https://issues.apache.org/jira/browse/HDFS-10175
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: hdfs-client
>            Reporter: Ram Venkatesh
>            Assignee: Mingliang Liu
>         Attachments: HDFS-10175.000.patch, HDFS-10175.001.patch, 
> HDFS-10175.002.patch, HDFS-10175.003.patch, HDFS-10175.004.patch, 
> TestStatisticsOverhead.java
>
>
> Currently FileSystem.Statistics exposes the following statistics:
> BytesRead
> BytesWritten
> ReadOps
> LargeReadOps
> WriteOps
> These are in-turn exposed as job counters by MapReduce and other frameworks. 
> There is logic within DfsClient to map operations to these counters that can 
> be confusing, for instance, mkdirs counts as a writeOp.
> Proposed enhancement:
> Add a statistic for each DfsClient operation including create, append, 
> createSymlink, delete, exists, mkdirs, rename and expose them as new 
> properties on the Statistics object. The operation-specific counters can be 
> used for analyzing the load imposed by a particular job on HDFS. 
> For example, we can use them to identify jobs that end up creating a large 
> number of files.
> Once this information is available in the Statistics object, the app 
> frameworks like MapReduce can expose them as additional counters to be 
> aggregated and recorded as part of job summary.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to