[ 
http://issues.apache.org/jira/browse/HADOOP-237?page=comments#action_12420710 ] 

David Bowen commented on HADOOP-237:
------------------------------------


OK, maybe this is no big deal since the records are small.  The idea of a 
record was to be a bunch of things that should be updated simultaneously, but 
maybe using it for a small number of things that are updated independently is 
OK.  Splitting the record into two would cost a bit of extra space in the 
client library (since the overhead of an extra record in a hash table outweighs 
the savings of  4 bytes per record) and would not save much in the cost of an 
update.



> Standard set of Performance Metrics for Hadoop
> ----------------------------------------------
>
>          Key: HADOOP-237
>          URL: http://issues.apache.org/jira/browse/HADOOP-237
>      Project: Hadoop
>         Type: Improvement

>   Components: metrics
>     Versions: 0.3.0
>  Environment: All
>     Reporter: Milind Bhandarkar
>     Assignee: Milind Bhandarkar
>  Attachments: hadoop-metrics.patch
>
> I am starting to use Hadoop's shiny new Metrics API to publish performance 
> (and other) Metrics of running jobs and other daemons.
> Which performance metrics are people interested in seeing ? If possible, 
> please group them according to modules, such as map-reduce, dfs, 
> general-cluster-related etc. I will follow this process:
> 1. collect this list
> 2. assess feasibility of obtaining metric
> 3. assign context/record/metrics names
> 4. seek approval for names
> 5. instrument the code.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
   http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see:
   http://www.atlassian.com/software/jira

Reply via email to