[ http://issues.apache.org/jira/browse/HADOOP-237?page=comments#action_12420710 ]
David Bowen commented on HADOOP-237: ------------------------------------ OK, maybe this is no big deal since the records are small. The idea of a record was to be a bunch of things that should be updated simultaneously, but maybe using it for a small number of things that are updated independently is OK. Splitting the record into two would cost a bit of extra space in the client library (since the overhead of an extra record in a hash table outweighs the savings of 4 bytes per record) and would not save much in the cost of an update. > Standard set of Performance Metrics for Hadoop > ---------------------------------------------- > > Key: HADOOP-237 > URL: http://issues.apache.org/jira/browse/HADOOP-237 > Project: Hadoop > Type: Improvement > Components: metrics > Versions: 0.3.0 > Environment: All > Reporter: Milind Bhandarkar > Assignee: Milind Bhandarkar > Attachments: hadoop-metrics.patch > > I am starting to use Hadoop's shiny new Metrics API to publish performance > (and other) Metrics of running jobs and other daemons. > Which performance metrics are people interested in seeing ? If possible, > please group them according to modules, such as map-reduce, dfs, > general-cluster-related etc. I will follow this process: > 1. collect this list > 2. assess feasibility of obtaining metric > 3. assign context/record/metrics names > 4. seek approval for names > 5. instrument the code. -- This message is automatically generated by JIRA. - If you think it was sent incorrectly contact one of the administrators: http://issues.apache.org/jira/secure/Administrators.jspa - For more information on JIRA, see: http://www.atlassian.com/software/jira
