[
https://issues.apache.org/jira/browse/HADOOP-17333?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
minchengbo updated HADOOP-17333:
--------------------------------
Description:
Got sink exception,when set
datanode.sink.ganglia.metric.filter.exclude=metricssystem in
hadoop-metrics2.properties ,
java.lang.ClassCastException:
org.apache.hadoop.metrics2.impl.MetricsRecordFiltered$1 cannot be cast to
java.util.Collection
at
org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30.putMetrics(GangliaSink30.java:165)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:184)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:43)
at org.apache.hadoop.metrics2.impl.SinkQueue.consumeAll(SinkQueue.java:87)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.publishMetricsFromQueue(MetricsSinkAdapter.java:135)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter$1.run(MetricsSinkAdapter.java:89)
//////////////////////////////////////////////////
This case can show the exception
public static void main(String[] args) {
List<AbstractMetric> metricsd=new
LinkedList<AbstractMetric>();
MetricsInfo info=MsInfo.ProcessName;
long timestamp=System.currentTimeMillis();
List<MetricsTag> tags=new LinkedList<>();
org.apache.hadoop.metrics2.impl.MetricsRecordImpl recordimp =
new MetricsRecordImpl(info, timestamp, tags, metricsd);
MetricsFilter filter=new RegexFilter();
MetricsRecordFiltered recordfilter=new
MetricsRecordFiltered(recordimp,filter);
SubsetConfiguration conf=new SubsetConfiguration(new
PropertyListConfiguration(),"test");
conf.addProperty(AbstractGangliaSink.SUPPORT_SPARSE_METRICS_PROPERTY, true);
GangliaSink30 ganliasink=new GangliaSink30();
ganliasink.init(conf);
ganliasink.putMetrics(recordfilter);
}
///////////////////////////////////////////////////////////////
The root cause is:
Gets a Iterable object in MetricsRecordFiltered.java:
@Override public Iterable<AbstractMetric> metrics() {
return new Iterable<AbstractMetric>() {
final Iterator<AbstractMetric> it = delegate.metrics().iterator();
@Override public Iterator<AbstractMetric> iterator() {
return new AbstractIterator<AbstractMetric>() {
@Override public AbstractMetric computeNext() {
while (it.hasNext()) {
AbstractMetric next = it.next();
if (filter.accepts(next.name())) {
return next;
}
}
return (AbstractMetric)endOfData();
}
};
}
};
}
but convert to Collection in GangliaSink30.java line 164
Collection<AbstractMetric> metrics = (Collection<AbstractMetric>) record
.metrics();
was:
Got sink exception,when set
datanode.sink.ganglia.metric.filter.exclude=metricssystem in
hadoop-metrics2.properties ,
java.lang.ClassCastException:
org.apache.hadoop.metrics2.impl.MetricsRecordFiltered$1 cannot be cast to
java.util.Collection
at
org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30.putMetrics(GangliaSink30.java:165)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:184)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:43)
at org.apache.hadoop.metrics2.impl.SinkQueue.consumeAll(SinkQueue.java:87)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.publishMetricsFromQueue(MetricsSinkAdapter.java:135)
at
org.apache.hadoop.metrics2.impl.MetricsSinkAdapter$1.run(MetricsSinkAdapter.java:89)
Environment:
was:
This case can show the exception
public static void main(String[] args) {
List<AbstractMetric> metricsd=new
LinkedList<AbstractMetric>();
MetricsInfo info=MsInfo.ProcessName;
long timestamp=System.currentTimeMillis();
List<MetricsTag> tags=new LinkedList<>();
org.apache.hadoop.metrics2.impl.MetricsRecordImpl recordimp =
new MetricsRecordImpl(info, timestamp, tags, metricsd);
MetricsFilter filter=new RegexFilter();
MetricsRecordFiltered recordfilter=new
MetricsRecordFiltered(recordimp,filter);
SubsetConfiguration conf=new SubsetConfiguration(new
PropertyListConfiguration(),"test");
conf.addProperty(AbstractGangliaSink.SUPPORT_SPARSE_METRICS_PROPERTY, true);
GangliaSink30 ganliasink=new GangliaSink30();
ganliasink.init(conf);
ganliasink.putMetrics(recordfilter);
}
///////////////////////////////////////////////////////////////
The root cause is:
Gets a Iterable object in MetricsRecordFiltered.java:
@Override public Iterable<AbstractMetric> metrics() {
return new Iterable<AbstractMetric>() {
final Iterator<AbstractMetric> it = delegate.metrics().iterator();
@Override public Iterator<AbstractMetric> iterator() {
return new AbstractIterator<AbstractMetric>() {
@Override public AbstractMetric computeNext() {
while (it.hasNext()) {
AbstractMetric next = it.next();
if (filter.accepts(next.name())) {
return next;
}
}
return (AbstractMetric)endOfData();
}
};
}
};
}
but convert to Collection in GangliaSink30.java line 164
Collection<AbstractMetric> metrics = (Collection<AbstractMetric>) record
.metrics();
> MetricsRecordFiltered error
> ---------------------------
>
> Key: HADOOP-17333
> URL: https://issues.apache.org/jira/browse/HADOOP-17333
> Project: Hadoop Common
> Issue Type: Bug
> Components: common
> Affects Versions: 3.2.1
> Environment:
> Reporter: minchengbo
> Priority: Minor
>
> Got sink exception,when set
> datanode.sink.ganglia.metric.filter.exclude=metricssystem in
> hadoop-metrics2.properties ,
> java.lang.ClassCastException:
> org.apache.hadoop.metrics2.impl.MetricsRecordFiltered$1 cannot be cast to
> java.util.Collection
> at
> org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30.putMetrics(GangliaSink30.java:165)
> at
> org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:184)
> at
> org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.consume(MetricsSinkAdapter.java:43)
> at org.apache.hadoop.metrics2.impl.SinkQueue.consumeAll(SinkQueue.java:87)
> at
> org.apache.hadoop.metrics2.impl.MetricsSinkAdapter.publishMetricsFromQueue(MetricsSinkAdapter.java:135)
> at
> org.apache.hadoop.metrics2.impl.MetricsSinkAdapter$1.run(MetricsSinkAdapter.java:89)
> //////////////////////////////////////////////////
> This case can show the exception
> public static void main(String[] args) {
> List<AbstractMetric> metricsd=new
> LinkedList<AbstractMetric>();
> MetricsInfo info=MsInfo.ProcessName;
> long timestamp=System.currentTimeMillis();
> List<MetricsTag> tags=new LinkedList<>();
> org.apache.hadoop.metrics2.impl.MetricsRecordImpl recordimp =
> new MetricsRecordImpl(info, timestamp, tags, metricsd);
> MetricsFilter filter=new RegexFilter();
> MetricsRecordFiltered recordfilter=new
> MetricsRecordFiltered(recordimp,filter);
> SubsetConfiguration conf=new SubsetConfiguration(new
> PropertyListConfiguration(),"test");
>
> conf.addProperty(AbstractGangliaSink.SUPPORT_SPARSE_METRICS_PROPERTY, true);
> GangliaSink30 ganliasink=new GangliaSink30();
> ganliasink.init(conf);
> ganliasink.putMetrics(recordfilter);
>
> }
> ///////////////////////////////////////////////////////////////
> The root cause is:
> Gets a Iterable object in MetricsRecordFiltered.java:
> @Override public Iterable<AbstractMetric> metrics() {
> return new Iterable<AbstractMetric>() {
> final Iterator<AbstractMetric> it = delegate.metrics().iterator();
> @Override public Iterator<AbstractMetric> iterator() {
> return new AbstractIterator<AbstractMetric>() {
> @Override public AbstractMetric computeNext() {
> while (it.hasNext()) {
> AbstractMetric next = it.next();
> if (filter.accepts(next.name())) {
> return next;
> }
> }
> return (AbstractMetric)endOfData();
> }
> };
> }
> };
> }
> but convert to Collection in GangliaSink30.java line 164
> Collection<AbstractMetric> metrics = (Collection<AbstractMetric>)
> record
> .metrics();
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]