[
https://issues.apache.org/jira/browse/FLUME-643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Ashish Paliwal resolved FLUME-643.
----------------------------------
Resolution: Won't Fix
Fix Version/s: v0.9.5
Won't fix. 0.X branch not maintained anymore
> Logging from Scribe to Hadoop via Flume breaks UTF-8 encoding
> -------------------------------------------------------------
>
> Key: FLUME-643
> URL: https://issues.apache.org/jira/browse/FLUME-643
> Project: Flume
> Issue Type: Bug
> Components: Sinks+Sources
> Affects Versions: v0.9.3
> Reporter: Disabled imported user
> Priority: Minor
> Fix For: v0.9.5
>
>
> Log messages with UTF-8 Characters like äöü end up with broken in Hadoop when
> logging via Scribe. We used a simple Setup with:
> exec config scribe_input scribe "scribe(1463)
> "collectorSink("hdfs://localhost/testing/", "test",1000)"
> exec spawn testserver scribe_input
> We usually use avrojson as collector output format and gzip for compression,
> but the chars are broken if we deactivate both.
> The Problem seems to occur when flume writes the files into Hadoop, as in a
> more complicated setup like:
> exec config scribe_input scribe "scribe(1463)" autoDFOChain
> exec config hdfs scribe autoCollectorSource
> "collectorSink("hdfs://localhost/testing/", "test",1000)"
> exec spawn testserver1 scribe_input
> exec spawn testserver2 hdfs
> the chars are still ok in the DFO Logs on testserver1
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)