[ 
https://issues.apache.org/jira/browse/FLUME-643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Paliwal resolved FLUME-643.
----------------------------------
       Resolution: Won't Fix
    Fix Version/s: v0.9.5

Won't fix. 0.X branch not maintained anymore

> Logging from Scribe to Hadoop via Flume breaks UTF-8 encoding
> -------------------------------------------------------------
>
>                 Key: FLUME-643
>                 URL: https://issues.apache.org/jira/browse/FLUME-643
>             Project: Flume
>          Issue Type: Bug
>          Components: Sinks+Sources
>    Affects Versions: v0.9.3
>            Reporter: Disabled imported user
>            Priority: Minor
>             Fix For: v0.9.5
>
>
> Log messages with UTF-8 Characters like äöü end up with broken in Hadoop when 
> logging via Scribe. We used a simple Setup with:
> exec config scribe_input scribe "scribe(1463) 
> "collectorSink("hdfs://localhost/testing/", "test",1000)"
> exec spawn testserver scribe_input
> We usually use avrojson as collector output format and gzip for compression, 
> but the chars are broken if we deactivate both. 
> The Problem seems to occur when flume writes the files into Hadoop, as in a 
> more complicated setup like:
> exec config scribe_input scribe "scribe(1463)"  autoDFOChain 
> exec config hdfs scribe autoCollectorSource 
> "collectorSink("hdfs://localhost/testing/", "test",1000)"
> exec spawn testserver1 scribe_input
> exec spawn testserver2 hdfs
> the chars are still ok in the DFO Logs on testserver1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to