If the field is declared as binary, you can do this select convert(binary_col using utf8), col2,... from table where ... $CONDITIONS
On Mon, Jul 15, 2013 at 6:52 PM, varun kumar gullipalli < [email protected]> wrote: > Hi Jarcec, > > I am validating the data by running the following command, > > hadoop fs -text <hdfs cluster> > > I think there is no issue with the shell (correct me if am wrong) because > I am connecting to MySQL database from the same shell(command line) and > could view the source data properly. > > Initially we observed that the following conf files doesn't have utf-8 > encoding. > <?xml version="1.0" encoding="UTF-8"?> > > sqoop-site.xml > sqoop=site-template.xml > > But no luck after making the changes too. > > Thanks, > Varun > > ------------------------------ > *From:* Jarek Jarcec Cecho <[email protected]> > *To:* [email protected]; varun kumar gullipalli < > [email protected]> > *Sent:* Monday, July 15, 2013 6:37 PM > *Subject:* Re: Sqoop - utf-8 data load issue > > Hi Varun, > we are usually not seeing any issues with transferring text data in UTF. > How are > you validating the imported file? I can imagine that your shell might be > messing > the encoding. > > Jarcec > > On Mon, Jul 15, 2013 at 06:27:25PM -0700, varun kumar gullipalli wrote: > > > > > > Hi, > > I am importing data from MySql to HDFS using free-form query import. > > It works fine but facing issue when the data is utf-8.The source(MySql) > db is utf-8 compatible but looks like sqoop is converting the data during > import. > > Example - The source value - elémeñt is loaded as elémeñt to HDFS. > > Please provide a solution for this. > > Thanks in advance! > >
