[ 
https://issues.apache.org/jira/browse/PHOENIX-5452?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16916000#comment-16916000
 ] 

sujit potluri edited comment on PHOENIX-5452 at 8/26/19 5:46 PM:
-----------------------------------------------------------------

Team,

I am facing following problem when trying to use lzo file:
{code:java}
//hdfs dfs -put data.csv.lzo /tmp{code}
{code:java}
HADOOP_CLASSPATH=/usr/hdp/2.6.4.0-91/hbase/lib/hbase-protocol-1.1.2.2.6.4.0-91.jar:/etc/hbase/conf
 hadoop jar /usr/hdp/current/phoenix-client/phoenix-4.7.0.2.6.4.0-91-client.jar 
org.apache.phoenix.mapreduce.CsvBulkLoadTool --table EXAMPLE --input 
/tmp/data.csv.lzo{code}
{code:java}
 19/08/22 15:28:47 INFO mapreduce.AbstractBulkLoadTool: Running MapReduce 
import job from /tmp/data.csv.lzo to 
/tmp/0d55520c-2db7-44a0-b2d6-926d4e4b70df{code}
{code:java}
19/08/22 15:29:54 INFO input.FileInputFormat: Total input paths to process : 1 
19/08/22 15:29:54 INFO lzo.GPLNativeCodeLoader: Loaded native gpl library 
19/08/22 15:29:54 INFO lzo.LzoCodec: Successfully loaded & initialized 
native-lzo library [hadoop-lzo rev c3bf57bd7310159c984a56403f3570d6ccb70f1c] 
19/08/22 15:29:54 INFO mapreduce.JobSubmitter: number of splits:1 19/08/22 
15:29:54 INFO mapreduce.JobSubmitter: Submitting tokens for job: 
job_1566486875553_0001{code}
{code:java}
19/08/22 15:29:55 INFO mapreduce.Job: Running job: 
job_1566486875553_000119/08/22 15:29:55 INFO mapreduce.Job: Running job: 
job_1566486875553_000119/08/22 15:30:08 INFO mapreduce.Job: Job 
job_1566486875553_0001 running in uber mode : false19/08/22 15:30:08 INFO 
mapreduce.Job:  map 0% reduce 0%19/08/22 15:30:20 INFO mapreduce.Job: Task Id : 
attempt_1566486875553_0001_m_000000_0, Status : FAILEDError: 
java.lang.RuntimeException: java.lang.IllegalArgumentException: CSV record does 
not have enough values (has 1, but needs 3) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:201)
 at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:73)
 at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) at 
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787) at 
org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) at 
org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:170) at 
java.security.AccessController.doPrivileged(Native Method) at 
javax.security.auth.Subject.doAs(Subject.java:422) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)
 at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:164)Caused by: 
java.lang.IllegalArgumentException: CSV record does not have enough values (has 
1, but needs 3) at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:81)
 at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:51)
 at org.apache.phoenix.util.UpsertExecutor.execute(UpsertExecutor.java:133) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:170)
 ... 9 more 19/08/22 15:30:29 INFO mapreduce.Job: Task Id : 
attempt_1566486875553_0001_m_000000_1, Status : FAILEDError: 
java.lang.RuntimeException: java.lang.IllegalArgumentException: CSV record does 
not have enough values (has 1, but needs 3) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:201)
 at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:73)
 at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) at 
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787) at 
org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) at 
org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:170) at 
java.security.AccessController.doPrivileged(Native Method) at 
javax.security.auth.Subject.doAs(Subject.java:422) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)
 at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:164)Caused by: 
java.lang.IllegalArgumentException: CSV record does not have enough values (has 
1, but needs 3) at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:81)
 at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:51)
 at org.apache.phoenix.util.UpsertExecutor.execute(UpsertExecutor.java:133) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:170)
 ... 9 more 19/08/22 15:30:36 INFO mapreduce.Job: Task Id : 
attempt_1566486875553_0001_m_000000_2, Status : FAILEDError: 
java.lang.RuntimeException: java.lang.IllegalArgumentException: CSV record does 
not have enough values (has 1, but needs 3) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:201)
 at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:73)
 at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) at 
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787) at 
org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) at 
org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:170) at 
java.security.AccessController.doPrivileged(Native Method) at 
javax.security.auth.Subject.doAs(Subject.java:422) at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)
 at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:164)Caused by: 
java.lang.IllegalArgumentException: CSV record does not have enough values (has 
1, but needs 3) at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:81)
 at 
org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:51)
 at org.apache.phoenix.util.UpsertExecutor.execute(UpsertExecutor.java:133) at 
org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:170)
 ... 9 more 19/08/22 15:30:49 INFO mapreduce.Job:  map 100% reduce 100%19/08/22 
15:30:49 INFO mapreduce.Job: Job job_1566486875553_0001 failed with state 
FAILED due to: Task failed task_1566486875553_0001_m_000000Job failed as tasks 
failed. failedMaps:1 failedReduces:0 19/08/22 15:30:49 INFO mapreduce.Job: 
Counters: 13 Job Counters  Failed map tasks=4 Killed reduce tasks=1 Launched 
map tasks=4 Other local map tasks=3 Data-local map tasks=1 Total time spent by 
all maps in occupied slots (ms)=34657 Total time spent by all reduces in 
occupied slots (ms)=0 Total time spent by all map tasks (ms)=34657 Total time 
spent by all reduce tasks (ms)=0 Total vcore-milliseconds taken by all map 
tasks=34657 Total vcore-milliseconds taken by all reduce tasks=0 Total 
megabyte-milliseconds taken by all map tasks=35488768 Total 
megabyte-milliseconds taken by all reduce tasks=0
{code}
 

any ideas why it may not be working.

Regards

Sujit P


was (Author: sujitp0123):
Team,

 

I am facing following problem when trying to use lzo file:


{code:java}
// 
{code}

> Document CsvBulkLoadTool can accept compressed files
> ----------------------------------------------------
>
>                 Key: PHOENIX-5452
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-5452
>             Project: Phoenix
>          Issue Type: Improvement
>    Affects Versions: 5.0.0, 4.15.0, 4.14.2, 4.14.3
>            Reporter: Artem Ervits
>            Assignee: Artem Ervits
>            Priority: Major
>              Labels: newbie
>         Attachments: PHOENIX-5452-site.patch, PHOENIX-5452.v0.patch
>
>
> today, the bulk import tool only accepts uncompressed csv files, this Jira is 
> to add ability to import compressed data.
> {code:java}
> Error: java.lang.RuntimeException: 
> org.apache.phoenix.schema.IllegalDataException: java.sql.SQLException: ERROR 
> 201 (22000): Illegal data.Error: java.lang.RuntimeException: 
> org.apache.phoenix.schema.IllegalDataException: java.sql.SQLException: ERROR 
> 201 (22000): Illegal data. at 
> org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:205)
>  at 
> org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:77)
>  at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:146) at 
> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787) at 
> org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) at 
> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:175) at 
> java.security.AccessController.doPrivileged(Native Method) at 
> javax.security.auth.Subject.doAs(Subject.java:422) at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
>  at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:169)Caused by: 
> org.apache.phoenix.schema.IllegalDataException: java.sql.SQLException: ERROR 
> 201 (22000): Illegal data. at 
> org.apache.phoenix.schema.types.PDataType.newIllegalDataException(PDataType.java:310)
>  at org.apache.phoenix.schema.types.PLong.toObject(PLong.java:228) at 
> org.apache.phoenix.util.csv.CsvUpsertExecutor$SimpleDatatypeConversionFunction.apply(CsvUpsertExecutor.java:204)
>  at 
> org.apache.phoenix.util.csv.CsvUpsertExecutor$SimpleDatatypeConversionFunction.apply(CsvUpsertExecutor.java:120)
>  at 
> org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:85)
>  at 
> org.apache.phoenix.util.csv.CsvUpsertExecutor.execute(CsvUpsertExecutor.java:52)
>  at org.apache.phoenix.util.UpsertExecutor.execute(UpsertExecutor.java:133) 
> at 
> org.apache.phoenix.mapreduce.FormatToBytesWritableMapper.map(FormatToBytesWritableMapper.java:174)
>  ... 9 moreCaused by: java.sql.SQLException: ERROR 201 (22000): Illegal data. 
> at 
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:498)
>  at 
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:150)
>  ... 17 moreCaused by: java.lang.NumberFormatException: For input string: 
> "data.csv0000664000175000017500000005065313527306733012416 0ustar  
> vagrantvagrant71027" at 
> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) 
> at java.lang.Long.parseLong(Long.java:589) at 
> java.lang.Long.parseLong(Long.java:631) at 
> org.apache.phoenix.schema.types.PLong.toObject(PLong.java:226) ... 15 more 
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

Reply via email to