Hi Hyunsik Choi, I will try it. Thanks a lot. :)
2013/12/6 Hyunsik Choi <[email protected]> > Hi Jae Lee, > > TAJO-304 (https://issues.apache.org/jira/browse/TAJO-304) was > resolved. Now, 'DROP TABLE' does not remove data, and just remove a > table information from catalog. And, 'DROP TABLE table_name PURGE' > removes a table information as well as removes data. > > - hyunsik > > On Wed, Nov 13, 2013 at 11:02 PM, Jae Lee <[email protected]> wrote: > > Hi Hyunsik Choi, > > > > Thank you for your reply and jira issues. > > I loot forward to meet 0.8-incubating. > > > > Regards, > > Jae. > > > > > > > > 2013/11/13 Hyunsik Choi <[email protected]> > > > >> Hi Jae Lee, > >> > >> Thank you for your interest in Tajo. From your questions, I've just > >> added two jira issues, and scheduled them to 0.8-incubating. They will > >> be resolved soon. > >> > >> TAJO-304: drop table command should not remove data files in default > >> (https://issues.apache.org/jira/browse/TAJO-304) > >> TAJO-305: Implement killQuery feature > >> (https://issues.apache.org/jira/browse/TAJO-305) > >> > >> - hyunsik > >> > >> On Tue, Nov 12, 2013 at 10:58 AM, Jinho Kim <[email protected]> wrote: > >> > Hi Jae, > >> > > >> > Could you try to do the following? I'm suspecting that some column > values > >> > may contain Hive's default null character '\\N'. > >> > > >> > CREATE EXTERNAL TABLE tablename( A int, B text, C text) > >> > > >> > using csv with('csvfile.delimiter'='\001','csvfile.null'='\\N') > >> > > >> > > >> > Besides, in Hive, some value that cannot be parsed are dealt as NULL. > >> But, > >> > in the current Tajo, they causes errors. > >> > > >> > > >> > Thanks > >> > -Jinho > >> > > >> > > >> > 2013/11/12 Jae Lee <[email protected]> > >> > > >> >> Hi Jihoon, > >> >> > >> >> Thank you for your answer. > >> >> About Q1 has more question. > >> >> I already waiting for query result long time. I think that is not > >> normaly. > >> >> COUNT(*) query got result only 300sec, but DISTINCT, GROUPBY and SUM > >> query > >> >> is excuting whole day. > >> >> I found another error message. Please see below message. > >> >> Error message is about integer type column name of "Year". > >> >> The query was "select distinct year from departuredelay;" > >> >> I was execute same query on Hive. It had no error. > >> >> But Year column has some null or blank data. > >> >> Table was create EXTERNAL table with several CSV files on HDFS. > >> >> > >> > --------------------------------------------------------------------------- > >> >> > >> >> 2013-11-11 18:34:01,436 ERROR worker.Task (Task.java:run(363)) - > >> >> java.lang.NumberFormatException: For input string: "Year" > >> >> at > >> >> > >> >> > >> > java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) > >> >> at java.lang.Integer.parseInt(Integer.java:492) > >> >> at java.lang.Integer.valueOf(Integer.java:582) > >> >> at > org.apache.tajo.datum.DatumFactory.createInt4(DatumFactory.java:140) > >> >> at > >> org.apache.tajo.storage.LazyTuple.createByTextBytes(LazyTuple.java:313) > >> >> at org.apache.tajo.storage.LazyTuple.get(LazyTuple.java:126) > >> >> at org.apache.tajo.engine.eval.FieldEval.eval(FieldEval.java:58) > >> >> at org.apache.tajo.engine.planner.Projector.eval(Projector.java:87) > >> >> at > >> >> > >> >> > >> > org.apache.tajo.engine.planner.physical.SeqScanExec.next(SeqScanExec.java:111) > >> >> at > >> >> > >> >> > >> > org.apache.tajo.engine.planner.physical.HashAggregateExec.compute(HashAggregateExec.java:57) > >> >> at > >> >> > >> >> > >> > org.apache.tajo.engine.planner.physical.HashAggregateExec.next(HashAggregateExec.java:83) > >> >> at > >> >> > >> >> > >> > org.apache.tajo.engine.planner.physical.PartitionedStoreExec.next(PartitionedStoreExec.java:121) > >> >> at org.apache.tajo.worker.Task.run(Task.java:355) > >> >> at org.apache.tajo.worker.TaskRunner$1.run(TaskRunner.java:376) > >> >> at java.lang.Thread.run(Thread.java:744) > >> >> > >> >> > >> >> > >> > ----------------------------------------------------------------------------- > >> >> > >> >> > >> >> > >> >> Also I attach file tajo-site.xml. > >> >> Please check my config is correct. > >> >> > >> >> hostname | hadoop | tajo | DBMS > >> >> namenode | namenode | master | Maria (Metastore) > >> >> snamenode | snamenode+datanode1 | worker > >> >> datanode02 | datanode2 | worker > >> >> datanode03 | datanode3 | worker > >> >> > >> >> > >> >> > >> > ----------------------------------------------------------------------------- > >> >> <configuration> > >> >> <property> > >> >> <name>tajo.rootdir</name> > >> >> <value>hdfs://namenode:9000/tajo</value> > >> >> </property> > >> >> <property> > >> >> <name>tajo.master.umbilical-rpc.address</name> > >> >> <value>namenode:26001</value> > >> >> </property> > >> >> <property> > >> >> <name>tajo.master.client-rpc.address</name> > >> >> <value>namenode:26002</value> > >> >> </property> > >> >> <property> > >> >> <name>tajo.master.info-http.address</name> > >> >> <value>namenode:26080</value> > >> >> </property> > >> >> <property> > >> >> <name>tajo.catalog.client-rpc.address</name> > >> >> <value>namenode:26005</value> > >> >> </property> > >> >> </configuration> > >> >> > >> >> > >> >> Regards, > >> >> Jae > >> >> > >> >> > >> >> 2013/11/11 Jihoon Son <[email protected]> > >> >> > >> >> > Hi Jae Lee, > >> >> > thanks for your interesting to Tajo. > >> >> > > >> >> > Here are my answers. > >> >> > > >> >> > 1. The timeout message looks like an error, but it does not mean > that > >> the > >> >> > query is failed. (We should change the message.) > >> >> > Would you wait for some time after executing a query, please? > >> >> > If any other errors occur, please report it to us. > >> >> > > >> >> > 2. Tajo's SQL commands are designed to follow those of traditional > >> >> > relational databases. > >> >> > In those systems, the 'DROP table' command deletes data from disks. > >> >> > However, we are also considering the Hive-style 'DROP table', > because > >> >> > tables are generally very large. > >> >> > > >> >> > 3. Tajo currently does not provide any commands to kill executing > >> >> queries. > >> >> > Instead, you should kill the master and every worker using the unix > >> >> 'kill' > >> >> > command. > >> >> > > >> >> > If you have any other questions, > >> >> > please feel free to ask us. > >> >> > > >> >> > Thanks, > >> >> > Jihoon > >> >> > > >> >> > > >> >> > 2013/11/11 Jae Lee <[email protected]> > >> >> > > >> >> > > Hello, > >> >> > > > >> >> > > :: I have error message and hang query with below. > >> >> > > It's from clustered tajo worker. > >> >> > > Centos 6.2 + hadoop 2.0.5 + tajo 0.2.0 > >> >> > > Just count(*) query is working but use distinct or group by > query > >> had > >> >> > hang > >> >> > > and this error messages > >> >> > > > >> >> > > :: have more question > >> >> > > Tajo delete files on hdfs when i drop EXTERNAL table. is it > normal? > >> >> > > Because Hive is not delete files when drop external table. > >> >> > > > >> >> > > :: How to can i kill tajo jobs (query)? > >> >> > > > >> >> > > > >> --------------------------------------------------------------------- > >> >> > > 2013-11-11 18:44:22,751 WARN worker.TaskRunner > >> >> > (TaskRunner.java:run(339)) > >> >> > > - Timeout > >> >> > > > >> >> > > > >> >> > > >> >> > >> > GetTask:eb_1384155011466_0005_000001,container_1384155011466_0005_01_000013, > >> >> > > but retry > >> >> > > java.util.concurrent.TimeoutException > >> >> > > at org.apache.tajo.rpc.CallFuture.get(CallFuture.java:81) > >> >> > > at org.apache.tajo.worker.TaskRunner$1.run(TaskRunner.java:328) > >> >> > > at java.lang.Thread.run(Thread.java:744) > >> >> > > > >> >> > > > >> >> > > Regards, > >> >> > > Jae > >> >> > > > >> >> > > >> >> > > >> >> > > >> >> > -- > >> >> > Jihoon Son > >> >> > > >> >> > Database & Information Systems Group, > >> >> > Prof. Yon Dohn Chung Lab. > >> >> > Dept. of Computer Science & Engineering, > >> >> > Korea University > >> >> > 1, 5-ga, Anam-dong, Seongbuk-gu, > >> >> > Seoul, 136-713, Republic of Korea > >> >> > > >> >> > Tel : +82-2-3290-3580 > >> >> > E-mail : [email protected] > >> >> > > >> >> > >> >
