for griffin log, please search in your spark cluster env, usually in worker
log dir.
One weird thing is how you submit a job to spark, if you disabled livy?

On Fri, Sep 11, 2020 at 8:46 PM Sunil Muniyal <[email protected]>
wrote:

> possible to please help the location where measure log would get created
> or from where can i check the location?
>
> Thanks and Regards,
> Sunil Muniyal
>
>
> On Fri, Sep 11, 2020 at 6:14 PM William Guo <[email protected]> wrote:
>
>> Livy is used to post jobs to your cluster, I don't think it is related to
>> livy.
>>
>> Could you also share the measure log in your cluster?
>>
>>
>> On Fri, Sep 11, 2020 at 8:03 PM Sunil Muniyal <[email protected]>
>> wrote:
>>
>>> Got below message as output of
>>>
>>> {"Test_Measure":[{"name":"Test_Job","type":"ACCURACY","owner":"test","metricValues":[]}]}
>>>
>>> metricValues seems empty. So is it like Griffin is not getting data from
>>> ES? whereas ES does have the data which we verified previously. By any
>>> chance, do you think not having Livy could be a problem?
>>>
>>> These are the latest logs from service.out:
>>> *[EL Fine]: sql: 2020-09-11
>>> 11:59:11.662--ServerSession(400064818)--Connection(754936662)--SELECT
>>> DISTINCT ID, APPID, APPURI, CREATEDDATE, DELETED, expire_timestamp,
>>> MODIFIEDDATE, predicate_job_deleted, predicate_group_name,
>>> predicate_job_name, SESSIONID, STATE, timestamp, TYPE, job_id FROM
>>> JOBINSTANCEBEAN WHERE (STATE IN (?,?,?,?,?,?))*
>>> *        bind => [6 parameters bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 11:59:51.044--ServerSession(400064818)--Connection(353930083)--SELECT ID,
>>> type, CREATEDDATE, CRONEXPRESSION, DELETED, quartz_group_name, JOBNAME,
>>> MEASUREID, METRICNAME, MODIFIEDDATE, quartz_job_name, PREDICATECONFIG,
>>> TIMEZONE FROM job WHERE (DELETED = ?)*
>>> *        bind => [1 parameter bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 11:59:51.046--ServerSession(400064818)--Connection(1245663749)--SELECT
>>> DISTINCT DTYPE FROM MEASURE WHERE (DELETED = ?)*
>>> *        bind => [1 parameter bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 11:59:51.046--ServerSession(400064818)--Connection(674248356)--SELECT
>>> t0.ID, t0.DTYPE, t0.CREATEDDATE, t0.DELETED, t0.DESCRIPTION, t0.DQTYPE,
>>> t0.MODIFIEDDATE, t0.NAME, t0.ORGANIZATION, t0.OWNER, t0.SINKS, t1.ID,
>>> t1.PROCESSTYPE, t1.RULEDESCRIPTION, t1.evaluate_rule_id FROM MEASURE t0,
>>> GRIFFINMEASURE t1 WHERE ((t0.DELETED = ?) AND ((t1.ID = t0.ID) AND
>>> (t0.DTYPE = ?)))*
>>> *        bind => [2 parameters bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 12:00:00.019--ClientSession(294162678)--Connection(98503327)--INSERT INTO
>>> JOBINSTANCEBEAN (ID, APPID, APPURI, CREATEDDATE, DELETED, expire_timestamp,
>>> MODIFIEDDATE, predicate_job_deleted, predicate_group_name,
>>> predicate_job_name, SESSIONID, STATE, timestamp, TYPE, job_id) VALUES (?,
>>> ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)*
>>> *        bind => [15 parameters bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 12:00:00.09--ServerSession(400064818)--Connection(491395630)--SELECT ID,
>>> APPID, APPURI, CREATEDDATE, DELETED, expire_timestamp, MODIFIEDDATE,
>>> predicate_job_deleted, predicate_group_name, predicate_job_name, SESSIONID,
>>> STATE, timestamp, TYPE, job_id FROM JOBINSTANCEBEAN WHERE
>>> (predicate_job_name = ?)*
>>> *        bind => [1 parameter bound]*
>>> *2020-09-11 12:00:00.117  INFO 10980 --- [ryBean_Worker-3]
>>> o.a.g.c.j.SparkSubmitJob                 : {*
>>> *  "measure.type" : "griffin",*
>>> *  "id" : 201,*
>>> *  "name" : "Test_Job",*
>>> *  "owner" : "test",*
>>> *  "description" : "Measure to check %age of id field values are same",*
>>> *  "deleted" : false,*
>>> *  "timestamp" : 1599822000000,*
>>> *  "dq.type" : "ACCURACY",*
>>> *  "sinks" : [ "ELASTICSEARCH", "HDFS" ],*
>>> *  "process.type" : "BATCH",*
>>> *  "data.sources" : [ {*
>>> *    "id" : 204,*
>>> *    "name" : "source",*
>>> *    "connectors" : [ {*
>>> *      "id" : 205,*
>>> *      "name" : "source1599568886803",*
>>> *      "type" : "HIVE",*
>>> *      "version" : "1.2",*
>>> *      "predicates" : [ ],*
>>> *      "data.unit" : "1hour",*
>>> *      "data.time.zone" : "",*
>>> *      "config" : {*
>>> *        "database" : "default",*
>>> *        "table.name <http://table.name>" : "demo_src",*
>>> *        "where" : "dt=20200911 AND hour=11"*
>>> *      }*
>>> *    } ],*
>>> *    "baseline" : false*
>>> *  }, {*
>>> *    "id" : 206,*
>>> *    "name" : "target",*
>>> *    "connectors" : [ {*
>>> *      "id" : 207,*
>>> *      "name" : "target1599568896874",*
>>> *      "type" : "HIVE",*
>>> *      "version" : "1.2",*
>>> *      "predicates" : [ ],*
>>> *      "data.unit" : "1hour",*
>>> *      "data.time.zone" : "",*
>>> *      "config" : {*
>>> *        "database" : "default",*
>>> *        "table.name <http://table.name>" : "demo_tgt",*
>>> *        "where" : "dt=20200911 AND hour=11"*
>>> *      }*
>>> *    } ],*
>>> *    "baseline" : false*
>>> *  } ],*
>>> *  "evaluate.rule" : {*
>>> *    "id" : 202,*
>>> *    "rules" : [ {*
>>> *      "id" : 203,*
>>> *      "rule" : "source.id <http://source.id>=target.id
>>> <http://target.id>",*
>>> *      "dsl.type" : "griffin-dsl",*
>>> *      "dq.type" : "ACCURACY",*
>>> *      "out.dataframe.name <http://out.dataframe.name>" : "accuracy"*
>>> *    } ]*
>>> *  },*
>>> *  "measure.type" : "griffin"*
>>> *}*
>>> *2020-09-11 12:00:00.119 ERROR 10980 --- [ryBean_Worker-3]
>>> o.a.g.c.j.SparkSubmitJob                 : Post to livy ERROR. I/O error on
>>> POST request for "http://localhost:8998/batches
>>> <http://localhost:8998/batches>": Connection refused (Connection refused);
>>> nested exception is java.net.ConnectException: Connection refused
>>> (Connection refused)*
>>> *2020-09-11 12:00:00.131  INFO 10980 --- [ryBean_Worker-3]
>>> o.a.g.c.j.SparkSubmitJob                 : Delete predicate
>>> job(PG,Test_Job_predicate_1599825600016) SUCCESS.*
>>> *[EL Fine]: sql: 2020-09-11
>>> 12:00:00.133--ClientSession(273634815)--Connection(296858203)--UPDATE
>>> JOBINSTANCEBEAN SET predicate_job_deleted = ?, STATE = ? WHERE (ID = ?)*
>>> *        bind => [3 parameters bound]*
>>> *[EL Fine]: sql: 2020-09-11
>>> 12:00:11.664--ServerSession(400064818)--Connection(1735064739)--SELECT
>>> DISTINCT ID, APPID, APPURI, CREATEDDATE, DELETED, expire_timestamp,
>>> MODIFIEDDATE, predicate_job_deleted, predicate_group_name,
>>> predicate_job_name, SESSIONID, STATE, timestamp, TYPE, job_id FROM
>>> JOBINSTANCEBEAN WHERE (STATE IN (?,?,?,?,?,?))*
>>> *        bind => [6 parameters bound]*
>>>
>>>
>>> Thanks and Regards,
>>> Sunil Muniyal
>>>
>>>
>>> On Fri, Sep 11, 2020 at 3:42 PM William Guo <[email protected]> wrote:
>>>
>>>> From the log, I didn't find any information related to metrics fetching.
>>>>
>>>> Could you try to call /api/v1/metrics, and show us the latest log again?
>>>>
>>>>
>>>>
>>>> On Fri, Sep 11, 2020 at 5:48 PM Sunil Muniyal <
>>>> [email protected]> wrote:
>>>>
>>>>> 1: I guest it is related to your login user and super user.
>>>>> I am less worried about unless this could be the cause of metrics not
>>>>> being displayed.
>>>>>
>>>>> 2: Could you share with us your griffin log , I suspect some exception
>>>>> happened when trying to connect with ES.
>>>>> Attached is the service.out file. I see an error is while submitting
>>>>> Spark jobs via Livy. Since Livy is not configured / deployed this is
>>>>> expected. I believe this should not be the reason since we are getting 
>>>>> data
>>>>> from hive (as part of batch processing). Please correct if my 
>>>>> understanding
>>>>> is incorrect.
>>>>>
>>>>> Thanks and Regards,
>>>>> Sunil Muniyal
>>>>>
>>>>>
>>>>> On Fri, Sep 11, 2020 at 3:09 PM William Guo <[email protected]> wrote:
>>>>>
>>>>>> 1: I guest it is related to your login user and super user.
>>>>>> 2: Could you share with us your griffin log , I suspect some
>>>>>> exception happened when try to connect with ES.
>>>>>>
>>>>>> On Fri, Sep 11, 2020 at 5:14 PM Sunil Muniyal <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> Hello William,
>>>>>>>
>>>>>>> Tried as suggested.
>>>>>>>
>>>>>>> 1. Ingested data into Hive tables using the provided script.
>>>>>>> The ownership still show as is (Source with Admin and Target with
>>>>>>> Root)
>>>>>>>
>>>>>>> 2. Updated env-batch.json and env-streaming.json files with IP
>>>>>>> address for ES and rebuilt Griffin.
>>>>>>> Still no metrics for the jobs executed.
>>>>>>> ES does have data as confirmed yesterday.
>>>>>>>
>>>>>>> Please help.
>>>>>>>
>>>>>>> Thanks and Regards,
>>>>>>> Sunil Muniyal
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Sep 10, 2020 at 7:41 PM William Guo <[email protected]>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> please enter ip directly.
>>>>>>>> not sure whether hostname can be resolved correctly or not.
>>>>>>>>
>>>>>>>> On Thu, Sep 10, 2020 at 10:06 PM Sunil Muniyal <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Hi William,
>>>>>>>>>
>>>>>>>>> Thank you for the reply.
>>>>>>>>>
>>>>>>>>> Regarding points 2 and 3. Possible to share some more details. I
>>>>>>>>> believe the env_batch.json is configured as it is expected. What 
>>>>>>>>> exactly
>>>>>>>>> needs to be updated correctly? ES Hostname or shall I enter IP or 
>>>>>>>>> something
>>>>>>>>> else? Please help.
>>>>>>>>>
>>>>>>>>> Thanks and Regards,
>>>>>>>>> Sunil Muniyal
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Thu, Sep 10, 2020 at 7:30 PM William Guo <[email protected]>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> 1 OK, We will fix this issue soon.
>>>>>>>>>> 2 Could you try ping es from your spark environment and input ES
>>>>>>>>>> endpoint correctly in env_batch.json
>>>>>>>>>> 3 Please put your es endpoint in env_batch.json
>>>>>>>>>> 6 Please try the following script to build your env.
>>>>>>>>>> ```
>>>>>>>>>>
>>>>>>>>>> #!/bin/bash
>>>>>>>>>> #create table
>>>>>>>>>> hive -f create-table.hqlecho "create table done"
>>>>>>>>>> #current hoursudo ./gen_demo_data.shcur_date=`date 
>>>>>>>>>> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
>>>>>>>>>> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>>>>>>>>>> insert-data.hql
>>>>>>>>>> hive -f 
>>>>>>>>>> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>>>>>>>>>> hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>>>>>>>>>> hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>>>>>>>>>> hadoop fs -touchz ${src_done_path}
>>>>>>>>>> hadoop fs -touchz ${tgt_done_path}echo "insert data 
>>>>>>>>>> [$partition_date] done"
>>>>>>>>>> #last hoursudo ./gen_demo_data.shcur_date=`date -d '1 hour ago' 
>>>>>>>>>> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
>>>>>>>>>> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>>>>>>>>>> insert-data.hql
>>>>>>>>>> hive -f 
>>>>>>>>>> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>>>>>>>>>> hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>>>>>>>>>> hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>>>>>>>>>> hadoop fs -touchz ${src_done_path}
>>>>>>>>>> hadoop fs -touchz ${tgt_done_path}echo "insert data 
>>>>>>>>>> [$partition_date] done"
>>>>>>>>>> #next hoursset +ewhile truedo
>>>>>>>>>>   sudo ./gen_demo_data.sh
>>>>>>>>>>   cur_date=`date +%Y%m%d%H`
>>>>>>>>>>   next_date=`date -d "+1hour" '+%Y%m%d%H'`
>>>>>>>>>>   dt=${next_date:0:8}
>>>>>>>>>>   hour=${next_date:8:2}
>>>>>>>>>>   partition_date="dt='$dt',hour='$hour'"
>>>>>>>>>>   sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>>>>>>>>>> insert-data.hql
>>>>>>>>>>   hive -f insert-data.hql
>>>>>>>>>>   
>>>>>>>>>> src_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONE
>>>>>>>>>>   
>>>>>>>>>> tgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>>>>>>>>>>   hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>>>>>>>>>>   hadoop fs -mkdir -p 
>>>>>>>>>> /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>>>>>>>>>>   hadoop fs -touchz ${src_done_path}
>>>>>>>>>>   hadoop fs -touchz ${tgt_done_path}
>>>>>>>>>>   echo "insert data [$partition_date] done"
>>>>>>>>>>   sleep 3600doneset -e
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>>
>>>>>>>>>> William
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Thu, Sep 10, 2020 at 4:58 PM Sunil Muniyal <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does
>>>>>>>>>>> it mean that only ES upto 6.8.x is supported for Griffin as of now? 
>>>>>>>>>>> If yes,
>>>>>>>>>>> what are the plans further? Is there a page from which I could get 
>>>>>>>>>>> updates?
>>>>>>>>>>> --please file a jira ticket for us to make our code ES
>>>>>>>>>>> compatible.
>>>>>>>>>>> [SM] GRIFFIN-346 - Support for Elastic Search latest version
>>>>>>>>>>> (7.9.1) <https://issues.apache.org/jira/browse/GRIFFIN-346> is
>>>>>>>>>>> submitted
>>>>>>>>>>>
>>>>>>>>>>> 2. I still do not see the metrics available (please refer below
>>>>>>>>>>> screenshots). Though the measure is now listed in the drop down of 
>>>>>>>>>>> *DQ
>>>>>>>>>>> Metrics* tab. But when I selected the test measure, nothing
>>>>>>>>>>> came up.
>>>>>>>>>>> --could you check the ES whether metrics have been injected or
>>>>>>>>>>> not.
>>>>>>>>>>> [SM] I used the link below and got the index that is created in
>>>>>>>>>>> ES. I believe the data is loaded. However, please correct if I
>>>>>>>>>>> understood incorrectly
>>>>>>>>>>> *"http://<ES Public IP>:9200/_cat/indices?v"*
>>>>>>>>>>> --------------> POC env is on public cloud so using Public IP.
>>>>>>>>>>>
>>>>>>>>>>> health status index   uuid                   pri rep docs.count 
>>>>>>>>>>> docs.deleted store.size pri.store.size
>>>>>>>>>>> yellow open   griffin ur_Kd3XFQBCsPzIM84j87Q   5   2          0     
>>>>>>>>>>>        0      1.2kb          1.2kb
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Docs in the index:* "http://<ES Public
>>>>>>>>>>> IP>:9200/griffin/_search"*
>>>>>>>>>>>
>>>>>>>>>>> {"took":44,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}}
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Index Mapping: *"http://<ES Public IP>:9200/griffin"*
>>>>>>>>>>>
>>>>>>>>>>> {"griffin":{"aliases":{},"mappings":{"accuracy":{"properties":{"name":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"tmst":{"type":"date"}}}},"settings":{"index":{"creation_date":"1599567930578","number_of_shards":"5","number_of_replicas":"2","uuid":"ur_Kd3XFQBCsPzIM84j87Q","version":{"created":"6081299"},"provided_name":"griffin"}}}}
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 3. At a step in deployment guide it is suggested to check URL: 
>>>>>>>>>>> "*http://<ES
>>>>>>>>>>> HOST IP>:9200/griffin/accuracy"* When navigated to this URL, I
>>>>>>>>>>> get below error. Please advise
>>>>>>>>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and
>>>>>>>>>>> method [GET], allowed: [POST]","status":405}*
>>>>>>>>>>> *-- it seems you need to use POST method.*
>>>>>>>>>>> [SM] I am using the POST method as suggested in the article.
>>>>>>>>>>> Below is the JSON of *env_batch.JSON*
>>>>>>>>>>> *    {*
>>>>>>>>>>> *      "type": "ELASTICSEARCH",*
>>>>>>>>>>> *      "config": {*
>>>>>>>>>>> *        "method": "post",*
>>>>>>>>>>> *        "api": "http://<ES Host Name>:9200/griffin/accuracy",
>>>>>>>>>>> ---------> *do we need IP here?
>>>>>>>>>>> *        "connection.timeout": "1m",*
>>>>>>>>>>> *        "retry": 10*
>>>>>>>>>>> *      }*
>>>>>>>>>>> *    }*
>>>>>>>>>>>
>>>>>>>>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by
>>>>>>>>>>> Admin whereas, *demo-tgt* by root. Would that make any
>>>>>>>>>>> difference? If yes, how to correct it? Reload HIVE data?
>>>>>>>>>>> -- could you show me your script for dataset setup?
>>>>>>>>>>> <https://issues.apache.org/jira/browse/GRIFFIN-346>
>>>>>>>>>>> [SM] Attached are the 3 scripts. gen-hive-data.sh is the master
>>>>>>>>>>> script which triggers demo_data and it further triggers delta_src.
>>>>>>>>>>> Have done it as it is instructed in the Github article and
>>>>>>>>>>> gen-hive-data.sh is triggered as root in the terminal.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Please advise.
>>>>>>>>>>>
>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Sep 9, 2020 at 8:41 PM William Guo <[email protected]>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> *Request you to please advise further on below points:*
>>>>>>>>>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does
>>>>>>>>>>>> it mean that only ES upto 6.8.x is supported for Griffin as of 
>>>>>>>>>>>> now? If yes,
>>>>>>>>>>>> what are the plans further? Is there a page from which I could get 
>>>>>>>>>>>> updates?
>>>>>>>>>>>> --please file a jira ticket for us to make our code ES
>>>>>>>>>>>> compatible.
>>>>>>>>>>>>
>>>>>>>>>>>> 2. I still do not see the metrics available (please refer below
>>>>>>>>>>>> screenshots). Though the measure is now listed in the drop down of 
>>>>>>>>>>>> *DQ
>>>>>>>>>>>> Metrics* tab. But when I selected the test measure, nothing
>>>>>>>>>>>> came up.
>>>>>>>>>>>> --could you check the ES whether metrics have been injected or
>>>>>>>>>>>> not.
>>>>>>>>>>>>
>>>>>>>>>>>> 3. At a step in deployment guide it is suggested to check URL: 
>>>>>>>>>>>> http://<ES
>>>>>>>>>>>> HOST IP>:9200/griffin/accuracy
>>>>>>>>>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated
>>>>>>>>>>>> to this URL, I get below error. Please advise
>>>>>>>>>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy]
>>>>>>>>>>>> and method [GET], allowed: [POST]","status":405}*
>>>>>>>>>>>> *-- it seems you need to use POST method.*
>>>>>>>>>>>>
>>>>>>>>>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by
>>>>>>>>>>>> Admin whereas, *demo-tgt* by root. Would that make any
>>>>>>>>>>>> difference? If yes, how to correct it? Reload HIVE data?
>>>>>>>>>>>>
>>>>>>>>>>>> -- could you show me your script for dataset setup?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Sep 8, 2020 at 9:02 PM Sunil Muniyal <
>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi William,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I was finally able to get Griffin up and ElasticSearch
>>>>>>>>>>>>> integrated along with Hadoop. Thanks a lot for your help and 
>>>>>>>>>>>>> guidance so
>>>>>>>>>>>>> far.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have created a test measure and a job which gets triggered
>>>>>>>>>>>>> at every 4 mins automatically (have referred to the user guide 
>>>>>>>>>>>>> available on
>>>>>>>>>>>>> GitHub at this link
>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md>
>>>>>>>>>>>>> .)
>>>>>>>>>>>>>
>>>>>>>>>>>>> *Request you to please advise further on below points:*
>>>>>>>>>>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated,
>>>>>>>>>>>>> does it mean that only ES upto 6.8.x is supported for Griffin as 
>>>>>>>>>>>>> of now? If
>>>>>>>>>>>>> yes, what are the plans further? Is there a page from which I 
>>>>>>>>>>>>> could get
>>>>>>>>>>>>> updates?
>>>>>>>>>>>>> 2. I still do not see the metrics available (please refer
>>>>>>>>>>>>> below screenshots). Though the measure is now listed in the drop 
>>>>>>>>>>>>> down of *DQ
>>>>>>>>>>>>> Metrics* tab. But when I selected the test measure, nothing
>>>>>>>>>>>>> came up.
>>>>>>>>>>>>> 3. At a step in deployment guide it is suggested to check URL: 
>>>>>>>>>>>>> http://<ES
>>>>>>>>>>>>> HOST IP>:9200/griffin/accuracy
>>>>>>>>>>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated
>>>>>>>>>>>>> to this URL, I get below error. Please advise
>>>>>>>>>>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy]
>>>>>>>>>>>>> and method [GET], allowed: [POST]","status":405}*
>>>>>>>>>>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by
>>>>>>>>>>>>> Admin whereas, *demo-tgt* by root. Would that make any
>>>>>>>>>>>>> difference? If yes, how to correct it? Reload HIVE data?
>>>>>>>>>>>>>
>>>>>>>>>>>>> *Screenshots:*
>>>>>>>>>>>>> *Data Assets:*
>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>
>>>>>>>>>>>>> *DQ Metrics (Test Measure selected):*
>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>
>>>>>>>>>>>>> *Job Triggered multiple times:*
>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>
>>>>>>>>>>>>> *Metrics page from job directly:*
>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Sep 8, 2020 at 4:38 PM Sunil Muniyal <
>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> I am unable to get repos for 6.4.1 instead I found 6.8.x.
>>>>>>>>>>>>>> Will try with this version of Elastic Search in sometime.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> In the meantime, would it be possible to confirm if 6.4.x or
>>>>>>>>>>>>>> 6.8.x is the only supported version for Griffin? Reason I am 
>>>>>>>>>>>>>> asking is, the
>>>>>>>>>>>>>> GitHub article for griffin deployment points to the latest 
>>>>>>>>>>>>>> version of ES.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Sep 8, 2020 at 4:06 PM Sunil Muniyal <
>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I will need to redeploy ElasticSearch, correct?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Sep 8, 2020 at 4:05 PM William Guo <[email protected]>
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Could you try with this version?
>>>>>>>>>>>>>>>> <elasticsearch.version>6.4.1</elasticsearch.version>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Sep 8, 2020 at 5:59 PM Sunil Muniyal <
>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi William / Dev group,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I have deployed ES 7.9 - latest version (single node) and
>>>>>>>>>>>>>>>>> the same is configured. I also get the default page when 
>>>>>>>>>>>>>>>>> hitting
>>>>>>>>>>>>>>>>> http://<ES HOST IP>:9200/
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Upon creating the griffin configurations using the JSON
>>>>>>>>>>>>>>>>> string given
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> curl -k -H "Content-Type: application/json" -X PUT 
>>>>>>>>>>>>>>>>> http://<replaced with my ES host IP>:9200/griffin \
>>>>>>>>>>>>>>>>>  -d '{
>>>>>>>>>>>>>>>>>     "aliases": {},
>>>>>>>>>>>>>>>>>     "mappings": {
>>>>>>>>>>>>>>>>>         "accuracy": {
>>>>>>>>>>>>>>>>>             "properties": {
>>>>>>>>>>>>>>>>>                 "name": {
>>>>>>>>>>>>>>>>>                     "fields": {
>>>>>>>>>>>>>>>>>                         "keyword": {
>>>>>>>>>>>>>>>>>                             "ignore_above": 256,
>>>>>>>>>>>>>>>>>                             "type": "keyword"
>>>>>>>>>>>>>>>>>                         }
>>>>>>>>>>>>>>>>>                     },
>>>>>>>>>>>>>>>>>                     "type": "text"
>>>>>>>>>>>>>>>>>                 },
>>>>>>>>>>>>>>>>>                 "tmst": {
>>>>>>>>>>>>>>>>>                     "type": "date"
>>>>>>>>>>>>>>>>>                 }
>>>>>>>>>>>>>>>>>             }
>>>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>>>     },
>>>>>>>>>>>>>>>>>     "settings": {
>>>>>>>>>>>>>>>>>         "index": {
>>>>>>>>>>>>>>>>>             "number_of_replicas": "2",
>>>>>>>>>>>>>>>>>             "number_of_shards": "5"
>>>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>>> }'
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *I get below error:*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *{"error":{"root_cause":[{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>>>>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>>>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, 
>>>>>>>>>>>>>>>>> type=keyword}},
>>>>>>>>>>>>>>>>> type=text},
>>>>>>>>>>>>>>>>> tmst={type=date}}}]"}],"type":"mapper_parsing_exception","reason":"Failed
>>>>>>>>>>>>>>>>> to parse mapping [_doc]: Root mapping definition has 
>>>>>>>>>>>>>>>>> unsupported
>>>>>>>>>>>>>>>>> parameters:  [accuracy :
>>>>>>>>>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, 
>>>>>>>>>>>>>>>>> type=keyword}},
>>>>>>>>>>>>>>>>> type=text},
>>>>>>>>>>>>>>>>> tmst={type=date}}}]","caused_by":{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>>>>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>>>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, 
>>>>>>>>>>>>>>>>> type=keyword}},
>>>>>>>>>>>>>>>>> type=text}, tmst={type=date}}}]"}},"status":400}*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Seems like the JSON string is missing some values or is
>>>>>>>>>>>>>>>>> incorrectly provided.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Would be great if you could please help.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 8:16 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thank you for the response, William.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I have started preparing for ES deployment and should
>>>>>>>>>>>>>>>>>> attempt the same tomorrow.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> In the meantime, I will also wait for the Dev team in
>>>>>>>>>>>>>>>>>> case they have any additional inputs.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 8:06 PM William Guo <
>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> If dev confirms it to be mandatory, as I understand
>>>>>>>>>>>>>>>>>>> correct, I will need to:
>>>>>>>>>>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>>>>>>>>>>> 2. Update application.properties to include ES details
>>>>>>>>>>>>>>>>>>> and create ES index
>>>>>>>>>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> *Right, you need to package es env configuration into
>>>>>>>>>>>>>>>>>>> your jar.*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive),
>>>>>>>>>>>>>>>>>>> correct?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> *No*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On a side note, is there any other documentation of
>>>>>>>>>>>>>>>>>>> Griffin available or underway which would help to get below 
>>>>>>>>>>>>>>>>>>> details while
>>>>>>>>>>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>>>>>>>>>>> 1. What are the exact ports requirements (internal and
>>>>>>>>>>>>>>>>>>> external)?
>>>>>>>>>>>>>>>>>>> *check log and make sure all extra connections in
>>>>>>>>>>>>>>>>>>> properties can accessible*
>>>>>>>>>>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>>>>>>>>>>> *no*
>>>>>>>>>>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>>>>>>>>>>> *java 1.8*
>>>>>>>>>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized
>>>>>>>>>>>>>>>>>>> (secured), what are the dependencies or additional 
>>>>>>>>>>>>>>>>>>> configurations needed?
>>>>>>>>>>>>>>>>>>> *Should no extra dependencies, except those transitive
>>>>>>>>>>>>>>>>>>> dependencies incurred by spark and hadoop.*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:42 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Ohh ok.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> If dev confirms it to be mandatory, as I understand
>>>>>>>>>>>>>>>>>>>> correct, I will need to:
>>>>>>>>>>>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>>>>>>>>>>>> 2. Update application.properties to include ES details
>>>>>>>>>>>>>>>>>>>> and create ES index
>>>>>>>>>>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive),
>>>>>>>>>>>>>>>>>>>> correct?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On a side note, is there any other documentation of
>>>>>>>>>>>>>>>>>>>> Griffin available or underway which would help to get 
>>>>>>>>>>>>>>>>>>>> below details while
>>>>>>>>>>>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>>>>>>>>>>>> 1. What are the exact ports requirements (internal and
>>>>>>>>>>>>>>>>>>>> external)?
>>>>>>>>>>>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>>>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized
>>>>>>>>>>>>>>>>>>>> (secured), what are the dependencies or additional 
>>>>>>>>>>>>>>>>>>>> configurations needed?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I know some of the above information can be fetched
>>>>>>>>>>>>>>>>>>>> from the deployment guide on Github. However, checking if 
>>>>>>>>>>>>>>>>>>>> any other formal
>>>>>>>>>>>>>>>>>>>> documentation has been made available for the same?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 4:05 PM William Guo <
>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> cc dev for double checking.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Measure will emit metrics and store them in elastic,
>>>>>>>>>>>>>>>>>>>>> UI fetch those metrics from elastic.
>>>>>>>>>>>>>>>>>>>>> So elastic should be mandatory.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:32 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Thank you for the quick response, William.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> I have not configured ElasticSearch since it is not
>>>>>>>>>>>>>>>>>>>>>> deployed.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> In the application.properties, I just added the dummy
>>>>>>>>>>>>>>>>>>>>>> information (as below) just to pass the validation test 
>>>>>>>>>>>>>>>>>>>>>> and get Griffin up
>>>>>>>>>>>>>>>>>>>>>> and running.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> # elasticsearch
>>>>>>>>>>>>>>>>>>>>>> # elasticsearch.host = <IP>
>>>>>>>>>>>>>>>>>>>>>> # elasticsearch.port = <elasticsearch rest port>
>>>>>>>>>>>>>>>>>>>>>> # elasticsearch.user = user
>>>>>>>>>>>>>>>>>>>>>> # elasticsearch.password = password
>>>>>>>>>>>>>>>>>>>>>> elasticsearch.host=localhost
>>>>>>>>>>>>>>>>>>>>>> elasticsearch.port=9200
>>>>>>>>>>>>>>>>>>>>>> elasticsearch.scheme=http
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Is ElasticSearch a mandatory requirement to use
>>>>>>>>>>>>>>>>>>>>>> Griffin?
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 3:58 PM William Guo <
>>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Could you check whether ES has been injected with
>>>>>>>>>>>>>>>>>>>>>>> those metrics or not?
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:23 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> I was able to bypass this error by entering the
>>>>>>>>>>>>>>>>>>>>>>>> default field values for LDAP, ElasticSearch and Livy 
>>>>>>>>>>>>>>>>>>>>>>>> in
>>>>>>>>>>>>>>>>>>>>>>>> application.properties and successfully get Griffin 
>>>>>>>>>>>>>>>>>>>>>>>> running.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> By following the below article, I have created a
>>>>>>>>>>>>>>>>>>>>>>>> test measure and then a job which triggers that 
>>>>>>>>>>>>>>>>>>>>>>>> measure.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Have allowed the job to get triggered multiple
>>>>>>>>>>>>>>>>>>>>>>>> times, however, still i can't see anything in metrics 
>>>>>>>>>>>>>>>>>>>>>>>> related to the job.
>>>>>>>>>>>>>>>>>>>>>>>> Neither I see anything in *health *or *mydashboard* 
>>>>>>>>>>>>>>>>>>>>>>>> tabs.
>>>>>>>>>>>>>>>>>>>>>>>> Also, if you notice in the screenshot below, being in 
>>>>>>>>>>>>>>>>>>>>>>>> the *DQ
>>>>>>>>>>>>>>>>>>>>>>>> Metrics* tab, I still do not see the created
>>>>>>>>>>>>>>>>>>>>>>>> measure in the drop down list.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> *Test job executed multiple times:*
>>>>>>>>>>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Please advise if anything is mis-configured.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 12:40 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Thank you for the reply.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> This helped, actually i had missed to add the
>>>>>>>>>>>>>>>>>>>>>>>>> property in application.properties.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Now the other challenge is, along with ES and
>>>>>>>>>>>>>>>>>>>>>>>>> Livy, I am also not using LDAP and it is hitting the 
>>>>>>>>>>>>>>>>>>>>>>>>> error *unable
>>>>>>>>>>>>>>>>>>>>>>>>> to resolve ldap.url property.* Of Course it will,
>>>>>>>>>>>>>>>>>>>>>>>>> since the property is not configured.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Please suggest.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> On Sun, Sep 6, 2020 at 7:26 PM William Guo <
>>>>>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> hi Sunil Muniyal,
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Could you check this property in your griffin
>>>>>>>>>>>>>>>>>>>>>>>>>> properties file?
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> internal.event.listeners
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> On Thu, Sep 3, 2020 at 11:05 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> I am attempting to integrate Griffin with
>>>>>>>>>>>>>>>>>>>>>>>>>>> Cloudera Hadoop by following below article:
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md>I
>>>>>>>>>>>>>>>>>>>>>>>>>>> have followed everything as instructed, apart from 
>>>>>>>>>>>>>>>>>>>>>>>>>>> below things:
>>>>>>>>>>>>>>>>>>>>>>>>>>> 1. Using Cloudera Hadoop 5.15 and relevant
>>>>>>>>>>>>>>>>>>>>>>>>>>> configurations instead of Apache Hadoop
>>>>>>>>>>>>>>>>>>>>>>>>>>> 2. Not using Elastic search as it is not
>>>>>>>>>>>>>>>>>>>>>>>>>>> applicable
>>>>>>>>>>>>>>>>>>>>>>>>>>> 3. Did not use Livy as it is not applicable.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Maven build is successful and has got 2 jars at
>>>>>>>>>>>>>>>>>>>>>>>>>>> service/target and measure/target which I have 
>>>>>>>>>>>>>>>>>>>>>>>>>>> uploaded to HDFS.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> However, *starting griffin-service.jar using
>>>>>>>>>>>>>>>>>>>>>>>>>>> nohup command* is failing with below error:
>>>>>>>>>>>>>>>>>>>>>>>>>>> *Caused by: java.lang.IllegalArgumentException:
>>>>>>>>>>>>>>>>>>>>>>>>>>> Could not resolve placeholder 
>>>>>>>>>>>>>>>>>>>>>>>>>>> 'internal.event.listeners' in string value
>>>>>>>>>>>>>>>>>>>>>>>>>>> "#{'${internal.event.listeners}'.split(',')}"*
>>>>>>>>>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.parseStringValue(PropertyPlaceholderHelper.java:174)
>>>>>>>>>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.replacePlaceholders(PropertyPlaceholderHelper.java:126)
>>>>>>>>>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>>>>>>>>>> org.springframework.core.env.AbstractPropertyResolver.doResolvePlaceholders(AbstractPropertyResolver.java:236)
>>>>>>>>>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> I have tried to search a lot of articles with no
>>>>>>>>>>>>>>>>>>>>>>>>>>> luck.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Would be great if someone could help me to fix
>>>>>>>>>>>>>>>>>>>>>>>>>>> this.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Also, attached is the output of nohup command
>>>>>>>>>>>>>>>>>>>>>>>>>>> that was written in service.out.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>> To unsubscribe, e-mail:
>>>>>>>>>>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>>>>>>>>>>> For additional commands, e-mail:
>>>>>>>>>>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>> To unsubscribe, e-mail: [email protected]
>>>>>>>>>>> For additional commands, e-mail: [email protected]
>>>>>>>>>>
>>>>>>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: [email protected]
>>>>> For additional commands, e-mail: [email protected]
>>>>
>>>>

Reply via email to