please enter ip directly.
not sure whether hostname can be resolved correctly or not.

On Thu, Sep 10, 2020 at 10:06 PM Sunil Muniyal <[email protected]>
wrote:

> Hi William,
>
> Thank you for the reply.
>
> Regarding points 2 and 3. Possible to share some more details. I believe
> the env_batch.json is configured as it is expected. What exactly needs to
> be updated correctly? ES Hostname or shall I enter IP or something else?
> Please help.
>
> Thanks and Regards,
> Sunil Muniyal
>
>
> On Thu, Sep 10, 2020 at 7:30 PM William Guo <[email protected]> wrote:
>
>> 1 OK, We will fix this issue soon.
>> 2 Could you try ping es from your spark environment and input ES endpoint
>> correctly in env_batch.json
>> 3 Please put your es endpoint in env_batch.json
>> 6 Please try the following script to build your env.
>> ```
>>
>> #!/bin/bash
>> #create table
>> hive -f create-table.hqlecho "create table done"
>> #current hoursudo ./gen_demo_data.shcur_date=`date 
>> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
>> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>> insert-data.hql
>> hive -f 
>> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>> hadoop fs -touchz ${src_done_path}
>> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
>> #last hoursudo ./gen_demo_data.shcur_date=`date -d '1 hour ago' 
>> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
>> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>> insert-data.hql
>> hive -f 
>> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>> hadoop fs -touchz ${src_done_path}
>> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
>> #next hoursset +ewhile truedo
>>   sudo ./gen_demo_data.sh
>>   cur_date=`date +%Y%m%d%H`
>>   next_date=`date -d "+1hour" '+%Y%m%d%H'`
>>   dt=${next_date:0:8}
>>   hour=${next_date:8:2}
>>   partition_date="dt='$dt',hour='$hour'"
>>   sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
>> insert-data.hql
>>   hive -f insert-data.hql
>>   src_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONE
>>   tgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>>   hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>>   hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>>   hadoop fs -touchz ${src_done_path}
>>   hadoop fs -touchz ${tgt_done_path}
>>   echo "insert data [$partition_date] done"
>>   sleep 3600doneset -e
>>
>>
>>
>> Thanks,
>>
>> William
>>
>>
>> On Thu, Sep 10, 2020 at 4:58 PM Sunil Muniyal <[email protected]>
>> wrote:
>>
>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>>> are the plans further? Is there a page from which I could get updates?
>>> --please file a jira ticket for us to make our code ES compatible.
>>> [SM] GRIFFIN-346 - Support for Elastic Search latest version (7.9.1)
>>> <https://issues.apache.org/jira/browse/GRIFFIN-346> is submitted
>>>
>>> 2. I still do not see the metrics available (please refer below
>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>> --could you check the ES whether metrics have been injected or not.
>>> [SM] I used the link below and got the index that is created in ES. I
>>> believe the data is loaded. However, please correct if I
>>> understood incorrectly
>>> *"http://<ES Public IP>:9200/_cat/indices?v"*
>>> --------------> POC env is on public cloud so using Public IP.
>>>
>>> health status index   uuid                   pri rep docs.count 
>>> docs.deleted store.size pri.store.size
>>> yellow open   griffin ur_Kd3XFQBCsPzIM84j87Q   5   2          0            
>>> 0      1.2kb          1.2kb
>>>
>>>
>>> Docs in the index:* "http://<ES Public IP>:9200/griffin/_search"*
>>>
>>> {"took":44,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}}
>>>
>>>
>>> Index Mapping: *"http://<ES Public IP>:9200/griffin"*
>>>
>>> {"griffin":{"aliases":{},"mappings":{"accuracy":{"properties":{"name":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"tmst":{"type":"date"}}}},"settings":{"index":{"creation_date":"1599567930578","number_of_shards":"5","number_of_replicas":"2","uuid":"ur_Kd3XFQBCsPzIM84j87Q","version":{"created":"6081299"},"provided_name":"griffin"}}}}
>>>
>>>
>>> 3. At a step in deployment guide it is suggested to check URL: "*http://<ES
>>> HOST IP>:9200/griffin/accuracy"* When navigated to this URL, I get
>>> below error. Please advise
>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>>> [GET], allowed: [POST]","status":405}*
>>> *-- it seems you need to use POST method.*
>>> [SM] I am using the POST method as suggested in the article. Below is
>>> the JSON of *env_batch.JSON*
>>> *    {*
>>> *      "type": "ELASTICSEARCH",*
>>> *      "config": {*
>>> *        "method": "post",*
>>> *        "api": "http://<ES Host Name>:9200/griffin/accuracy",
>>> ---------> *do we need IP here?
>>> *        "connection.timeout": "1m",*
>>> *        "retry": 10*
>>> *      }*
>>> *    }*
>>>
>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>> how to correct it? Reload HIVE data?
>>> -- could you show me your script for dataset setup?
>>> <https://issues.apache.org/jira/browse/GRIFFIN-346>
>>> [SM] Attached are the 3 scripts. gen-hive-data.sh is the master script
>>> which triggers demo_data and it further triggers delta_src.
>>> Have done it as it is instructed in the Github article and
>>> gen-hive-data.sh is triggered as root in the terminal.
>>>
>>>
>>> Please advise.
>>>
>>> Thanks and Regards,
>>> Sunil Muniyal
>>>
>>>
>>> On Wed, Sep 9, 2020 at 8:41 PM William Guo <[email protected]> wrote:
>>>
>>>> *Request you to please advise further on below points:*
>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>>>> are the plans further? Is there a page from which I could get updates?
>>>> --please file a jira ticket for us to make our code ES compatible.
>>>>
>>>> 2. I still do not see the metrics available (please refer below
>>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>>> --could you check the ES whether metrics have been injected or not.
>>>>
>>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>>>> HOST IP>:9200/griffin/accuracy
>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>>>> URL, I get below error. Please advise
>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>>>> [GET], allowed: [POST]","status":405}*
>>>> *-- it seems you need to use POST method.*
>>>>
>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>>> how to correct it? Reload HIVE data?
>>>>
>>>> -- could you show me your script for dataset setup?
>>>>
>>>>
>>>> On Tue, Sep 8, 2020 at 9:02 PM Sunil Muniyal <
>>>> [email protected]> wrote:
>>>>
>>>>> Hi William,
>>>>>
>>>>> I was finally able to get Griffin up and ElasticSearch integrated
>>>>> along with Hadoop. Thanks a lot for your help and guidance so far.
>>>>>
>>>>> I have created a test measure and a job which gets triggered at every
>>>>> 4 mins automatically (have referred to the user guide available on GitHub
>>>>> at this link
>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md>
>>>>> .)
>>>>>
>>>>> *Request you to please advise further on below points:*
>>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it
>>>>> mean that only ES upto 6.8.x is supported for Griffin as of now? If yes,
>>>>> what are the plans further? Is there a page from which I could get 
>>>>> updates?
>>>>> 2. I still do not see the metrics available (please refer below
>>>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>>>>> HOST IP>:9200/griffin/accuracy
>>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>>>>> URL, I get below error. Please advise
>>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and
>>>>> method [GET], allowed: [POST]","status":405}*
>>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>>>> how to correct it? Reload HIVE data?
>>>>>
>>>>> *Screenshots:*
>>>>> *Data Assets:*
>>>>> [image: image.png]
>>>>>
>>>>> *DQ Metrics (Test Measure selected):*
>>>>> [image: image.png]
>>>>>
>>>>> *Job Triggered multiple times:*
>>>>> [image: image.png]
>>>>>
>>>>> *Metrics page from job directly:*
>>>>> [image: image.png]
>>>>>
>>>>> Thanks and Regards,
>>>>> Sunil Muniyal
>>>>>
>>>>>
>>>>> On Tue, Sep 8, 2020 at 4:38 PM Sunil Muniyal <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> I am unable to get repos for 6.4.1 instead I found 6.8.x. Will try
>>>>>> with this version of Elastic Search in sometime.
>>>>>>
>>>>>> In the meantime, would it be possible to confirm if 6.4.x or 6.8.x is
>>>>>> the only supported version for Griffin? Reason I am asking is, the GitHub
>>>>>> article for griffin deployment points to the latest version of ES.
>>>>>>
>>>>>> Thanks and Regards,
>>>>>> Sunil Muniyal
>>>>>>
>>>>>>
>>>>>> On Tue, Sep 8, 2020 at 4:06 PM Sunil Muniyal <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> I will need to redeploy ElasticSearch, correct?
>>>>>>>
>>>>>>> Thanks and Regards,
>>>>>>> Sunil Muniyal
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Sep 8, 2020 at 4:05 PM William Guo <[email protected]> wrote:
>>>>>>>
>>>>>>>> Could you try with this version?
>>>>>>>> <elasticsearch.version>6.4.1</elasticsearch.version>
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> William
>>>>>>>>
>>>>>>>> On Tue, Sep 8, 2020 at 5:59 PM Sunil Muniyal <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Hi William / Dev group,
>>>>>>>>>
>>>>>>>>> I have deployed ES 7.9 - latest version (single node) and the same
>>>>>>>>> is configured. I also get the default page when hitting
>>>>>>>>> http://<ES HOST IP>:9200/
>>>>>>>>>
>>>>>>>>> Upon creating the griffin configurations using the JSON string
>>>>>>>>> given
>>>>>>>>>
>>>>>>>>> curl -k -H "Content-Type: application/json" -X PUT http://<replaced 
>>>>>>>>> with my ES host IP>:9200/griffin \
>>>>>>>>>  -d '{
>>>>>>>>>     "aliases": {},
>>>>>>>>>     "mappings": {
>>>>>>>>>         "accuracy": {
>>>>>>>>>             "properties": {
>>>>>>>>>                 "name": {
>>>>>>>>>                     "fields": {
>>>>>>>>>                         "keyword": {
>>>>>>>>>                             "ignore_above": 256,
>>>>>>>>>                             "type": "keyword"
>>>>>>>>>                         }
>>>>>>>>>                     },
>>>>>>>>>                     "type": "text"
>>>>>>>>>                 },
>>>>>>>>>                 "tmst": {
>>>>>>>>>                     "type": "date"
>>>>>>>>>                 }
>>>>>>>>>             }
>>>>>>>>>         }
>>>>>>>>>     },
>>>>>>>>>     "settings": {
>>>>>>>>>         "index": {
>>>>>>>>>             "number_of_replicas": "2",
>>>>>>>>>             "number_of_shards": "5"
>>>>>>>>>         }
>>>>>>>>>     }
>>>>>>>>> }'
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> *I get below error:*
>>>>>>>>>
>>>>>>>>> *{"error":{"root_cause":[{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>>> type=text},
>>>>>>>>> tmst={type=date}}}]"}],"type":"mapper_parsing_exception","reason":"Failed
>>>>>>>>> to parse mapping [_doc]: Root mapping definition has unsupported
>>>>>>>>> parameters:  [accuracy :
>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>>> type=text},
>>>>>>>>> tmst={type=date}}}]","caused_by":{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>>> type=text}, tmst={type=date}}}]"}},"status":400}*
>>>>>>>>>
>>>>>>>>> Seems like the JSON string is missing some values or is
>>>>>>>>> incorrectly provided.
>>>>>>>>>
>>>>>>>>> Would be great if you could please help.
>>>>>>>>>
>>>>>>>>> Thanks and Regards,
>>>>>>>>> Sunil Muniyal
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Sep 7, 2020 at 8:16 PM Sunil Muniyal <
>>>>>>>>> [email protected]> wrote:
>>>>>>>>>
>>>>>>>>>> Thank you for the response, William.
>>>>>>>>>>
>>>>>>>>>> I have started preparing for ES deployment and should attempt the
>>>>>>>>>> same tomorrow.
>>>>>>>>>>
>>>>>>>>>> In the meantime, I will also wait for the Dev team in case they
>>>>>>>>>> have any additional inputs.
>>>>>>>>>>
>>>>>>>>>> Thanks and Regards,
>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Sep 7, 2020 at 8:06 PM William Guo <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>>>> will need to:
>>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>>> 2. Update application.properties to include ES details and
>>>>>>>>>>> create ES index
>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>>
>>>>>>>>>>> *Right, you need to package es env configuration into your jar.*
>>>>>>>>>>>
>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>>>
>>>>>>>>>>> *No*
>>>>>>>>>>>
>>>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>>> 1. What are the exact ports requirements (internal and external)?
>>>>>>>>>>> *check log and make sure all extra connections in properties can
>>>>>>>>>>> accessible*
>>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>>> *no*
>>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>>> *java 1.8*
>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what
>>>>>>>>>>> are the dependencies or additional configurations needed?
>>>>>>>>>>> *Should no extra dependencies, except those transitive
>>>>>>>>>>> dependencies incurred by spark and hadoop.*
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:42 PM Sunil Muniyal <
>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Ohh ok.
>>>>>>>>>>>>
>>>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>>>>> will need to:
>>>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>>>> 2. Update application.properties to include ES details and
>>>>>>>>>>>> create ES index
>>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>>>
>>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>>>>
>>>>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>>>> 1. What are the exact ports requirements (internal and
>>>>>>>>>>>> external)?
>>>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured),
>>>>>>>>>>>> what are the dependencies or additional configurations needed?
>>>>>>>>>>>>
>>>>>>>>>>>> I know some of the above information can be fetched from the
>>>>>>>>>>>> deployment guide on Github. However, checking if any other formal
>>>>>>>>>>>> documentation has been made available for the same?
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Sep 7, 2020 at 4:05 PM William Guo <[email protected]>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> cc dev for double checking.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Measure will emit metrics and store them in elastic, UI fetch
>>>>>>>>>>>>> those metrics from elastic.
>>>>>>>>>>>>> So elastic should be mandatory.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> William
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:32 PM Sunil Muniyal <
>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thank you for the quick response, William.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I have not configured ElasticSearch since it is not deployed.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> In the application.properties, I just added the dummy
>>>>>>>>>>>>>> information (as below) just to pass the validation test and get 
>>>>>>>>>>>>>> Griffin up
>>>>>>>>>>>>>> and running.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> # elasticsearch
>>>>>>>>>>>>>> # elasticsearch.host = <IP>
>>>>>>>>>>>>>> # elasticsearch.port = <elasticsearch rest port>
>>>>>>>>>>>>>> # elasticsearch.user = user
>>>>>>>>>>>>>> # elasticsearch.password = password
>>>>>>>>>>>>>> elasticsearch.host=localhost
>>>>>>>>>>>>>> elasticsearch.port=9200
>>>>>>>>>>>>>> elasticsearch.scheme=http
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Is ElasticSearch a mandatory requirement to use Griffin?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 3:58 PM William Guo <[email protected]>
>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Could you check whether ES has been injected with those
>>>>>>>>>>>>>>> metrics or not?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:23 PM Sunil Muniyal <
>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I was able to bypass this error by entering the default
>>>>>>>>>>>>>>>> field values for LDAP, ElasticSearch and Livy in 
>>>>>>>>>>>>>>>> application.properties and
>>>>>>>>>>>>>>>> successfully get Griffin running.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> By following the below article, I have created a test
>>>>>>>>>>>>>>>> measure and then a job which triggers that measure.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Have allowed the job to get triggered multiple times,
>>>>>>>>>>>>>>>> however, still i can't see anything in metrics related to the 
>>>>>>>>>>>>>>>> job. Neither
>>>>>>>>>>>>>>>> I see anything in *health *or *mydashboard* tabs. Also, if
>>>>>>>>>>>>>>>> you notice in the screenshot below, being in the *DQ
>>>>>>>>>>>>>>>> Metrics* tab, I still do not see the created measure in
>>>>>>>>>>>>>>>> the drop down list.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Test job executed multiple times:*
>>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Please advise if anything is mis-configured.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 12:40 PM Sunil Muniyal <
>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thank you for the reply.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> This helped, actually i had missed to add the property in
>>>>>>>>>>>>>>>>> application.properties.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Now the other challenge is, along with ES and Livy, I am
>>>>>>>>>>>>>>>>> also not using LDAP and it is hitting the error *unable
>>>>>>>>>>>>>>>>> to resolve ldap.url property.* Of Course it will, since
>>>>>>>>>>>>>>>>> the property is not configured.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Please suggest.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Sun, Sep 6, 2020 at 7:26 PM William Guo <
>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> hi Sunil Muniyal,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Could you check this property in your griffin properties
>>>>>>>>>>>>>>>>>> file?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> internal.event.listeners
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Thu, Sep 3, 2020 at 11:05 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> I am attempting to integrate Griffin with Cloudera
>>>>>>>>>>>>>>>>>>> Hadoop by following below article:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md>I
>>>>>>>>>>>>>>>>>>> have followed everything as instructed, apart from below 
>>>>>>>>>>>>>>>>>>> things:
>>>>>>>>>>>>>>>>>>> 1. Using Cloudera Hadoop 5.15 and relevant
>>>>>>>>>>>>>>>>>>> configurations instead of Apache Hadoop
>>>>>>>>>>>>>>>>>>> 2. Not using Elastic search as it is not applicable
>>>>>>>>>>>>>>>>>>> 3. Did not use Livy as it is not applicable.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Maven build is successful and has got 2 jars at
>>>>>>>>>>>>>>>>>>> service/target and measure/target which I have uploaded to 
>>>>>>>>>>>>>>>>>>> HDFS.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> However, *starting griffin-service.jar using nohup
>>>>>>>>>>>>>>>>>>> command* is failing with below error:
>>>>>>>>>>>>>>>>>>> *Caused by: java.lang.IllegalArgumentException: Could
>>>>>>>>>>>>>>>>>>> not resolve placeholder 'internal.event.listeners' in 
>>>>>>>>>>>>>>>>>>> string value
>>>>>>>>>>>>>>>>>>> "#{'${internal.event.listeners}'.split(',')}"*
>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.parseStringValue(PropertyPlaceholderHelper.java:174)
>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.replacePlaceholders(PropertyPlaceholderHelper.java:126)
>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>>> org.springframework.core.env.AbstractPropertyResolver.doResolvePlaceholders(AbstractPropertyResolver.java:236)
>>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> I have tried to search a lot of articles with no luck.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Would be great if someone could help me to fix this.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Also, attached is the output of nohup command that was
>>>>>>>>>>>>>>>>>>> written in service.out.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>> To unsubscribe, e-mail:
>>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>>> For additional commands, e-mail:
>>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: [email protected]
>>> For additional commands, e-mail: [email protected]
>>
>>

Reply via email to