Hi William,

Thank you for the reply.

Regarding points 2 and 3. Possible to share some more details. I believe
the env_batch.json is configured as it is expected. What exactly needs to
be updated correctly? ES Hostname or shall I enter IP or something else?
Please help.

Thanks and Regards,
Sunil Muniyal


On Thu, Sep 10, 2020 at 7:30 PM William Guo <[email protected]> wrote:

> 1 OK, We will fix this issue soon.
> 2 Could you try ping es from your spark environment and input ES endpoint
> correctly in env_batch.json
> 3 Please put your es endpoint in env_batch.json
> 6 Please try the following script to build your env.
> ```
>
> #!/bin/bash
> #create table
> hive -f create-table.hqlecho "create table done"
> #current hoursudo ./gen_demo_data.shcur_date=`date 
> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
> insert-data.hql
> hive -f 
> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
> hadoop fs -touchz ${src_done_path}
> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
> #last hoursudo ./gen_demo_data.shcur_date=`date -d '1 hour ago' 
> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
> insert-data.hql
> hive -f 
> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
> hadoop fs -touchz ${src_done_path}
> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
> #next hoursset +ewhile truedo
>   sudo ./gen_demo_data.sh
>   cur_date=`date +%Y%m%d%H`
>   next_date=`date -d "+1hour" '+%Y%m%d%H'`
>   dt=${next_date:0:8}
>   hour=${next_date:8:2}
>   partition_date="dt='$dt',hour='$hour'"
>   sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > 
> insert-data.hql
>   hive -f insert-data.hql
>   src_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONE
>   tgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
>   hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
>   hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
>   hadoop fs -touchz ${src_done_path}
>   hadoop fs -touchz ${tgt_done_path}
>   echo "insert data [$partition_date] done"
>   sleep 3600doneset -e
>
>
>
> Thanks,
>
> William
>
>
> On Thu, Sep 10, 2020 at 4:58 PM Sunil Muniyal <[email protected]>
> wrote:
>
>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>> are the plans further? Is there a page from which I could get updates?
>> --please file a jira ticket for us to make our code ES compatible.
>> [SM] GRIFFIN-346 - Support for Elastic Search latest version (7.9.1)
>> <https://issues.apache.org/jira/browse/GRIFFIN-346> is submitted
>>
>> 2. I still do not see the metrics available (please refer below
>> screenshots). Though the measure is now listed in the drop down of *DQ
>> Metrics* tab. But when I selected the test measure, nothing came up.
>> --could you check the ES whether metrics have been injected or not.
>> [SM] I used the link below and got the index that is created in ES. I
>> believe the data is loaded. However, please correct if I
>> understood incorrectly
>> *"http://<ES Public IP>:9200/_cat/indices?v"*
>> --------------> POC env is on public cloud so using Public IP.
>>
>> health status index   uuid                   pri rep docs.count docs.deleted 
>> store.size pri.store.size
>> yellow open   griffin ur_Kd3XFQBCsPzIM84j87Q   5   2          0            0 
>>      1.2kb          1.2kb
>>
>>
>> Docs in the index:* "http://<ES Public IP>:9200/griffin/_search"*
>>
>> {"took":44,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}}
>>
>>
>> Index Mapping: *"http://<ES Public IP>:9200/griffin"*
>>
>> {"griffin":{"aliases":{},"mappings":{"accuracy":{"properties":{"name":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"tmst":{"type":"date"}}}},"settings":{"index":{"creation_date":"1599567930578","number_of_shards":"5","number_of_replicas":"2","uuid":"ur_Kd3XFQBCsPzIM84j87Q","version":{"created":"6081299"},"provided_name":"griffin"}}}}
>>
>>
>> 3. At a step in deployment guide it is suggested to check URL: "*http://<ES
>> HOST IP>:9200/griffin/accuracy"* When navigated to this URL, I get below
>> error. Please advise
>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>> [GET], allowed: [POST]","status":405}*
>> *-- it seems you need to use POST method.*
>> [SM] I am using the POST method as suggested in the article. Below is the
>> JSON of *env_batch.JSON*
>> *    {*
>> *      "type": "ELASTICSEARCH",*
>> *      "config": {*
>> *        "method": "post",*
>> *        "api": "http://<ES Host Name>:9200/griffin/accuracy", --------->
>> *do we need IP here?
>> *        "connection.timeout": "1m",*
>> *        "retry": 10*
>> *      }*
>> *    }*
>>
>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>> whereas, *demo-tgt* by root. Would that make any difference? If yes, how
>> to correct it? Reload HIVE data?
>> -- could you show me your script for dataset setup?
>> <https://issues.apache.org/jira/browse/GRIFFIN-346>
>> [SM] Attached are the 3 scripts. gen-hive-data.sh is the master script
>> which triggers demo_data and it further triggers delta_src.
>> Have done it as it is instructed in the Github article and
>> gen-hive-data.sh is triggered as root in the terminal.
>>
>>
>> Please advise.
>>
>> Thanks and Regards,
>> Sunil Muniyal
>>
>>
>> On Wed, Sep 9, 2020 at 8:41 PM William Guo <[email protected]> wrote:
>>
>>> *Request you to please advise further on below points:*
>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>>> are the plans further? Is there a page from which I could get updates?
>>> --please file a jira ticket for us to make our code ES compatible.
>>>
>>> 2. I still do not see the metrics available (please refer below
>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>> --could you check the ES whether metrics have been injected or not.
>>>
>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>>> HOST IP>:9200/griffin/accuracy
>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>>> URL, I get below error. Please advise
>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>>> [GET], allowed: [POST]","status":405}*
>>> *-- it seems you need to use POST method.*
>>>
>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>> how to correct it? Reload HIVE data?
>>>
>>> -- could you show me your script for dataset setup?
>>>
>>>
>>> On Tue, Sep 8, 2020 at 9:02 PM Sunil Muniyal <[email protected]>
>>> wrote:
>>>
>>>> Hi William,
>>>>
>>>> I was finally able to get Griffin up and ElasticSearch integrated along
>>>> with Hadoop. Thanks a lot for your help and guidance so far.
>>>>
>>>> I have created a test measure and a job which gets triggered at every 4
>>>> mins automatically (have referred to the user guide available on GitHub at
>>>> this link
>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md>
>>>> .)
>>>>
>>>> *Request you to please advise further on below points:*
>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>>>> are the plans further? Is there a page from which I could get updates?
>>>> 2. I still do not see the metrics available (please refer below
>>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>>>> HOST IP>:9200/griffin/accuracy
>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>>>> URL, I get below error. Please advise
>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>>>> [GET], allowed: [POST]","status":405}*
>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>>> how to correct it? Reload HIVE data?
>>>>
>>>> *Screenshots:*
>>>> *Data Assets:*
>>>> [image: image.png]
>>>>
>>>> *DQ Metrics (Test Measure selected):*
>>>> [image: image.png]
>>>>
>>>> *Job Triggered multiple times:*
>>>> [image: image.png]
>>>>
>>>> *Metrics page from job directly:*
>>>> [image: image.png]
>>>>
>>>> Thanks and Regards,
>>>> Sunil Muniyal
>>>>
>>>>
>>>> On Tue, Sep 8, 2020 at 4:38 PM Sunil Muniyal <
>>>> [email protected]> wrote:
>>>>
>>>>> I am unable to get repos for 6.4.1 instead I found 6.8.x. Will try
>>>>> with this version of Elastic Search in sometime.
>>>>>
>>>>> In the meantime, would it be possible to confirm if 6.4.x or 6.8.x is
>>>>> the only supported version for Griffin? Reason I am asking is, the GitHub
>>>>> article for griffin deployment points to the latest version of ES.
>>>>>
>>>>> Thanks and Regards,
>>>>> Sunil Muniyal
>>>>>
>>>>>
>>>>> On Tue, Sep 8, 2020 at 4:06 PM Sunil Muniyal <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> I will need to redeploy ElasticSearch, correct?
>>>>>>
>>>>>> Thanks and Regards,
>>>>>> Sunil Muniyal
>>>>>>
>>>>>>
>>>>>> On Tue, Sep 8, 2020 at 4:05 PM William Guo <[email protected]> wrote:
>>>>>>
>>>>>>> Could you try with this version?
>>>>>>> <elasticsearch.version>6.4.1</elasticsearch.version>
>>>>>>>
>>>>>>>
>>>>>>> Thanks,
>>>>>>> William
>>>>>>>
>>>>>>> On Tue, Sep 8, 2020 at 5:59 PM Sunil Muniyal <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Hi William / Dev group,
>>>>>>>>
>>>>>>>> I have deployed ES 7.9 - latest version (single node) and the same
>>>>>>>> is configured. I also get the default page when hitting
>>>>>>>> http://<ES HOST IP>:9200/
>>>>>>>>
>>>>>>>> Upon creating the griffin configurations using the JSON string given
>>>>>>>>
>>>>>>>> curl -k -H "Content-Type: application/json" -X PUT http://<replaced 
>>>>>>>> with my ES host IP>:9200/griffin \
>>>>>>>>  -d '{
>>>>>>>>     "aliases": {},
>>>>>>>>     "mappings": {
>>>>>>>>         "accuracy": {
>>>>>>>>             "properties": {
>>>>>>>>                 "name": {
>>>>>>>>                     "fields": {
>>>>>>>>                         "keyword": {
>>>>>>>>                             "ignore_above": 256,
>>>>>>>>                             "type": "keyword"
>>>>>>>>                         }
>>>>>>>>                     },
>>>>>>>>                     "type": "text"
>>>>>>>>                 },
>>>>>>>>                 "tmst": {
>>>>>>>>                     "type": "date"
>>>>>>>>                 }
>>>>>>>>             }
>>>>>>>>         }
>>>>>>>>     },
>>>>>>>>     "settings": {
>>>>>>>>         "index": {
>>>>>>>>             "number_of_replicas": "2",
>>>>>>>>             "number_of_shards": "5"
>>>>>>>>         }
>>>>>>>>     }
>>>>>>>> }'
>>>>>>>>
>>>>>>>>
>>>>>>>> *I get below error:*
>>>>>>>>
>>>>>>>> *{"error":{"root_cause":[{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>> type=text},
>>>>>>>> tmst={type=date}}}]"}],"type":"mapper_parsing_exception","reason":"Failed
>>>>>>>> to parse mapping [_doc]: Root mapping definition has unsupported
>>>>>>>> parameters:  [accuracy :
>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>> type=text},
>>>>>>>> tmst={type=date}}}]","caused_by":{"type":"mapper_parsing_exception","reason":"Root
>>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>>> type=text}, tmst={type=date}}}]"}},"status":400}*
>>>>>>>>
>>>>>>>> Seems like the JSON string is missing some values or is incorrectly
>>>>>>>> provided.
>>>>>>>>
>>>>>>>> Would be great if you could please help.
>>>>>>>>
>>>>>>>> Thanks and Regards,
>>>>>>>> Sunil Muniyal
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Sep 7, 2020 at 8:16 PM Sunil Muniyal <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Thank you for the response, William.
>>>>>>>>>
>>>>>>>>> I have started preparing for ES deployment and should attempt the
>>>>>>>>> same tomorrow.
>>>>>>>>>
>>>>>>>>> In the meantime, I will also wait for the Dev team in case they
>>>>>>>>> have any additional inputs.
>>>>>>>>>
>>>>>>>>> Thanks and Regards,
>>>>>>>>> Sunil Muniyal
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Sep 7, 2020 at 8:06 PM William Guo <[email protected]>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>>> will need to:
>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>> 2. Update application.properties to include ES details and create
>>>>>>>>>> ES index
>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>
>>>>>>>>>> *Right, you need to package es env configuration into your jar.*
>>>>>>>>>>
>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>>
>>>>>>>>>> *No*
>>>>>>>>>>
>>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>> 1. What are the exact ports requirements (internal and external)?
>>>>>>>>>> *check log and make sure all extra connections in properties can
>>>>>>>>>> accessible*
>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>> *no*
>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>> *java 1.8*
>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what
>>>>>>>>>> are the dependencies or additional configurations needed?
>>>>>>>>>> *Should no extra dependencies, except those transitive
>>>>>>>>>> dependencies incurred by spark and hadoop.*
>>>>>>>>>>
>>>>>>>>>> On Mon, Sep 7, 2020 at 6:42 PM Sunil Muniyal <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> Ohh ok.
>>>>>>>>>>>
>>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>>>> will need to:
>>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>>> 2. Update application.properties to include ES details and
>>>>>>>>>>> create ES index
>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>>
>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>>>
>>>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>>> 1. What are the exact ports requirements (internal and external)?
>>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what
>>>>>>>>>>> are the dependencies or additional configurations needed?
>>>>>>>>>>>
>>>>>>>>>>> I know some of the above information can be fetched from the
>>>>>>>>>>> deployment guide on Github. However, checking if any other formal
>>>>>>>>>>> documentation has been made available for the same?
>>>>>>>>>>>
>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Sep 7, 2020 at 4:05 PM William Guo <[email protected]>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> cc dev for double checking.
>>>>>>>>>>>>
>>>>>>>>>>>> Measure will emit metrics and store them in elastic, UI fetch
>>>>>>>>>>>> those metrics from elastic.
>>>>>>>>>>>> So elastic should be mandatory.
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> William
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:32 PM Sunil Muniyal <
>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Thank you for the quick response, William.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have not configured ElasticSearch since it is not deployed.
>>>>>>>>>>>>>
>>>>>>>>>>>>> In the application.properties, I just added the dummy
>>>>>>>>>>>>> information (as below) just to pass the validation test and get 
>>>>>>>>>>>>> Griffin up
>>>>>>>>>>>>> and running.
>>>>>>>>>>>>>
>>>>>>>>>>>>> # elasticsearch
>>>>>>>>>>>>> # elasticsearch.host = <IP>
>>>>>>>>>>>>> # elasticsearch.port = <elasticsearch rest port>
>>>>>>>>>>>>> # elasticsearch.user = user
>>>>>>>>>>>>> # elasticsearch.password = password
>>>>>>>>>>>>> elasticsearch.host=localhost
>>>>>>>>>>>>> elasticsearch.port=9200
>>>>>>>>>>>>> elasticsearch.scheme=http
>>>>>>>>>>>>>
>>>>>>>>>>>>> Is ElasticSearch a mandatory requirement to use Griffin?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 3:58 PM William Guo <[email protected]>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Could you check whether ES has been injected with those
>>>>>>>>>>>>>> metrics or not?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:23 PM Sunil Muniyal <
>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I was able to bypass this error by entering the default
>>>>>>>>>>>>>>> field values for LDAP, ElasticSearch and Livy in 
>>>>>>>>>>>>>>> application.properties and
>>>>>>>>>>>>>>> successfully get Griffin running.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> By following the below article, I have created a test
>>>>>>>>>>>>>>> measure and then a job which triggers that measure.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Have allowed the job to get triggered multiple times,
>>>>>>>>>>>>>>> however, still i can't see anything in metrics related to the 
>>>>>>>>>>>>>>> job. Neither
>>>>>>>>>>>>>>> I see anything in *health *or *mydashboard* tabs. Also, if
>>>>>>>>>>>>>>> you notice in the screenshot below, being in the *DQ
>>>>>>>>>>>>>>> Metrics* tab, I still do not see the created measure in the
>>>>>>>>>>>>>>> drop down list.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Test job executed multiple times:*
>>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Please advise if anything is mis-configured.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 12:40 PM Sunil Muniyal <
>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thank you for the reply.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> This helped, actually i had missed to add the property in
>>>>>>>>>>>>>>>> application.properties.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Now the other challenge is, along with ES and Livy, I am
>>>>>>>>>>>>>>>> also not using LDAP and it is hitting the error *unable to
>>>>>>>>>>>>>>>> resolve ldap.url property.* Of Course it will, since the
>>>>>>>>>>>>>>>> property is not configured.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Please suggest.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Sun, Sep 6, 2020 at 7:26 PM William Guo <
>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> hi Sunil Muniyal,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Could you check this property in your griffin properties
>>>>>>>>>>>>>>>>> file?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> internal.event.listeners
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Thu, Sep 3, 2020 at 11:05 PM Sunil Muniyal <
>>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I am attempting to integrate Griffin with Cloudera Hadoop
>>>>>>>>>>>>>>>>>> by following below article:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md>I
>>>>>>>>>>>>>>>>>> have followed everything as instructed, apart from below 
>>>>>>>>>>>>>>>>>> things:
>>>>>>>>>>>>>>>>>> 1. Using Cloudera Hadoop 5.15 and relevant configurations
>>>>>>>>>>>>>>>>>> instead of Apache Hadoop
>>>>>>>>>>>>>>>>>> 2. Not using Elastic search as it is not applicable
>>>>>>>>>>>>>>>>>> 3. Did not use Livy as it is not applicable.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Maven build is successful and has got 2 jars at
>>>>>>>>>>>>>>>>>> service/target and measure/target which I have uploaded to 
>>>>>>>>>>>>>>>>>> HDFS.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> However, *starting griffin-service.jar using nohup
>>>>>>>>>>>>>>>>>> command* is failing with below error:
>>>>>>>>>>>>>>>>>> *Caused by: java.lang.IllegalArgumentException: Could not
>>>>>>>>>>>>>>>>>> resolve placeholder 'internal.event.listeners' in string 
>>>>>>>>>>>>>>>>>> value
>>>>>>>>>>>>>>>>>> "#{'${internal.event.listeners}'.split(',')}"*
>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.parseStringValue(PropertyPlaceholderHelper.java:174)
>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.replacePlaceholders(PropertyPlaceholderHelper.java:126)
>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>>> org.springframework.core.env.AbstractPropertyResolver.doResolvePlaceholders(AbstractPropertyResolver.java:236)
>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I have tried to search a lot of articles with no luck.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Would be great if someone could help me to fix this.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Also, attached is the output of nohup command that was
>>>>>>>>>>>>>>>>>> written in service.out.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> To unsubscribe, e-mail:
>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>> For additional commands, e-mail:
>>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: [email protected]
>> For additional commands, e-mail: [email protected]
>
>

Reply via email to