please enter ip directly. not sure whether hostname can be resolved correctly or not.
On Thu, Sep 10, 2020 at 10:06 PM Sunil Muniyal <[email protected]> wrote: > Hi William, > > Thank you for the reply. > > Regarding points 2 and 3. Possible to share some more details. I believe > the env_batch.json is configured as it is expected. What exactly needs to > be updated correctly? ES Hostname or shall I enter IP or something else? > Please help. > > Thanks and Regards, > Sunil Muniyal > > > On Thu, Sep 10, 2020 at 7:30 PM William Guo <[email protected]> wrote: > >> 1 OK, We will fix this issue soon. >> 2 Could you try ping es from your spark environment and input ES endpoint >> correctly in env_batch.json >> 3 Please put your es endpoint in env_batch.json >> 6 Please try the following script to build your env. >> ``` >> >> #!/bin/bash >> #create table >> hive -f create-table.hqlecho "create table done" >> #current hoursudo ./gen_demo_data.shcur_date=`date >> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'" >> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > >> insert-data.hql >> hive -f >> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE >> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour} >> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour} >> hadoop fs -touchz ${src_done_path} >> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done" >> #last hoursudo ./gen_demo_data.shcur_date=`date -d '1 hour ago' >> +%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'" >> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > >> insert-data.hql >> hive -f >> insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE >> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour} >> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour} >> hadoop fs -touchz ${src_done_path} >> hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done" >> #next hoursset +ewhile truedo >> sudo ./gen_demo_data.sh >> cur_date=`date +%Y%m%d%H` >> next_date=`date -d "+1hour" '+%Y%m%d%H'` >> dt=${next_date:0:8} >> hour=${next_date:8:2} >> partition_date="dt='$dt',hour='$hour'" >> sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template > >> insert-data.hql >> hive -f insert-data.hql >> src_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONE >> tgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE >> hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour} >> hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour} >> hadoop fs -touchz ${src_done_path} >> hadoop fs -touchz ${tgt_done_path} >> echo "insert data [$partition_date] done" >> sleep 3600doneset -e >> >> >> >> Thanks, >> >> William >> >> >> On Thu, Sep 10, 2020 at 4:58 PM Sunil Muniyal <[email protected]> >> wrote: >> >>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean >>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what >>> are the plans further? Is there a page from which I could get updates? >>> --please file a jira ticket for us to make our code ES compatible. >>> [SM] GRIFFIN-346 - Support for Elastic Search latest version (7.9.1) >>> <https://issues.apache.org/jira/browse/GRIFFIN-346> is submitted >>> >>> 2. I still do not see the metrics available (please refer below >>> screenshots). Though the measure is now listed in the drop down of *DQ >>> Metrics* tab. But when I selected the test measure, nothing came up. >>> --could you check the ES whether metrics have been injected or not. >>> [SM] I used the link below and got the index that is created in ES. I >>> believe the data is loaded. However, please correct if I >>> understood incorrectly >>> *"http://<ES Public IP>:9200/_cat/indices?v"* >>> --------------> POC env is on public cloud so using Public IP. >>> >>> health status index uuid pri rep docs.count >>> docs.deleted store.size pri.store.size >>> yellow open griffin ur_Kd3XFQBCsPzIM84j87Q 5 2 0 >>> 0 1.2kb 1.2kb >>> >>> >>> Docs in the index:* "http://<ES Public IP>:9200/griffin/_search"* >>> >>> {"took":44,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}} >>> >>> >>> Index Mapping: *"http://<ES Public IP>:9200/griffin"* >>> >>> {"griffin":{"aliases":{},"mappings":{"accuracy":{"properties":{"name":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"tmst":{"type":"date"}}}},"settings":{"index":{"creation_date":"1599567930578","number_of_shards":"5","number_of_replicas":"2","uuid":"ur_Kd3XFQBCsPzIM84j87Q","version":{"created":"6081299"},"provided_name":"griffin"}}}} >>> >>> >>> 3. At a step in deployment guide it is suggested to check URL: "*http://<ES >>> HOST IP>:9200/griffin/accuracy"* When navigated to this URL, I get >>> below error. Please advise >>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method >>> [GET], allowed: [POST]","status":405}* >>> *-- it seems you need to use POST method.* >>> [SM] I am using the POST method as suggested in the article. Below is >>> the JSON of *env_batch.JSON* >>> * {* >>> * "type": "ELASTICSEARCH",* >>> * "config": {* >>> * "method": "post",* >>> * "api": "http://<ES Host Name>:9200/griffin/accuracy", >>> ---------> *do we need IP here? >>> * "connection.timeout": "1m",* >>> * "retry": 10* >>> * }* >>> * }* >>> >>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin >>> whereas, *demo-tgt* by root. Would that make any difference? If yes, >>> how to correct it? Reload HIVE data? >>> -- could you show me your script for dataset setup? >>> <https://issues.apache.org/jira/browse/GRIFFIN-346> >>> [SM] Attached are the 3 scripts. gen-hive-data.sh is the master script >>> which triggers demo_data and it further triggers delta_src. >>> Have done it as it is instructed in the Github article and >>> gen-hive-data.sh is triggered as root in the terminal. >>> >>> >>> Please advise. >>> >>> Thanks and Regards, >>> Sunil Muniyal >>> >>> >>> On Wed, Sep 9, 2020 at 8:41 PM William Guo <[email protected]> wrote: >>> >>>> *Request you to please advise further on below points:* >>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean >>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what >>>> are the plans further? Is there a page from which I could get updates? >>>> --please file a jira ticket for us to make our code ES compatible. >>>> >>>> 2. I still do not see the metrics available (please refer below >>>> screenshots). Though the measure is now listed in the drop down of *DQ >>>> Metrics* tab. But when I selected the test measure, nothing came up. >>>> --could you check the ES whether metrics have been injected or not. >>>> >>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES >>>> HOST IP>:9200/griffin/accuracy >>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this >>>> URL, I get below error. Please advise >>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method >>>> [GET], allowed: [POST]","status":405}* >>>> *-- it seems you need to use POST method.* >>>> >>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin >>>> whereas, *demo-tgt* by root. Would that make any difference? If yes, >>>> how to correct it? Reload HIVE data? >>>> >>>> -- could you show me your script for dataset setup? >>>> >>>> >>>> On Tue, Sep 8, 2020 at 9:02 PM Sunil Muniyal < >>>> [email protected]> wrote: >>>> >>>>> Hi William, >>>>> >>>>> I was finally able to get Griffin up and ElasticSearch integrated >>>>> along with Hadoop. Thanks a lot for your help and guidance so far. >>>>> >>>>> I have created a test measure and a job which gets triggered at every >>>>> 4 mins automatically (have referred to the user guide available on GitHub >>>>> at this link >>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md> >>>>> .) >>>>> >>>>> *Request you to please advise further on below points:* >>>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it >>>>> mean that only ES upto 6.8.x is supported for Griffin as of now? If yes, >>>>> what are the plans further? Is there a page from which I could get >>>>> updates? >>>>> 2. I still do not see the metrics available (please refer below >>>>> screenshots). Though the measure is now listed in the drop down of *DQ >>>>> Metrics* tab. But when I selected the test measure, nothing came up. >>>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES >>>>> HOST IP>:9200/griffin/accuracy >>>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this >>>>> URL, I get below error. Please advise >>>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and >>>>> method [GET], allowed: [POST]","status":405}* >>>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin >>>>> whereas, *demo-tgt* by root. Would that make any difference? If yes, >>>>> how to correct it? Reload HIVE data? >>>>> >>>>> *Screenshots:* >>>>> *Data Assets:* >>>>> [image: image.png] >>>>> >>>>> *DQ Metrics (Test Measure selected):* >>>>> [image: image.png] >>>>> >>>>> *Job Triggered multiple times:* >>>>> [image: image.png] >>>>> >>>>> *Metrics page from job directly:* >>>>> [image: image.png] >>>>> >>>>> Thanks and Regards, >>>>> Sunil Muniyal >>>>> >>>>> >>>>> On Tue, Sep 8, 2020 at 4:38 PM Sunil Muniyal < >>>>> [email protected]> wrote: >>>>> >>>>>> I am unable to get repos for 6.4.1 instead I found 6.8.x. Will try >>>>>> with this version of Elastic Search in sometime. >>>>>> >>>>>> In the meantime, would it be possible to confirm if 6.4.x or 6.8.x is >>>>>> the only supported version for Griffin? Reason I am asking is, the GitHub >>>>>> article for griffin deployment points to the latest version of ES. >>>>>> >>>>>> Thanks and Regards, >>>>>> Sunil Muniyal >>>>>> >>>>>> >>>>>> On Tue, Sep 8, 2020 at 4:06 PM Sunil Muniyal < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> I will need to redeploy ElasticSearch, correct? >>>>>>> >>>>>>> Thanks and Regards, >>>>>>> Sunil Muniyal >>>>>>> >>>>>>> >>>>>>> On Tue, Sep 8, 2020 at 4:05 PM William Guo <[email protected]> wrote: >>>>>>> >>>>>>>> Could you try with this version? >>>>>>>> <elasticsearch.version>6.4.1</elasticsearch.version> >>>>>>>> >>>>>>>> >>>>>>>> Thanks, >>>>>>>> William >>>>>>>> >>>>>>>> On Tue, Sep 8, 2020 at 5:59 PM Sunil Muniyal < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hi William / Dev group, >>>>>>>>> >>>>>>>>> I have deployed ES 7.9 - latest version (single node) and the same >>>>>>>>> is configured. I also get the default page when hitting >>>>>>>>> http://<ES HOST IP>:9200/ >>>>>>>>> >>>>>>>>> Upon creating the griffin configurations using the JSON string >>>>>>>>> given >>>>>>>>> >>>>>>>>> curl -k -H "Content-Type: application/json" -X PUT http://<replaced >>>>>>>>> with my ES host IP>:9200/griffin \ >>>>>>>>> -d '{ >>>>>>>>> "aliases": {}, >>>>>>>>> "mappings": { >>>>>>>>> "accuracy": { >>>>>>>>> "properties": { >>>>>>>>> "name": { >>>>>>>>> "fields": { >>>>>>>>> "keyword": { >>>>>>>>> "ignore_above": 256, >>>>>>>>> "type": "keyword" >>>>>>>>> } >>>>>>>>> }, >>>>>>>>> "type": "text" >>>>>>>>> }, >>>>>>>>> "tmst": { >>>>>>>>> "type": "date" >>>>>>>>> } >>>>>>>>> } >>>>>>>>> } >>>>>>>>> }, >>>>>>>>> "settings": { >>>>>>>>> "index": { >>>>>>>>> "number_of_replicas": "2", >>>>>>>>> "number_of_shards": "5" >>>>>>>>> } >>>>>>>>> } >>>>>>>>> }' >>>>>>>>> >>>>>>>>> >>>>>>>>> *I get below error:* >>>>>>>>> >>>>>>>>> *{"error":{"root_cause":[{"type":"mapper_parsing_exception","reason":"Root >>>>>>>>> mapping definition has unsupported parameters: [accuracy : >>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}}, >>>>>>>>> type=text}, >>>>>>>>> tmst={type=date}}}]"}],"type":"mapper_parsing_exception","reason":"Failed >>>>>>>>> to parse mapping [_doc]: Root mapping definition has unsupported >>>>>>>>> parameters: [accuracy : >>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}}, >>>>>>>>> type=text}, >>>>>>>>> tmst={type=date}}}]","caused_by":{"type":"mapper_parsing_exception","reason":"Root >>>>>>>>> mapping definition has unsupported parameters: [accuracy : >>>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}}, >>>>>>>>> type=text}, tmst={type=date}}}]"}},"status":400}* >>>>>>>>> >>>>>>>>> Seems like the JSON string is missing some values or is >>>>>>>>> incorrectly provided. >>>>>>>>> >>>>>>>>> Would be great if you could please help. >>>>>>>>> >>>>>>>>> Thanks and Regards, >>>>>>>>> Sunil Muniyal >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Sep 7, 2020 at 8:16 PM Sunil Muniyal < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> Thank you for the response, William. >>>>>>>>>> >>>>>>>>>> I have started preparing for ES deployment and should attempt the >>>>>>>>>> same tomorrow. >>>>>>>>>> >>>>>>>>>> In the meantime, I will also wait for the Dev team in case they >>>>>>>>>> have any additional inputs. >>>>>>>>>> >>>>>>>>>> Thanks and Regards, >>>>>>>>>> Sunil Muniyal >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Sep 7, 2020 at 8:06 PM William Guo <[email protected]> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I >>>>>>>>>>> will need to: >>>>>>>>>>> 1. Deploy and Configure ES >>>>>>>>>>> 2. Update application.properties to include ES details and >>>>>>>>>>> create ES index >>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service >>>>>>>>>>> >>>>>>>>>>> *Right, you need to package es env configuration into your jar.* >>>>>>>>>>> >>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct? >>>>>>>>>>> >>>>>>>>>>> *No* >>>>>>>>>>> >>>>>>>>>>> On a side note, is there any other documentation of Griffin >>>>>>>>>>> available or underway which would help to get below details while >>>>>>>>>>> integrating it with Cloudera Hadoop? >>>>>>>>>>> 1. What are the exact ports requirements (internal and external)? >>>>>>>>>>> *check log and make sure all extra connections in properties can >>>>>>>>>>> accessible* >>>>>>>>>>> 2. Which all packages will be required? >>>>>>>>>>> *no* >>>>>>>>>>> 3. Any Java dependencies? >>>>>>>>>>> *java 1.8* >>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what >>>>>>>>>>> are the dependencies or additional configurations needed? >>>>>>>>>>> *Should no extra dependencies, except those transitive >>>>>>>>>>> dependencies incurred by spark and hadoop.* >>>>>>>>>>> >>>>>>>>>>> On Mon, Sep 7, 2020 at 6:42 PM Sunil Muniyal < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Ohh ok. >>>>>>>>>>>> >>>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I >>>>>>>>>>>> will need to: >>>>>>>>>>>> 1. Deploy and Configure ES >>>>>>>>>>>> 2. Update application.properties to include ES details and >>>>>>>>>>>> create ES index >>>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service >>>>>>>>>>>> >>>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct? >>>>>>>>>>>> >>>>>>>>>>>> On a side note, is there any other documentation of Griffin >>>>>>>>>>>> available or underway which would help to get below details while >>>>>>>>>>>> integrating it with Cloudera Hadoop? >>>>>>>>>>>> 1. What are the exact ports requirements (internal and >>>>>>>>>>>> external)? >>>>>>>>>>>> 2. Which all packages will be required? >>>>>>>>>>>> 3. Any Java dependencies? >>>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), >>>>>>>>>>>> what are the dependencies or additional configurations needed? >>>>>>>>>>>> >>>>>>>>>>>> I know some of the above information can be fetched from the >>>>>>>>>>>> deployment guide on Github. However, checking if any other formal >>>>>>>>>>>> documentation has been made available for the same? >>>>>>>>>>>> >>>>>>>>>>>> Thanks and Regards, >>>>>>>>>>>> Sunil Muniyal >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 7, 2020 at 4:05 PM William Guo <[email protected]> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> cc dev for double checking. >>>>>>>>>>>>> >>>>>>>>>>>>> Measure will emit metrics and store them in elastic, UI fetch >>>>>>>>>>>>> those metrics from elastic. >>>>>>>>>>>>> So elastic should be mandatory. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> William >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:32 PM Sunil Muniyal < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Thank you for the quick response, William. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I have not configured ElasticSearch since it is not deployed. >>>>>>>>>>>>>> >>>>>>>>>>>>>> In the application.properties, I just added the dummy >>>>>>>>>>>>>> information (as below) just to pass the validation test and get >>>>>>>>>>>>>> Griffin up >>>>>>>>>>>>>> and running. >>>>>>>>>>>>>> >>>>>>>>>>>>>> # elasticsearch >>>>>>>>>>>>>> # elasticsearch.host = <IP> >>>>>>>>>>>>>> # elasticsearch.port = <elasticsearch rest port> >>>>>>>>>>>>>> # elasticsearch.user = user >>>>>>>>>>>>>> # elasticsearch.password = password >>>>>>>>>>>>>> elasticsearch.host=localhost >>>>>>>>>>>>>> elasticsearch.port=9200 >>>>>>>>>>>>>> elasticsearch.scheme=http >>>>>>>>>>>>>> >>>>>>>>>>>>>> Is ElasticSearch a mandatory requirement to use Griffin? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks and Regards, >>>>>>>>>>>>>> Sunil Muniyal >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 3:58 PM William Guo <[email protected]> >>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Could you check whether ES has been injected with those >>>>>>>>>>>>>>> metrics or not? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:23 PM Sunil Muniyal < >>>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hello William, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I was able to bypass this error by entering the default >>>>>>>>>>>>>>>> field values for LDAP, ElasticSearch and Livy in >>>>>>>>>>>>>>>> application.properties and >>>>>>>>>>>>>>>> successfully get Griffin running. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> By following the below article, I have created a test >>>>>>>>>>>>>>>> measure and then a job which triggers that measure. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Have allowed the job to get triggered multiple times, >>>>>>>>>>>>>>>> however, still i can't see anything in metrics related to the >>>>>>>>>>>>>>>> job. Neither >>>>>>>>>>>>>>>> I see anything in *health *or *mydashboard* tabs. Also, if >>>>>>>>>>>>>>>> you notice in the screenshot below, being in the *DQ >>>>>>>>>>>>>>>> Metrics* tab, I still do not see the created measure in >>>>>>>>>>>>>>>> the drop down list. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [image: image.png] >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> *Test job executed multiple times:* >>>>>>>>>>>>>>>> [image: image.png] >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Please advise if anything is mis-configured. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks and Regards, >>>>>>>>>>>>>>>> Sunil Muniyal >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 12:40 PM Sunil Muniyal < >>>>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hello William, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thank you for the reply. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> This helped, actually i had missed to add the property in >>>>>>>>>>>>>>>>> application.properties. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Now the other challenge is, along with ES and Livy, I am >>>>>>>>>>>>>>>>> also not using LDAP and it is hitting the error *unable >>>>>>>>>>>>>>>>> to resolve ldap.url property.* Of Course it will, since >>>>>>>>>>>>>>>>> the property is not configured. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Please suggest. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks and Regards, >>>>>>>>>>>>>>>>> Sunil Muniyal >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Sun, Sep 6, 2020 at 7:26 PM William Guo < >>>>>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> hi Sunil Muniyal, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Could you check this property in your griffin properties >>>>>>>>>>>>>>>>>> file? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> internal.event.listeners >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> William >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Thu, Sep 3, 2020 at 11:05 PM Sunil Muniyal < >>>>>>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I am attempting to integrate Griffin with Cloudera >>>>>>>>>>>>>>>>>>> Hadoop by following below article: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md>I >>>>>>>>>>>>>>>>>>> have followed everything as instructed, apart from below >>>>>>>>>>>>>>>>>>> things: >>>>>>>>>>>>>>>>>>> 1. Using Cloudera Hadoop 5.15 and relevant >>>>>>>>>>>>>>>>>>> configurations instead of Apache Hadoop >>>>>>>>>>>>>>>>>>> 2. Not using Elastic search as it is not applicable >>>>>>>>>>>>>>>>>>> 3. Did not use Livy as it is not applicable. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Maven build is successful and has got 2 jars at >>>>>>>>>>>>>>>>>>> service/target and measure/target which I have uploaded to >>>>>>>>>>>>>>>>>>> HDFS. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> However, *starting griffin-service.jar using nohup >>>>>>>>>>>>>>>>>>> command* is failing with below error: >>>>>>>>>>>>>>>>>>> *Caused by: java.lang.IllegalArgumentException: Could >>>>>>>>>>>>>>>>>>> not resolve placeholder 'internal.event.listeners' in >>>>>>>>>>>>>>>>>>> string value >>>>>>>>>>>>>>>>>>> "#{'${internal.event.listeners}'.split(',')}"* >>>>>>>>>>>>>>>>>>> * at >>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.parseStringValue(PropertyPlaceholderHelper.java:174) >>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]* >>>>>>>>>>>>>>>>>>> * at >>>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.replacePlaceholders(PropertyPlaceholderHelper.java:126) >>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]* >>>>>>>>>>>>>>>>>>> * at >>>>>>>>>>>>>>>>>>> org.springframework.core.env.AbstractPropertyResolver.doResolvePlaceholders(AbstractPropertyResolver.java:236) >>>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]* >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I have tried to search a lot of articles with no luck. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Would be great if someone could help me to fix this. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Also, attached is the output of nohup command that was >>>>>>>>>>>>>>>>>>> written in service.out. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Thanks and Regards, >>>>>>>>>>>>>>>>>>> Sunil Muniyal >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------- >>>>>>>>>>>>>>>>>>> To unsubscribe, e-mail: >>>>>>>>>>>>>>>>>>> [email protected] >>>>>>>>>>>>>>>>>>> For additional commands, e-mail: >>>>>>>>>>>>>>>>>>> [email protected] >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: [email protected] >>> For additional commands, e-mail: [email protected] >> >>
