1 OK, We will fix this issue soon.
2 Could you try ping es from your spark environment and input ES endpoint
correctly in env_batch.json
3 Please put your es endpoint in env_batch.json
6 Please try the following script to build your env.
```

#!/bin/bash
#create table
hive -f create-table.hqlecho "create table done"
#current hoursudo ./gen_demo_data.shcur_date=`date
+%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template >
insert-data.hql
hive -f 
insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
hadoop fs -touchz ${src_done_path}
hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
#last hoursudo ./gen_demo_data.shcur_date=`date -d '1 hour ago'
+%Y%m%d%H`dt=${cur_date:0:8}hour=${cur_date:8:2}partition_date="dt='$dt',hour='$hour'"
sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template >
insert-data.hql
hive -f 
insert-data.hqlsrc_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONEtgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
hadoop fs -touchz ${src_done_path}
hadoop fs -touchz ${tgt_done_path}echo "insert data [$partition_date] done"
#next hoursset +ewhile truedo
  sudo ./gen_demo_data.sh
  cur_date=`date +%Y%m%d%H`
  next_date=`date -d "+1hour" '+%Y%m%d%H'`
  dt=${next_date:0:8}
  hour=${next_date:8:2}
  partition_date="dt='$dt',hour='$hour'"
  sed s/PARTITION_DATE/$partition_date/ ./insert-data.hql.template >
insert-data.hql
  hive -f insert-data.hql
  src_done_path=/griffin/data/batch/demo_src/dt=${dt}/hour=${hour}/_DONE
  tgt_done_path=/griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}/_DONE
  hadoop fs -mkdir -p /griffin/data/batch/demo_src/dt=${dt}/hour=${hour}
  hadoop fs -mkdir -p /griffin/data/batch/demo_tgt/dt=${dt}/hour=${hour}
  hadoop fs -touchz ${src_done_path}
  hadoop fs -touchz ${tgt_done_path}
  echo "insert data [$partition_date] done"
  sleep 3600doneset -e



Thanks,

William


On Thu, Sep 10, 2020 at 4:58 PM Sunil Muniyal <[email protected]>
wrote:

> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
> are the plans further? Is there a page from which I could get updates?
> --please file a jira ticket for us to make our code ES compatible.
> [SM] GRIFFIN-346 - Support for Elastic Search latest version (7.9.1)
> <https://issues.apache.org/jira/browse/GRIFFIN-346> is submitted
>
> 2. I still do not see the metrics available (please refer below
> screenshots). Though the measure is now listed in the drop down of *DQ
> Metrics* tab. But when I selected the test measure, nothing came up.
> --could you check the ES whether metrics have been injected or not.
> [SM] I used the link below and got the index that is created in ES. I
> believe the data is loaded. However, please correct if I
> understood incorrectly
> *"http://<ES Public IP>:9200/_cat/indices?v"*
> --------------> POC env is on public cloud so using Public IP.
>
> health status index   uuid                   pri rep docs.count docs.deleted 
> store.size pri.store.size
> yellow open   griffin ur_Kd3XFQBCsPzIM84j87Q   5   2          0            0  
>     1.2kb          1.2kb
>
>
> Docs in the index:* "http://<ES Public IP>:9200/griffin/_search"*
>
> {"took":44,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}}
>
>
> Index Mapping: *"http://<ES Public IP>:9200/griffin"*
>
> {"griffin":{"aliases":{},"mappings":{"accuracy":{"properties":{"name":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"tmst":{"type":"date"}}}},"settings":{"index":{"creation_date":"1599567930578","number_of_shards":"5","number_of_replicas":"2","uuid":"ur_Kd3XFQBCsPzIM84j87Q","version":{"created":"6081299"},"provided_name":"griffin"}}}}
>
>
> 3. At a step in deployment guide it is suggested to check URL: "*http://<ES
> HOST IP>:9200/griffin/accuracy"* When navigated to this URL, I get below
> error. Please advise
> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
> [GET], allowed: [POST]","status":405}*
> *-- it seems you need to use POST method.*
> [SM] I am using the POST method as suggested in the article. Below is the
> JSON of *env_batch.JSON*
> *    {*
> *      "type": "ELASTICSEARCH",*
> *      "config": {*
> *        "method": "post",*
> *        "api": "http://<ES Host Name>:9200/griffin/accuracy", ---------> *do
> we need IP here?
> *        "connection.timeout": "1m",*
> *        "retry": 10*
> *      }*
> *    }*
>
> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
> whereas, *demo-tgt* by root. Would that make any difference? If yes, how
> to correct it? Reload HIVE data?
> -- could you show me your script for dataset setup?
> <https://issues.apache.org/jira/browse/GRIFFIN-346>
> [SM] Attached are the 3 scripts. gen-hive-data.sh is the master script
> which triggers demo_data and it further triggers delta_src.
> Have done it as it is instructed in the Github article and
> gen-hive-data.sh is triggered as root in the terminal.
>
>
> Please advise.
>
> Thanks and Regards,
> Sunil Muniyal
>
>
> On Wed, Sep 9, 2020 at 8:41 PM William Guo <[email protected]> wrote:
>
>> *Request you to please advise further on below points:*
>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>> are the plans further? Is there a page from which I could get updates?
>> --please file a jira ticket for us to make our code ES compatible.
>>
>> 2. I still do not see the metrics available (please refer below
>> screenshots). Though the measure is now listed in the drop down of *DQ
>> Metrics* tab. But when I selected the test measure, nothing came up.
>> --could you check the ES whether metrics have been injected or not.
>>
>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>> HOST IP>:9200/griffin/accuracy
>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>> URL, I get below error. Please advise
>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>> [GET], allowed: [POST]","status":405}*
>> *-- it seems you need to use POST method.*
>>
>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>> whereas, *demo-tgt* by root. Would that make any difference? If yes, how
>> to correct it? Reload HIVE data?
>>
>> -- could you show me your script for dataset setup?
>>
>>
>> On Tue, Sep 8, 2020 at 9:02 PM Sunil Muniyal <[email protected]>
>> wrote:
>>
>>> Hi William,
>>>
>>> I was finally able to get Griffin up and ElasticSearch integrated along
>>> with Hadoop. Thanks a lot for your help and guidance so far.
>>>
>>> I have created a test measure and a job which gets triggered at every 4
>>> mins automatically (have referred to the user guide available on GitHub at
>>> this link
>>> <https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md>
>>> .)
>>>
>>> *Request you to please advise further on below points:*
>>> 1. Since I was able to get ElasticSearch 6.8.x integrated, does it mean
>>> that only ES upto 6.8.x is supported for Griffin as of now? If yes, what
>>> are the plans further? Is there a page from which I could get updates?
>>> 2. I still do not see the metrics available (please refer below
>>> screenshots). Though the measure is now listed in the drop down of *DQ
>>> Metrics* tab. But when I selected the test measure, nothing came up.
>>> 3. At a step in deployment guide it is suggested to check URL: http://<ES
>>> HOST IP>:9200/griffin/accuracy
>>> <http://13.126.127.141:9200/griffin/accuracy> When navigated to this
>>> URL, I get below error. Please advise
>>> *{"error":"Incorrect HTTP method for uri [/griffin/accuracy] and method
>>> [GET], allowed: [POST]","status":405}*
>>> 6. I also noticed that in Data Assets, *demo_src* is owned by Admin
>>> whereas, *demo-tgt* by root. Would that make any difference? If yes,
>>> how to correct it? Reload HIVE data?
>>>
>>> *Screenshots:*
>>> *Data Assets:*
>>> [image: image.png]
>>>
>>> *DQ Metrics (Test Measure selected):*
>>> [image: image.png]
>>>
>>> *Job Triggered multiple times:*
>>> [image: image.png]
>>>
>>> *Metrics page from job directly:*
>>> [image: image.png]
>>>
>>> Thanks and Regards,
>>> Sunil Muniyal
>>>
>>>
>>> On Tue, Sep 8, 2020 at 4:38 PM Sunil Muniyal <[email protected]>
>>> wrote:
>>>
>>>> I am unable to get repos for 6.4.1 instead I found 6.8.x. Will try with
>>>> this version of Elastic Search in sometime.
>>>>
>>>> In the meantime, would it be possible to confirm if 6.4.x or 6.8.x is
>>>> the only supported version for Griffin? Reason I am asking is, the GitHub
>>>> article for griffin deployment points to the latest version of ES.
>>>>
>>>> Thanks and Regards,
>>>> Sunil Muniyal
>>>>
>>>>
>>>> On Tue, Sep 8, 2020 at 4:06 PM Sunil Muniyal <
>>>> [email protected]> wrote:
>>>>
>>>>> I will need to redeploy ElasticSearch, correct?
>>>>>
>>>>> Thanks and Regards,
>>>>> Sunil Muniyal
>>>>>
>>>>>
>>>>> On Tue, Sep 8, 2020 at 4:05 PM William Guo <[email protected]> wrote:
>>>>>
>>>>>> Could you try with this version?
>>>>>> <elasticsearch.version>6.4.1</elasticsearch.version>
>>>>>>
>>>>>>
>>>>>> Thanks,
>>>>>> William
>>>>>>
>>>>>> On Tue, Sep 8, 2020 at 5:59 PM Sunil Muniyal <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> Hi William / Dev group,
>>>>>>>
>>>>>>> I have deployed ES 7.9 - latest version (single node) and the same
>>>>>>> is configured. I also get the default page when hitting
>>>>>>> http://<ES HOST IP>:9200/
>>>>>>>
>>>>>>> Upon creating the griffin configurations using the JSON string given
>>>>>>>
>>>>>>> curl -k -H "Content-Type: application/json" -X PUT http://<replaced 
>>>>>>> with my ES host IP>:9200/griffin \
>>>>>>>  -d '{
>>>>>>>     "aliases": {},
>>>>>>>     "mappings": {
>>>>>>>         "accuracy": {
>>>>>>>             "properties": {
>>>>>>>                 "name": {
>>>>>>>                     "fields": {
>>>>>>>                         "keyword": {
>>>>>>>                             "ignore_above": 256,
>>>>>>>                             "type": "keyword"
>>>>>>>                         }
>>>>>>>                     },
>>>>>>>                     "type": "text"
>>>>>>>                 },
>>>>>>>                 "tmst": {
>>>>>>>                     "type": "date"
>>>>>>>                 }
>>>>>>>             }
>>>>>>>         }
>>>>>>>     },
>>>>>>>     "settings": {
>>>>>>>         "index": {
>>>>>>>             "number_of_replicas": "2",
>>>>>>>             "number_of_shards": "5"
>>>>>>>         }
>>>>>>>     }
>>>>>>> }'
>>>>>>>
>>>>>>>
>>>>>>> *I get below error:*
>>>>>>>
>>>>>>> *{"error":{"root_cause":[{"type":"mapper_parsing_exception","reason":"Root
>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>> type=text},
>>>>>>> tmst={type=date}}}]"}],"type":"mapper_parsing_exception","reason":"Failed
>>>>>>> to parse mapping [_doc]: Root mapping definition has unsupported
>>>>>>> parameters:  [accuracy :
>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>> type=text},
>>>>>>> tmst={type=date}}}]","caused_by":{"type":"mapper_parsing_exception","reason":"Root
>>>>>>> mapping definition has unsupported parameters:  [accuracy :
>>>>>>> {properties={name={fields={keyword={ignore_above=256, type=keyword}},
>>>>>>> type=text}, tmst={type=date}}}]"}},"status":400}*
>>>>>>>
>>>>>>> Seems like the JSON string is missing some values or is incorrectly
>>>>>>> provided.
>>>>>>>
>>>>>>> Would be great if you could please help.
>>>>>>>
>>>>>>> Thanks and Regards,
>>>>>>> Sunil Muniyal
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Sep 7, 2020 at 8:16 PM Sunil Muniyal <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Thank you for the response, William.
>>>>>>>>
>>>>>>>> I have started preparing for ES deployment and should attempt the
>>>>>>>> same tomorrow.
>>>>>>>>
>>>>>>>> In the meantime, I will also wait for the Dev team in case they
>>>>>>>> have any additional inputs.
>>>>>>>>
>>>>>>>> Thanks and Regards,
>>>>>>>> Sunil Muniyal
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Sep 7, 2020 at 8:06 PM William Guo <[email protected]>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>> will need to:
>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>> 2. Update application.properties to include ES details and create
>>>>>>>>> ES index
>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>
>>>>>>>>> *Right, you need to package es env configuration into your jar.*
>>>>>>>>>
>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>
>>>>>>>>> *No*
>>>>>>>>>
>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>> 1. What are the exact ports requirements (internal and external)?
>>>>>>>>> *check log and make sure all extra connections in properties can
>>>>>>>>> accessible*
>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>> *no*
>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>> *java 1.8*
>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what
>>>>>>>>> are the dependencies or additional configurations needed?
>>>>>>>>> *Should no extra dependencies, except those transitive
>>>>>>>>> dependencies incurred by spark and hadoop.*
>>>>>>>>>
>>>>>>>>> On Mon, Sep 7, 2020 at 6:42 PM Sunil Muniyal <
>>>>>>>>> [email protected]> wrote:
>>>>>>>>>
>>>>>>>>>> Ohh ok.
>>>>>>>>>>
>>>>>>>>>> If dev confirms it to be mandatory, as I understand correct, I
>>>>>>>>>> will need to:
>>>>>>>>>> 1. Deploy and Configure ES
>>>>>>>>>> 2. Update application.properties to include ES details and create
>>>>>>>>>> ES index
>>>>>>>>>> 3. Rebuild Maven package and rerun the Griffin service
>>>>>>>>>>
>>>>>>>>>> There is no need to reload the data into Hadoop (Hive), correct?
>>>>>>>>>>
>>>>>>>>>> On a side note, is there any other documentation of Griffin
>>>>>>>>>> available or underway which would help to get below details while
>>>>>>>>>> integrating it with Cloudera Hadoop?
>>>>>>>>>> 1. What are the exact ports requirements (internal and external)?
>>>>>>>>>> 2. Which all packages will be required?
>>>>>>>>>> 3. Any Java dependencies?
>>>>>>>>>> 4. If we have Cloudera Hadoop cluster kerberized (secured), what
>>>>>>>>>> are the dependencies or additional configurations needed?
>>>>>>>>>>
>>>>>>>>>> I know some of the above information can be fetched from the
>>>>>>>>>> deployment guide on Github. However, checking if any other formal
>>>>>>>>>> documentation has been made available for the same?
>>>>>>>>>>
>>>>>>>>>> Thanks and Regards,
>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Sep 7, 2020 at 4:05 PM William Guo <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> cc dev for double checking.
>>>>>>>>>>>
>>>>>>>>>>> Measure will emit metrics and store them in elastic, UI fetch
>>>>>>>>>>> those metrics from elastic.
>>>>>>>>>>> So elastic should be mandatory.
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> William
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:32 PM Sunil Muniyal <
>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Thank you for the quick response, William.
>>>>>>>>>>>>
>>>>>>>>>>>> I have not configured ElasticSearch since it is not deployed.
>>>>>>>>>>>>
>>>>>>>>>>>> In the application.properties, I just added the dummy
>>>>>>>>>>>> information (as below) just to pass the validation test and get 
>>>>>>>>>>>> Griffin up
>>>>>>>>>>>> and running.
>>>>>>>>>>>>
>>>>>>>>>>>> # elasticsearch
>>>>>>>>>>>> # elasticsearch.host = <IP>
>>>>>>>>>>>> # elasticsearch.port = <elasticsearch rest port>
>>>>>>>>>>>> # elasticsearch.user = user
>>>>>>>>>>>> # elasticsearch.password = password
>>>>>>>>>>>> elasticsearch.host=localhost
>>>>>>>>>>>> elasticsearch.port=9200
>>>>>>>>>>>> elasticsearch.scheme=http
>>>>>>>>>>>>
>>>>>>>>>>>> Is ElasticSearch a mandatory requirement to use Griffin?
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Sep 7, 2020 at 3:58 PM William Guo <[email protected]>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Could you check whether ES has been injected with those
>>>>>>>>>>>>> metrics or not?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 6:23 PM Sunil Muniyal <
>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I was able to bypass this error by entering the default field
>>>>>>>>>>>>>> values for LDAP, ElasticSearch and Livy in 
>>>>>>>>>>>>>> application.properties and
>>>>>>>>>>>>>> successfully get Griffin running.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> By following the below article, I have created a test measure
>>>>>>>>>>>>>> and then a job which triggers that measure.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/ui/user-guide.md
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Have allowed the job to get triggered multiple times,
>>>>>>>>>>>>>> however, still i can't see anything in metrics related to the 
>>>>>>>>>>>>>> job. Neither
>>>>>>>>>>>>>> I see anything in *health *or *mydashboard* tabs. Also, if
>>>>>>>>>>>>>> you notice in the screenshot below, being in the *DQ Metrics*
>>>>>>>>>>>>>> tab, I still do not see the created measure in the drop down 
>>>>>>>>>>>>>> list.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Test job executed multiple times:*
>>>>>>>>>>>>>> [image: image.png]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Please advise if anything is mis-configured.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Sep 7, 2020 at 12:40 PM Sunil Muniyal <
>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hello William,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thank you for the reply.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> This helped, actually i had missed to add the property in
>>>>>>>>>>>>>>> application.properties.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Now the other challenge is, along with ES and Livy, I am
>>>>>>>>>>>>>>> also not using LDAP and it is hitting the error *unable to
>>>>>>>>>>>>>>> resolve ldap.url property.* Of Course it will, since the
>>>>>>>>>>>>>>> property is not configured.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Please suggest.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Sun, Sep 6, 2020 at 7:26 PM William Guo <[email protected]>
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> hi Sunil Muniyal,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Could you check this property in your griffin properties
>>>>>>>>>>>>>>>> file?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> internal.event.listeners
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> William
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Thu, Sep 3, 2020 at 11:05 PM Sunil Muniyal <
>>>>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I am attempting to integrate Griffin with Cloudera Hadoop
>>>>>>>>>>>>>>>>> by following below article:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> <https://github.com/apache/griffin/blob/master/griffin-doc/deploy/deploy-guide.md>I
>>>>>>>>>>>>>>>>> have followed everything as instructed, apart from below 
>>>>>>>>>>>>>>>>> things:
>>>>>>>>>>>>>>>>> 1. Using Cloudera Hadoop 5.15 and relevant configurations
>>>>>>>>>>>>>>>>> instead of Apache Hadoop
>>>>>>>>>>>>>>>>> 2. Not using Elastic search as it is not applicable
>>>>>>>>>>>>>>>>> 3. Did not use Livy as it is not applicable.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Maven build is successful and has got 2 jars at
>>>>>>>>>>>>>>>>> service/target and measure/target which I have uploaded to 
>>>>>>>>>>>>>>>>> HDFS.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> However, *starting griffin-service.jar using nohup
>>>>>>>>>>>>>>>>> command* is failing with below error:
>>>>>>>>>>>>>>>>> *Caused by: java.lang.IllegalArgumentException: Could not
>>>>>>>>>>>>>>>>> resolve placeholder 'internal.event.listeners' in string value
>>>>>>>>>>>>>>>>> "#{'${internal.event.listeners}'.split(',')}"*
>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.parseStringValue(PropertyPlaceholderHelper.java:174)
>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>> org.springframework.util.PropertyPlaceholderHelper.replacePlaceholders(PropertyPlaceholderHelper.java:126)
>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>> *        at
>>>>>>>>>>>>>>>>> org.springframework.core.env.AbstractPropertyResolver.doResolvePlaceholders(AbstractPropertyResolver.java:236)
>>>>>>>>>>>>>>>>> ~[spring-core-4.3.6.RELEASE.jar!/:4.3.6.RELEASE]*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I have tried to search a lot of articles with no luck.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Would be great if someone could help me to fix this.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Also, attached is the output of nohup command that was
>>>>>>>>>>>>>>>>> written in service.out.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks and Regards,
>>>>>>>>>>>>>>>>> Sunil Muniyal
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>>>>>>>> To unsubscribe, e-mail:
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>> For additional commands, e-mail:
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [email protected]
> For additional commands, e-mail: [email protected]

Reply via email to