Joe, if you're using the CA from NiFi 1.10, this is a known issue (already fixed in master, I don't have the JIRA handy). You can try with the CA from toolkit 1.9.2 if that's not too complex in your environment.
Le lun. 30 déc. 2019 à 21:29, Joe Gresock <jgres...@gmail.com> a écrit : > Looks like it works when I exclude the --subjectAlternativeNames > parameter, whereas it fails with the null error when I specify this > parameter. > For example, this works: > /opt/nifi/nifi-toolkit-current/bin/tls-toolkit.sh client -c nifi-ca-cs -t > a5s4d545h48s6d87fa6sd7g45f6g4fga84sd6a8e7f6ga786a --dn "CN=$(hostname > -f),OU=NIFI" > > But this doesn't: > /opt/nifi/nifi-toolkit-current/bin/tls-toolkit.sh client -c nifi-ca-cs -t > a5s4d545h48s6d87fa6sd7g45f6g4fga84sd6a8e7f6ga786a --dn "CN=$(hostname > -f),OU=NIFI" --subjectAlternativeNames "<my proxy server's hostname>" > > I'll dig further to see if there is any way I can get the SAN to work. > > On Mon, Dec 30, 2019 at 4:51 PM Joe Gresock <jgres...@gmail.com> wrote: > >> Thanks for the suggestion, apparently there was a typo on my nifi-ca-cs >> service that caused the pod not to be matched. Unfortunately, even once I >> corrected this, I'm still getting the "Service client error: null". >> >> I'll keep working on it. Thanks for your help! >> >> On Sat, Dec 28, 2019 at 5:53 AM Swarup Karavadi <r...@swazza.io> wrote: >> >>> Hey Joe, >>> >>> I'm glad that the article was of some help. I, unfortunately, do not >>> recollect running into that specific error scenario. At the risk of stating >>> the obvious, can you check if the nifi-ca-cs service is reachable from your >>> nifi pod? >>> >>> On the CRD bit, we are considering going this route with NiFi stateless. >>> We are experimenting to see if we can have something like "NiFi K" that can >>> be inline with the way "Camel K" works. Based on our current (limited) >>> understanding of NiFi stateless, we think the CRD approach will help scale >>> NiFi stateless horizontally with much more ease. >>> >>> Cheers, >>> Swarup. >>> >>> On Sat, Dec 28, 2019 at 4:32 AM Mike Thomsen <mikerthom...@gmail.com> >>> wrote: >>> >>>> If you don't see a value when you run "echo %JAVA_HOME%" then you need >>>> to check to see if it was set globally in Windows, and if it was, you need >>>> to open a new command shell. >>>> >>>> On Mon, Oct 21, 2019 at 12:37 PM <crzy...@gmail.com> wrote: >>>> >>>>> Any suggestions? >>>>> >>>>> >>>>> >>>>> I downloaded NiFi, but when I run the runnifi from the bin folder >>>>> nothing happens. I get the following message. The JAVA_HOME environment >>>>> variable is not defined correctly. I downloaded the latest JRE, but still >>>>> get the same error message. >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> *From:* Swarup Karavadi <r...@swazza.io> >>>>> *Sent:* Monday, October 21, 2019 12:19 PM >>>>> *To:* users@nifi.apache.org >>>>> *Cc:* jgres...@gmail.com >>>>> *Subject:* Re: NiFi Kubernetes question >>>>> >>>>> >>>>> >>>>> If you are hosting on the cloud, I'd recommend going for dedicated >>>>> worker nodes for the NiFi cluster. There might be rare (or not) occasions >>>>> when a worker node is under high load and needs to evict pods. If your >>>>> NiFi >>>>> deployment's pod disruption budget allows for eviction of pods then there >>>>> are always chances that an evicted NiFi pod can be rescheduled on a >>>>> different node that is tainted (tainted because the node may not meet the >>>>> pod's volume affinity requirements). Your best case scenario when this >>>>> happens is that the pod will keep getting rescheduled on different nodes >>>>> until it starts up again. The worst case scenario is that it'll be stuck >>>>> in >>>>> a CrashLoopBackoff limbo. >>>>> >>>>> >>>>> >>>>> Disclaimer - I speak from my experience on a non production >>>>> environment. Our NiFi clusters will be deployed to a production k8s >>>>> environment in a few weeks from now. I am only sharing some learnings I've >>>>> had w.r.t. k8s statefulsets along the way. >>>>> >>>>> >>>>> >>>>> Hope this helps, >>>>> >>>>> Swarup. >>>>> >>>>> >>>>> >>>>> On Mon, Oct 21, 2019, 9:32 PM Wyllys Ingersoll < >>>>> wyllys.ingers...@keepertech.com> wrote: >>>>> >>>>> >>>>> >>>>> We had success running a 3-node cluster in kubernetes using modified >>>>> configuration scripts from the AlexJones github repo - >>>>> https://github.com/AlexsJones/nifi >>>>> >>>>> Ours is on an internal bare-metal k8s lab configuration, not in a >>>>> public cloud at this time, but the basics are the same either way. >>>>> >>>>> >>>>> >>>>> - setup nifi as a stateful set so you can scale up or down as needed. >>>>> When a pod fails, k8s will spawn another to take its place and zookeeper >>>>> will manage the election of the master during transitions. >>>>> >>>>> - manage your certs as K8S secrets. >>>>> >>>>> - you also need to also have a stateful set of zookeeper pods for >>>>> managing the nifi servers. >>>>> >>>>> - use persistent volume mounts to hold the flowfile, database, >>>>> content, and provenance _repository directories >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Mon, Oct 21, 2019 at 11:21 AM Joe Gresock <jgres...@gmail.com> >>>>> wrote: >>>>> >>>>> Apologies if this has been answered on the list already.. >>>>> >>>>> >>>>> >>>>> Does anyone have knowledge of the latest in the realm of nifi >>>>> kubernetes support? I see some pages like >>>>> https://hub.helm.sh/charts/cetic/nifi, and >>>>> https://github.com/AlexsJones/nifi but am unsure which example to >>>>> pick to start with. >>>>> >>>>> >>>>> >>>>> I'm curious how well kubernetes maintains the nifi cluster state with >>>>> pod failures. I.e., do any of the k8s implementations play well with the >>>>> nifi cluster list so that we don't have dangling downed nodes in the >>>>> cluster? Also, I'm wondering how certs are managed in a secured cluster. >>>>> >>>>> >>>>> >>>>> Appreciate any nudge in the right direction, >>>>> >>>>> Joe >>>>> >>>>> >>>>> >>>>> On Mon, Oct 21, 2019, 9:32 PM Wyllys Ingersoll < >>>>> wyllys.ingers...@keepertech.com> wrote: >>>>> >>>>> >>>>> >>>>> We had success running a 3-node cluster in kubernetes using modified >>>>> configuration scripts from the AlexJones github repo - >>>>> https://github.com/AlexsJones/nifi >>>>> >>>>> Ours is on an internal bare-metal k8s lab configuration, not in a >>>>> public cloud at this time, but the basics are the same either way. >>>>> >>>>> >>>>> >>>>> - setup nifi as a stateful set so you can scale up or down as needed. >>>>> When a pod fails, k8s will spawn another to take its place and zookeeper >>>>> will manage the election of the master during transitions. >>>>> >>>>> - manage your certs as K8S secrets. >>>>> >>>>> - you also need to also have a stateful set of zookeeper pods for >>>>> managing the nifi servers. >>>>> >>>>> - use persistent volume mounts to hold the flowfile, database, >>>>> content, and provenance _repository directories >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Mon, Oct 21, 2019 at 11:21 AM Joe Gresock <jgres...@gmail.com> >>>>> wrote: >>>>> >>>>> Apologies if this has been answered on the list already.. >>>>> >>>>> >>>>> >>>>> Does anyone have knowledge of the latest in the realm of nifi >>>>> kubernetes support? I see some pages like >>>>> https://hub.helm.sh/charts/cetic/nifi, and >>>>> https://github.com/AlexsJones/nifi but am unsure which example to >>>>> pick to start with. >>>>> >>>>> >>>>> >>>>> I'm curious how well kubernetes maintains the nifi cluster state with >>>>> pod failures. I.e., do any of the k8s implementations play well with the >>>>> nifi cluster list so that we don't have dangling downed nodes in the >>>>> cluster? Also, I'm wondering how certs are managed in a secured cluster. >>>>> >>>>> >>>>> >>>>> Appreciate any nudge in the right direction, >>>>> >>>>> Joe >>>>> >>>>> >> >> -- >> Be on your guard; stand firm in the faith; be courageous; be strong. Do >> everything in love. -*1 Corinthians 16:13-14* >> > > > -- > Be on your guard; stand firm in the faith; be courageous; be strong. Do > everything in love. -*1 Corinthians 16:13-14* >