If you don't see a value when you run "echo %JAVA_HOME%" then you need to
check to see if it was set globally in Windows, and if it was, you need to
open a new command shell.

On Mon, Oct 21, 2019 at 12:37 PM <[email protected]> wrote:

> Any suggestions?
>
>
>
> I downloaded NiFi, but when I run the runnifi from the bin folder nothing
> happens. I get the following message. The JAVA_HOME environment variable is
> not defined correctly. I downloaded the latest JRE, but still get the same
> error message.
>
>
>
>
>
> *From:* Swarup Karavadi <[email protected]>
> *Sent:* Monday, October 21, 2019 12:19 PM
> *To:* [email protected]
> *Cc:* [email protected]
> *Subject:* Re: NiFi Kubernetes question
>
>
>
> If you are hosting on the cloud, I'd recommend going for dedicated worker
> nodes for the NiFi cluster. There might be rare (or not) occasions when a
> worker node is under high load and needs to evict pods. If your NiFi
> deployment's pod disruption budget allows for eviction of pods then there
> are always chances that an evicted NiFi pod can be rescheduled on a
> different node that is tainted (tainted because the node may not meet the
> pod's volume affinity requirements). Your best case scenario when this
> happens is that the pod will keep getting rescheduled on different nodes
> until it starts up again. The worst case scenario is that it'll be stuck in
> a CrashLoopBackoff limbo.
>
>
>
> Disclaimer - I speak from my experience on a non production environment.
> Our NiFi clusters will be deployed to a production k8s environment in a few
> weeks from now. I am only sharing some learnings I've had w.r.t. k8s
> statefulsets along the way.
>
>
>
> Hope this helps,
>
> Swarup.
>
>
>
> On Mon, Oct 21, 2019, 9:32 PM Wyllys Ingersoll <
> [email protected]> wrote:
>
>
>
> We had success running  a 3-node cluster in kubernetes using modified
> configuration scripts from the AlexJones github repo -
> https://github.com/AlexsJones/nifi
>
> Ours is on an internal bare-metal k8s lab configuration, not in a public
> cloud at this time, but the basics are the same either way.
>
>
>
> - setup nifi as a stateful set so you can scale up or down as needed. When
> a pod fails, k8s will spawn another to take its place and zookeeper will
> manage the election of the master during transitions.
>
> - manage your certs as K8S secrets.
>
> - you also need to also have a stateful set of zookeeper pods for managing
> the nifi servers.
>
> - use persistent volume mounts to hold the flowfile, database, content,
> and provenance _repository directories
>
>
>
>
>
>
>
> On Mon, Oct 21, 2019 at 11:21 AM Joe Gresock <[email protected]> wrote:
>
> Apologies if this has been answered on the list already..
>
>
>
> Does anyone have knowledge of the latest in the realm of nifi kubernetes
> support?  I see some pages like https://hub.helm.sh/charts/cetic/nifi,
> and https://github.com/AlexsJones/nifi but am unsure which example to
> pick to start with.
>
>
>
> I'm curious how well kubernetes maintains the nifi cluster state with pod
> failures.  I.e., do any of the k8s implementations play well with the nifi
> cluster list so that we don't have dangling downed nodes in the cluster?
> Also, I'm wondering how certs are managed in a secured cluster.
>
>
>
> Appreciate any nudge in the right direction,
>
> Joe
>
>
>
> On Mon, Oct 21, 2019, 9:32 PM Wyllys Ingersoll <
> [email protected]> wrote:
>
>
>
> We had success running  a 3-node cluster in kubernetes using modified
> configuration scripts from the AlexJones github repo -
> https://github.com/AlexsJones/nifi
>
> Ours is on an internal bare-metal k8s lab configuration, not in a public
> cloud at this time, but the basics are the same either way.
>
>
>
> - setup nifi as a stateful set so you can scale up or down as needed. When
> a pod fails, k8s will spawn another to take its place and zookeeper will
> manage the election of the master during transitions.
>
> - manage your certs as K8S secrets.
>
> - you also need to also have a stateful set of zookeeper pods for managing
> the nifi servers.
>
> - use persistent volume mounts to hold the flowfile, database, content,
> and provenance _repository directories
>
>
>
>
>
>
>
> On Mon, Oct 21, 2019 at 11:21 AM Joe Gresock <[email protected]> wrote:
>
> Apologies if this has been answered on the list already..
>
>
>
> Does anyone have knowledge of the latest in the realm of nifi kubernetes
> support?  I see some pages like https://hub.helm.sh/charts/cetic/nifi,
> and https://github.com/AlexsJones/nifi but am unsure which example to
> pick to start with.
>
>
>
> I'm curious how well kubernetes maintains the nifi cluster state with pod
> failures.  I.e., do any of the k8s implementations play well with the nifi
> cluster list so that we don't have dangling downed nodes in the cluster?
> Also, I'm wondering how certs are managed in a secured cluster.
>
>
>
> Appreciate any nudge in the right direction,
>
> Joe
>
>

Reply via email to