Hi!
First of all, Ignite 3 uses separate node and cluster configurations. I
presume you're talking about node configuration?
Let's say you want to run two nodes from a single distribution.
Unfortunately, Ignite 3 is not very flexible in this case and you will need
to manually edit configuration files.
I would probably copy-paste vars.env files to something like vars1.env and
vars2.env.
Then in each file I'd change NODE_NAME to unique name, let's say node1 and
node2.
Then as you already did - modify CONFIG_FILE to point to the configuration
file for each node.
Then I would copy-paste ignite3db files as well and change the .
${IGNITE_HOME}/etc/vars.env line to point to the vars.env file for the
specific node.
I think we could do better in this regard, let me create a ticket for
improvement and let me know if the above suggestion end up working.
Thanks,
Vadim
On Thu, Mar 12, 2026 at 3:08 PM Charlin S <[email protected]> wrote:
> Hi All,
>
> I am currently working with two different cluster setup configurations
> (using different port numbers) for Ignite 2.17. I am able to start both
> clusters sequentially so they run simultaneously using the following
> commands:
>
> 1. ignite.sh configfile1.xml
> 2. ignite.sh configfile2.xml
>
> For Ignite 3.1, I have migrated these files to configfile1.conf and
> configfile2.conf using the Ignite migration tool and placed them in the
> ignite3-db-3.1.0/etc/ folder. I successfully started the first cluster by
> editing vars.env to set CONFIG_FILE=${CONF_DIR}/configfile1.conf.
>
> Could you please advise on how to start the second cluster using
> configfile2.conf in this new environment?
>
> Thanks & Regards,
> Charlin S
>