Re: [Wien] parallel ssh error

2019-09-30 Thread Gavin Abo

Good to see.  As you know, "hup: Command not found" can be ignored:

http://zeus.theochem.tuwien.ac.at/pipermail/wien/2011-April/014484.html
http://zeus.theochem.tuwien.ac.at/pipermail/wien/2010-September/013598.html
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg11131.html

On 9/30/2019 10:35 AM, Indranil mal wrote:
Thank you Sir for your instantaneous support. Now it is working 
smoothly only with hup: Command not found

___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-30 Thread Indranil mal
 Thank you Sir for your instantaneous support. Now it is working smoothly
only with  hup: Command not found.



On Sun, Sep 29, 2019 at 6:32 PM Gavin Abo  wrote:

> Checking with "which lapw1c" on each node (vlsi1, vlsi2, vlsi3, and vlsi4)
> is a good idea.  However, since WIENROOT is (blank) [1], it probably won't
> work until that is resolved.
>
> It was mentioned that the WIEN2k .bashrc block was setup on each node by
> running userconfig [2].  So it definitely seems strange that WIENROOT is
> (blank) on the client nodes since I would think it would work if both
> WIENROOT and PATH are both defined from userconfig in .bashrc:
>
> username@computername:~$ ssh vlsi1
> ...
> username@computername:~$ cd ~/WIEN2k
> username@computername:~/WIEN2k$ which lapw1c
> username@computername:~/WIEN2k$ grep "export WIENROOT" ~/.bashrc
> username@computername:~/WIEN2k$ grep "export PATH" ~/.bashrc
> username@computername:~/WIEN2k$ ./userconfig
> ...
> username@computername:~/WIEN2k$ grep "export WIENROOT" ~/.bashrc
> export WIENROOT=/servernode1
> username@computername:~/WIEN2k$ grep "export PATH" ~/.bashrc
> export
> PATH=$WIENROOT:$STRUCTEDIT_PATH:$WIENROOT/SRC_IRelast/script-elastic:$PATH:.
> export PATH=$PATH:$WIENROOT:.
> username@computername:~/WIEN2k$ source ~/.bashrc
> username@computername:~/WIEN2k$ which lapw1c
> /home/username/WIEN2k/lapw1c
> username@computername:~/WIEN2k$ exit
> logout
> Connection to vlsi1 closed.
>
> Though, I suppose if something like a conf file [3] was setup by the user
> to override .bashrc or a job queue scheduler system is in use [4] it might
> also cause the issue.
> [1]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19052.html
> [2]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html
> [3]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg08016.html
> [4]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg15985.html
>
> On 9/29/2019 6:11 AM, Laurence Marks wrote:
>
> What does
>
> ssh vlsi1 which lapw1c
> give, what does "cat *.error" give in the case directory?
> _
> Professor Laurence Marks
> "Research is to see what everybody else has seen, and to think what nobody
> else has thought", Albert Szent-Gyorgi
> www.numis.northwestern.edu
>
> On Sun, Sep 29, 2019, 01:17 Indranil mal  wrote:
>
>> Now echo $WIENROOT is giving the $WIENROOT location.
>>
>> echo $WIENROOT/lapw*
>>
>> /home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
>> /home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
>> /home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
>> /home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
>> /home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
>> /home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
>> /home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
>> /home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
>> /home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
>> /home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
>> /home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
>> /home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
>> /home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
>> /home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
>> /home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
>> /home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
>> /home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw
>>
>> ssh vlsi1 'echo $WIENROOT/lapw*'
>>
>> /home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
>> /home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
>> /home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
>> /home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
>> /home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
>> /home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
>> /home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
>> /home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
>> /home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
>> /home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
>> /home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
>> /home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
>> /home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
>> /home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
>> /home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
>> /home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
>> /home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw
>>
>>
>> However getting the same error
>>
>>
>> >   stop error
>>
>> grep: *scf1*: No such file or directory
>> cp: cannot stat '.in.tmp': No such file or directory
>> FERMI - Error
>> grep: *scf1*: No such file or directory
>> Parallel.scf1_1: No such file or directory.

Re: [Wien] parallel ssh error

2019-09-30 Thread Gavin Abo
An additional comment, /home/username/WIEN2k (or ~/WIEN2k) is where I 
have WIEN2k installed.  Whereas, you have installed WIEN2k at 
/servernode1 [1].  In the examples of my previous posts (e.g. [2]) you 
might find some typographical errors were I forget to replace my 
/home/username/WIEN2k with your /servernode1.


It is best to have WIEN2k setup at a common path location on all nodes 
[3,4] (i.e., your vlsi1-vsli4).  Therefore, I recommend to not have 
WIEN2k at other locations among your system nodes like the 
/home/username/WIEN2K as you mentioned below.


[1] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html
[2] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19061.html
[3] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg17988.html
[4] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg09229.html


On 9/30/2019 12:59 AM, Peter Blaha wrote:
So there is progress as now the environment seems to be accepted in 
the remote shell.


lapw1para (called by x_lapw, which is called by run_lapw -p) creates 
the splitted klists-files (case.klist_1,...) and def files 
lapw1_1.def,...


It uses the $cwd variable and executes basically:

ssh vlsi1 "cd $cwd; lapw1c lapw1_1.def "

Does this work on your computers ?



On 9/29/19 7:16 PM, Indranil mal wrote:

Now echo $WIENROOT is giving the $WIENROOT location.

echo $WIENROOT/lapw*

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw


ssh vlsi1 'echo $WIENROOT/lapw*'

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw



However getting the same error





  stop error


grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
  LAPW0 END
hup: Command not found.


and lapw2 error file

  'LAPW2' - can't open unit: 30
  'LAPW2' -        filename: Parallel.energy_1
**  testerror: Error in Parallel LAPW2



On Sat, Sep 28, 2019 at 11:58 PM Gavin Abo > wrote:


    The "sudo service sshd restart" step, which I forgot to copy and
    paste, that is missing is corrected below.

    On 9/28/2019 12:18 PM, Gavin Abo wrote:


    After you set both "SendEnv *" and "AcceptEnv *", did you restart
    the sshd service [1]?  The following illustrates steps that might
    help you verify that WIENROOT appears on a remote vlsi node:

    username@computername:~$ echo $WIENROOT

    

Re: [Wien] parallel ssh error

2019-09-30 Thread Peter Blaha
So there is progress as now the environment seems to be accepted in the 
remote shell.


lapw1para (called by x_lapw, which is called by run_lapw -p) creates the 
splitted klists-files (case.klist_1,...) and def files lapw1_1.def,...


It uses the $cwd variable and executes basically:

ssh vlsi1 "cd $cwd; lapw1c lapw1_1.def "

Does this work on your computers ?



On 9/29/19 7:16 PM, Indranil mal wrote:

Now echo $WIENROOT is giving the $WIENROOT location.

echo $WIENROOT/lapw*

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw


ssh vlsi1 'echo $WIENROOT/lapw*'

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw



However getting the same error





  stop error


grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
  LAPW0 END
hup: Command not found.


and lapw2 error file

  'LAPW2' - can't open unit: 30
  'LAPW2' -        filename: Parallel.energy_1
**  testerror: Error in Parallel LAPW2



On Sat, Sep 28, 2019 at 11:58 PM Gavin Abo > wrote:


The "sudo service sshd restart" step, which I forgot to copy and
paste, that is missing is corrected below.

On 9/28/2019 12:18 PM, Gavin Abo wrote:


After you set both "SendEnv *" and "AcceptEnv *", did you restart
the sshd service [1]?  The following illustrates steps that might
help you verify that WIENROOT appears on a remote vlsi node:

username@computername:~$ echo $WIENROOT

username@computername:~$ export WIENROOT=/servernode1
username@computername:~$ echo $WIENROOT
/servernode1
username@computername:~$ ssh vlsi
Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 4.15.0-64-generic x86_64)
...
Last login: Sat Sep 28 12:04:07 2019 from xxx.x.x.x
username@computername:~$ echo $WIENROOT

username@computername:~$ exit
logout
Connection to vlsi closed.
username@computername:~$ sudo gedit /etc/ssh/ssh_config
[sudo] password for username:

username@computername:~$ sudo gedit /etc/ssh/sshd_config

username@computername:~$ grep SendEnv /etc/ssh/ssh_config
    SendEnv LANG LC_* WIENROOT
username@computername:~$ grep AcceptEnv /etc/ssh/sshd_config
AcceptEnv LANG LC_* WIENROOT


    username@computername:~$ sudo service sshd restart


username@computername:~$ ssh vlsi
...
username@computername:~$ echo $WIENROOT

Re: [Wien] parallel ssh error

2019-09-29 Thread Gavin Abo
Checking with "which lapw1c" on each node (vlsi1, vlsi2, vlsi3, and 
vlsi4) is a good idea.  However, since WIENROOT is (blank) [1], it 
probably won't work until that is resolved.


It was mentioned that the WIEN2k .bashrc block was setup on each node by 
running userconfig [2]. So it definitely seems strange that WIENROOT is 
(blank) on the client nodes since I would think it would work if both 
WIENROOT and PATH are both defined from userconfig in .bashrc:


username@computername:~$ ssh vlsi1
...
username@computername:~$ cd ~/WIEN2k
username@computername:~/WIEN2k$ which lapw1c
username@computername:~/WIEN2k$ grep "export WIENROOT" ~/.bashrc
username@computername:~/WIEN2k$ grep "export PATH" ~/.bashrc
username@computername:~/WIEN2k$ ./userconfig
...
username@computername:~/WIEN2k$ grep "export WIENROOT" ~/.bashrc
export WIENROOT=/servernode1
username@computername:~/WIEN2k$ grep "export PATH" ~/.bashrc
export 
PATH=$WIENROOT:$STRUCTEDIT_PATH:$WIENROOT/SRC_IRelast/script-elastic:$PATH:.

export PATH=$PATH:$WIENROOT:.
username@computername:~/WIEN2k$ source ~/.bashrc
username@computername:~/WIEN2k$ which lapw1c
/home/username/WIEN2k/lapw1c
username@computername:~/WIEN2k$ exit
logout
Connection to vlsi1 closed.

Though, I suppose if something like a conf file [3] was setup by the 
user to override .bashrc or a job queue scheduler system is in use [4] 
it might also cause the issue.


[1] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19052.html
[2] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html
[3] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg08016.html
[4] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg15985.html


On 9/29/2019 6:11 AM, Laurence Marks wrote:

What does

ssh vlsi1 which lapw1c
give, what does "cat *.error" give in the case directory?
_
Professor Laurence Marks
"Research is to see what everybody else has seen, and to think what 
nobody else has thought", Albert Szent-Gyorgi

www.numis.northwestern.edu 

On Sun, Sep 29, 2019, 01:17 Indranil mal > wrote:


Now echo $WIENROOT is giving the $WIENROOT location.

echo $WIENROOT/lapw*

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
/home/username/WIEN2K/lapw0para
/home/username/WIEN2K/lapw0para_lapw /home/username/WIEN2K/lapw1
/home/username/WIEN2K/lapw1c /home/username/WIEN2K/lapw1c_mpi
/home/username/WIEN2K/lapw1cpara /home/username/WIEN2K/lapw1_mpi
/home/username/WIEN2K/lapw1para
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
/home/username/WIEN2K/lapw2para
/home/username/WIEN2K/lapw2para_lapw /home/username/WIEN2K/lapw3
/home/username/WIEN2K/lapw3c /home/username/WIEN2K/lapw5
/home/username/WIEN2K/lapw5c /home/username/WIEN2K/lapw7
/home/username/WIEN2K/lapw7c /home/username/WIEN2K/lapwdm
/home/username/WIEN2K/lapwdmc /home/username/WIEN2K/lapwdmcpara
/home/username/WIEN2K/lapwdmpara
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw

ssh vlsi1 'echo $WIENROOT/lapw*'

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
/home/username/WIEN2K/lapw0para
/home/username/WIEN2K/lapw0para_lapw /home/username/WIEN2K/lapw1
/home/username/WIEN2K/lapw1c /home/username/WIEN2K/lapw1c_mpi
/home/username/WIEN2K/lapw1cpara /home/username/WIEN2K/lapw1_mpi
/home/username/WIEN2K/lapw1para
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
/home/username/WIEN2K/lapw2para
/home/username/WIEN2K/lapw2para_lapw /home/username/WIEN2K/lapw3
/home/username/WIEN2K/lapw3c /home/username/WIEN2K/lapw5
/home/username/WIEN2K/lapw5c /home/username/WIEN2K/lapw7
/home/username/WIEN2K/lapw7c /home/username/WIEN2K/lapwdm
/home/username/WIEN2K/lapwdmc /home/username/WIEN2K/lapwdmcpara
/home/username/WIEN2K/lapwdmpara
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw


However getting the same error




>   stop error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found

Re: [Wien] parallel ssh error

2019-09-29 Thread Laurence Marks
What does

ssh vlsi1 which lapw1c
give, what does "cat *.error" give in the case directory?
_
Professor Laurence Marks
"Research is to see what everybody else has seen, and to think what nobody
else has thought", Albert Szent-Gyorgi
www.numis.northwestern.edu

On Sun, Sep 29, 2019, 01:17 Indranil mal  wrote:

> Now echo $WIENROOT is giving the $WIENROOT location.
>
> echo $WIENROOT/lapw*
>
> /home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
> /home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
> /home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
> /home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
> /home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
> /home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
> /home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
> /home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
> /home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
> /home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
> /home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
> /home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
> /home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
> /home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
> /home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
> /home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
> /home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw
>
> ssh vlsi1 'echo $WIENROOT/lapw*'
>
> /home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
> /home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
> /home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
> /home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
> /home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
> /home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
> /home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
> /home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
> /home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
> /home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
> /home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
> /home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
> /home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
> /home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
> /home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
> /home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
> /home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw
>
>
> However getting the same error
>
>
> >   stop error
>
> grep: *scf1*: No such file or directory
> cp: cannot stat '.in.tmp': No such file or directory
> FERMI - Error
> grep: *scf1*: No such file or directory
> Parallel.scf1_1: No such file or directory.
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
> bash: fixerror_lapw: command not found
> bash: lapw1c: command not found
>  LAPW0 END
> hup: Command not found.
>
>
> and lapw2 error file
>
>  'LAPW2' - can't open unit: 30
>
>  'LAPW2' -filename: Parallel.energy_1
>
> **  testerror: Error in Parallel LAPW2
>
>
> On Sat, Sep 28, 2019 at 11:58 PM Gavin Abo  wrote:
>
>> The "sudo service sshd restart" step, which I forgot to copy and paste,
>> that is missing is corrected below.
>> On 9/28/2019 12:18 PM, Gavin Abo wrote:
>>
>> After you set both "SendEnv *" and "AcceptEnv *", did you restart the
>> sshd service [1]?  The following illustrates steps that might help you
>> verify that WIENROOT appears on a remote vlsi node:
>>
>> username@computername:~$ echo $WIENROOT
>>
>> username@computername:~$ export WIENROOT=/servernode1
>> username@computername:~$ echo $WIENROOT
>> /servernode1
>> username@computername:~$ ssh vlsi
>> Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 4.15.0-64-generic x86_64)
>> ...
>> Last login: Sat Sep 28 12:04:07 2019 from xxx.x.x.x
>> username@computername:~$ echo $WIENROOT
>>
>> username@computername:~$ exit
>> logout
>> Connection to vlsi closed.
>> username@computername:~$ sudo gedit /etc/ssh/ssh_config
>> [sudo] password for username:
>>
>> username@computername:~$ sudo gedit /etc/ssh/sshd_config
>>
>> username@computername:~$ grep SendEnv /etc/ssh/ssh_config
>> SendEnv LANG LC_* WIENROOT
>> username@computername:~$ grep AcceptEnv /etc/ssh/sshd_config
>> AcceptEnv LANG LC_* WIENROOT
>>
>>username@computername:~$ sudo service sshd restart
>>
>> username@computername:~$ ssh vlsi
>> ...
>> username@computername:~$ echo $WIENROOT
>> /servernode1
>> username@computername:~$ exit
>>
>> [1]

Re: [Wien] parallel ssh error

2019-09-29 Thread Gavin Abo
I had noticed that ssh vlsi1 'echo $WIENROOT/lapw*' seems to pick up the 
local environment.  Since you are interested in the remote environment, 
make sure you issue them as separate commands [1] for vlsi1, vlsi2, 
vlsi3, and vlsi4:


ssh vlsi1
echo $WIENROOT/lapw*
exit



ssh vlsi4
echo $WIENROOT/lapw*
exit

If the ssh_config and in sshd_config approach is not working. Did you 
also try the alternative .bashrc approach [2] and also check that it is 
not a non-interactive shell login issue [3]?


Check .bashrc locally:

username@computername:~$ grep "export WIENROOT" ~/.bashrc
export WIENROOT=/home/username/WIEN2k

Check .bashrc remotely for vlsi1, vlsi2, vlsi3, and vlsi4:

username@computername:~$ ssh vlsi1
...
username@computername:~$ grep "export WIENROOT" ~/.bashrc
export WIENROOT=/home/username/WIEN2k
username@computername:~$ exit
logout
Connection to vlsi1 closed.

...

username@computername:~$ ssh vlsi4
...
username@computername:~$ grep "export WIENROOT" ~/.bashrc
export WIENROOT=/home/username/WIEN2k
username@computername:~$ exit
logout
Connection to vlsi4 closed.

[1] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19057.html
[2] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19053.html
[3] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg18685.html


On 9/29/2019 11:16 AM, Indranil mal wrote:

Now echo $WIENROOT is giving the $WIENROOT location.

echo $WIENROOT/lapw*

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw


ssh vlsi1 'echo $WIENROOT/lapw*'

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi 
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw 
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c 
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara 
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para 
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2 
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi 
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi 
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw 
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c 
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c 
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c 
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc 
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara 
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso 
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi 
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw



However getting the same error




>   stop error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
  LAPW0 END
hup: Command not found.


and lapw2 error file

 'LAPW2' - can't open unit: 30
 'LAPW2' -        filename: Parallel.energy_1
**  testerror: Error in Parallel LAPW2

___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-29 Thread Indranil mal
Now echo $WIENROOT is giving the $WIENROOT location.

echo $WIENROOT/lapw*

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw

ssh vlsi1 'echo $WIENROOT/lapw*'

/home/username/WIEN2K/lapw0 /home/username/WIEN2K/lapw0_mpi
/home/username/WIEN2K/lapw0para /home/username/WIEN2K/lapw0para_lapw
/home/username/WIEN2K/lapw1 /home/username/WIEN2K/lapw1c
/home/username/WIEN2K/lapw1c_mpi /home/username/WIEN2K/lapw1cpara
/home/username/WIEN2K/lapw1_mpi /home/username/WIEN2K/lapw1para
/home/username/WIEN2K/lapw1para_lapw /home/username/WIEN2K/lapw2
/home/username/WIEN2K/lapw2c /home/username/WIEN2K/lapw2c_mpi
/home/username/WIEN2K/lapw2cpara /home/username/WIEN2K/lapw2_mpi
/home/username/WIEN2K/lapw2para /home/username/WIEN2K/lapw2para_lapw
/home/username/WIEN2K/lapw3 /home/username/WIEN2K/lapw3c
/home/username/WIEN2K/lapw5 /home/username/WIEN2K/lapw5c
/home/username/WIEN2K/lapw7 /home/username/WIEN2K/lapw7c
/home/username/WIEN2K/lapwdm /home/username/WIEN2K/lapwdmc
/home/username/WIEN2K/lapwdmcpara /home/username/WIEN2K/lapwdmpara
/home/username/WIEN2K/lapwdmpara_lapw /home/username/WIEN2K/lapwso
/home/username/WIEN2K/lapwsocpara /home/username/WIEN2K/lapwso_mpi
/home/username/WIEN2K/lapwsopara /home/username/WIEN2K/lapwsopara_lapw


However getting the same error


>   stop error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
 LAPW0 END
hup: Command not found.


and lapw2 error file

 'LAPW2' - can't open unit: 30

 'LAPW2' -filename: Parallel.energy_1

**  testerror: Error in Parallel LAPW2


On Sat, Sep 28, 2019 at 11:58 PM Gavin Abo  wrote:

> The "sudo service sshd restart" step, which I forgot to copy and paste,
> that is missing is corrected below.
> On 9/28/2019 12:18 PM, Gavin Abo wrote:
>
> After you set both "SendEnv *" and "AcceptEnv *", did you restart the sshd
> service [1]?  The following illustrates steps that might help you verify
> that WIENROOT appears on a remote vlsi node:
>
> username@computername:~$ echo $WIENROOT
>
> username@computername:~$ export WIENROOT=/servernode1
> username@computername:~$ echo $WIENROOT
> /servernode1
> username@computername:~$ ssh vlsi
> Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 4.15.0-64-generic x86_64)
> ...
> Last login: Sat Sep 28 12:04:07 2019 from xxx.x.x.x
> username@computername:~$ echo $WIENROOT
>
> username@computername:~$ exit
> logout
> Connection to vlsi closed.
> username@computername:~$ sudo gedit /etc/ssh/ssh_config
> [sudo] password for username:
>
> username@computername:~$ sudo gedit /etc/ssh/sshd_config
>
> username@computername:~$ grep SendEnv /etc/ssh/ssh_config
> SendEnv LANG LC_* WIENROOT
> username@computername:~$ grep AcceptEnv /etc/ssh/sshd_config
> AcceptEnv LANG LC_* WIENROOT
>
>username@computername:~$ sudo service sshd restart
>
> username@computername:~$ ssh vlsi
> ...
> username@computername:~$ echo $WIENROOT
> /servernode1
> username@computername:~$ exit
>
> [1]
> https://askubuntu.com/questions/462968/take-changes-in-file-sshd-config-file-without-server-reboot
> On 9/28/2019 11:22 AM, Indranil mal wrote:
>
> Sir I have tried with " SetEnv * " Still nothing is coming with echo
>  commad and user name by mistake I posted wrong Otherwise no issue with
> user name and I have set the parallel options file taksset "no" and remote
> options are 1 1 in server and client machines.
>
>
> On Sat, 28 Sep 2019 11:36 Gavin Abo,  wrote:
>
>> Respected Sir, In my 

Re: [Wien] parallel ssh error

2019-09-28 Thread Gavin Abo
The "sudo service sshd restart" step, which I forgot to copy and paste, 
that is missing is corrected below.


On 9/28/2019 12:18 PM, Gavin Abo wrote:


After you set both "SendEnv *" and "AcceptEnv *", did you restart the 
sshd service [1]?  The following illustrates steps that might help you 
verify that WIENROOT appears on a remote vlsi node:


username@computername:~$ echo $WIENROOT

username@computername:~$ export WIENROOT=/servernode1
username@computername:~$ echo $WIENROOT
/servernode1
username@computername:~$ ssh vlsi
Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 4.15.0-64-generic x86_64)
...
Last login: Sat Sep 28 12:04:07 2019 from xxx.x.x.x
username@computername:~$ echo $WIENROOT

username@computername:~$ exit
logout
Connection to vlsi closed.
username@computername:~$ sudo gedit /etc/ssh/ssh_config
[sudo] password for username:

username@computername:~$ sudo gedit /etc/ssh/sshd_config

username@computername:~$ grep SendEnv /etc/ssh/ssh_config
    SendEnv LANG LC_* WIENROOT
username@computername:~$ grep AcceptEnv /etc/ssh/sshd_config
AcceptEnv LANG LC_* WIENROOT


   username@computername:~$ sudo service sshd restart


username@computername:~$ ssh vlsi
...
username@computername:~$ echo $WIENROOT
/servernode1
username@computername:~$ exit

[1] 
https://askubuntu.com/questions/462968/take-changes-in-file-sshd-config-file-without-server-reboot


On 9/28/2019 11:22 AM, Indranil mal wrote:
Sir I have tried with " SetEnv * " Still nothing is coming with echo 
 commad and user name by mistake I posted wrong Otherwise no issue 
with user name and I have set the parallel options file taksset "no" 
and remote options are 1 1 in server and client machines.



On Sat, 28 Sep 2019 11:36 Gavin Abo, > wrote:



Respected Sir, In my linux(Ubuntu 18.04 LTS) in ssh_config, and
in sshd_config there are two line already "SendEnv LANG LC_*"
"AcceptEnv LANG LC_*" respectively.


The "LANG LC_*" probably only puts just the local language
variables in the remote environment.  Did you follow the previous
advice [1] of trying to use "*" to put all variables from the
local environment?

[1]
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19049.html


However, ssh vsli1 'echo $WIENROOT' gives nothing (blank).


That seems to be the main cause of the problem as it should not
return (blank) but needs to return "/servernode1" as you
previously mentioned [2].

[2]
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html

Perhaps the message below is a clue.  It you had set the WIENROOT
variable in .bashrc of your /home/vlsi accounts on each system,
you likely have to login and use that same /home/vlsi account on
the head node as the output below seems to indicate login to a
different /home/niel account.  Alternatively, setting the
WIENROOT variable in .bashrc of all /home/niel accounts on each
node might work too.


   The command ssh vsli1 'pwd $WIENROOT' print "/home/vlsi" the
common home directory and
ssh vlsi1 "env"
...
USER=niel
PWD=/home/niel
HOME=/home/niel
...
this is similar as server, and other nodes.

Sir After changing the parallel option file in $WIENROOT in
server to

setenv TASKSET *"yes" from "no"*
if ( ! $?USE_REMOTE ) setenv USE_REMOTE 1
if ( ! $?MPI_REMOTE ) setenv MPI_REMOTE 1
setenv WIEN_GRANULARITY 1
setenv DELAY 0.1
setenv SLEEPY 1
setenv WIEN_MPIRUN "mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_"
setenv CORES_PER_NODE 1

the error is not coming but the program is not increasing steps
after lapw0 it stuck in lapw1


Since it seemed to be throwing an appropriate error message with
TASKSET previously unlike when set to "yes", probably you should
change it back to "no".

___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-28 Thread Gavin Abo
After you set both "SendEnv *" and "AcceptEnv *", did you restart the 
sshd service [1]?  The following illustrates steps that might help you 
verify that WIENROOT appears on a remote vlsi node:


username@computername:~$ echo $WIENROOT

username@computername:~$ export WIENROOT=/servernode1
username@computername:~$ echo $WIENROOT
/servernode1
username@computername:~$ ssh vlsi
Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 4.15.0-64-generic x86_64)
...
Last login: Sat Sep 28 12:04:07 2019 from xxx.x.x.x
username@computername:~$ echo $WIENROOT

username@computername:~$ exit
logout
Connection to vlsi closed.
username@computername:~$ sudo gedit /etc/ssh/ssh_config
[sudo] password for username:

username@computername:~$ sudo gedit /etc/ssh/sshd_config

username@computername:~$ grep SendEnv /etc/ssh/ssh_config
    SendEnv LANG LC_* WIENROOT
username@computername:~$ grep AcceptEnv /etc/ssh/sshd_config
AcceptEnv LANG LC_* WIENROOT
username@computername:~$ ssh vlsi
...
username@computername:~$ echo $WIENROOT
/servernode1
username@computername:~$ exit

[1] 
https://askubuntu.com/questions/462968/take-changes-in-file-sshd-config-file-without-server-reboot


On 9/28/2019 11:22 AM, Indranil mal wrote:
Sir I have tried with " SetEnv * " Still nothing is coming with echo 
 commad and user name by mistake I posted wrong Otherwise no issue 
with user name and I have set the parallel options file taksset "no" 
and remote options are 1 1 in server and client machines.



On Sat, 28 Sep 2019 11:36 Gavin Abo, > wrote:



Respected Sir, In my linux(Ubuntu 18.04 LTS) in ssh_config, and
in sshd_config there are two line already "SendEnv LANG LC_*"
"AcceptEnv LANG LC_*" respectively.


The "LANG LC_*" probably only puts just the local language
variables in the remote environment.  Did you follow the previous
advice [1] of trying to use "*" to put all variables from the
local environment?

[1]
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19049.html


However, ssh vsli1 'echo $WIENROOT' gives nothing (blank).


That seems to be the main cause of the problem as it should not
return (blank) but needs to return "/servernode1" as you
previously mentioned [2].

[2]
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html

Perhaps the message below is a clue.  It you had set the WIENROOT
variable in .bashrc of your /home/vlsi accounts on each system,
you likely have to login and use that same /home/vlsi account on
the head node as the output below seems to indicate login to a
different /home/niel account. Alternatively, setting the WIENROOT
variable in .bashrc of all /home/niel accounts on each node might
work too.


   The command ssh vsli1 'pwd $WIENROOT' print "/home/vlsi" the
common home directory and
ssh vlsi1 "env"
...
USER=niel
PWD=/home/niel
HOME=/home/niel
...
this is similar as server, and other nodes.

Sir After changing the parallel option file in $WIENROOT in
server to

setenv TASKSET *"yes" from "no"*
if ( ! $?USE_REMOTE ) setenv USE_REMOTE 1
if ( ! $?MPI_REMOTE ) setenv MPI_REMOTE 1
setenv WIEN_GRANULARITY 1
setenv DELAY 0.1
setenv SLEEPY 1
setenv WIEN_MPIRUN "mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_"
setenv CORES_PER_NODE 1

the error is not coming but the program is not increasing steps
after lapw0 it stuck in lapw1


Since it seemed to be throwing an appropriate error message with
TASKSET previously unlike when set to "yes", probably you should
change it back to "no".

___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-28 Thread Indranil mal
Sir I have tried with " SetEnv * " Still nothing is coming with echo
 commad and user name by mistake I posted wrong Otherwise no issue with
user name and I have set the parallel options file taksset "no" and remote
options are 1 1 in server and client machines.


On Sat, 28 Sep 2019 11:36 Gavin Abo,  wrote:

> Respected Sir, In my linux(Ubuntu 18.04 LTS) in ssh_config, and in
> sshd_config there are two line already "SendEnv LANG LC_*" "AcceptEnv LANG
> LC_*" respectively.
>
> The "LANG LC_*" probably only puts just the local language variables in
> the remote environment.  Did you follow the previous advice [1] of trying
> to use "*" to put all variables from the local environment?
>
> [1]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19049.html
>
> However, ssh vsli1 'echo $WIENROOT' gives nothing (blank).
>
> That seems to be the main cause of the problem as it should not return
> (blank) but needs to return "/servernode1" as you previously mentioned [2].
>
> [2]
> https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html
>
> Perhaps the message below is a clue.  It you had set the WIENROOT variable
> in .bashrc of your /home/vlsi accounts on each system, you likely have to
> login and use that same /home/vlsi account on the head node as the output
> below seems to indicate login to a different /home/niel account.
> Alternatively, setting the WIENROOT variable in .bashrc of all /home/niel
> accounts on each node might work too.
>
>The command ssh vsli1 'pwd $WIENROOT' print "/home/vlsi" the common
> home directory and
> ssh vlsi1 "env"
> ...
> USER=niel
> PWD=/home/niel
> HOME=/home/niel
> ...
> this is similar as server, and other nodes.
>
> Sir After changing the parallel option file in $WIENROOT in server to
>
> setenv TASKSET *"yes" from "no"*
> if ( ! $?USE_REMOTE ) setenv USE_REMOTE 1
> if ( ! $?MPI_REMOTE ) setenv MPI_REMOTE 1
> setenv WIEN_GRANULARITY 1
> setenv DELAY 0.1
> setenv SLEEPY 1
> setenv WIEN_MPIRUN "mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_"
> setenv CORES_PER_NODE 1
>
> the error is not coming but the program is not increasing steps after
> lapw0 it stuck in lapw1
>
> Since it seemed to be throwing an appropriate error message with TASKSET
> previously unlike when set to "yes", probably you should change it back to
> "no".
> ___
> Wien mailing list
> Wien@zeus.theochem.tuwien.ac.at
> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> SEARCH the MAILING-LIST at:
> http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
>
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-28 Thread Gavin Abo
Respected Sir, In my linux(Ubuntu 18.04 LTS) in ssh_config, and in 
sshd_config there are two line already "SendEnv LANG LC_*" "AcceptEnv 
LANG LC_*" respectively.


The "LANG LC_*" probably only puts just the local language variables in 
the remote environment.  Did you follow the previous advice [1] of 
trying to use "*" to put all variables from the local environment?


[1] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19049.html



However, ssh vsli1 'echo $WIENROOT' gives nothing (blank).


That seems to be the main cause of the problem as it should not return 
(blank) but needs to return "/servernode1" as you previously mentioned [2].


[2] 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg19036.html


Perhaps the message below is a clue.  It you had set the WIENROOT 
variable in .bashrc of your /home/vlsi accounts on each system, you 
likely have to login and use that same /home/vlsi account on the head 
node as the output below seems to indicate login to a different 
/home/niel account. Alternatively, setting the WIENROOT variable in 
.bashrc of all /home/niel accounts on each node might work too.


   The command ssh vsli1 'pwd $WIENROOT' print "/home/vlsi" the common 
home directory and

ssh vlsi1 "env"
...
USER=niel
PWD=/home/niel
HOME=/home/niel
...
this is similar as server, and other nodes.

Sir After changing the parallel option file in $WIENROOT in server to

setenv TASKSET *"yes" from "no"*
if ( ! $?USE_REMOTE ) setenv USE_REMOTE 1
if ( ! $?MPI_REMOTE ) setenv MPI_REMOTE 1
setenv WIEN_GRANULARITY 1
setenv DELAY 0.1
setenv SLEEPY 1
setenv WIEN_MPIRUN "mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_"
setenv CORES_PER_NODE 1

the error is not coming but the program is not increasing steps after 
lapw0 it stuck in lapw1


Since it seemed to be throwing an appropriate error message with TASKSET 
previously unlike when set to "yes", probably you should change it back 
to "no".


___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


Re: [Wien] parallel ssh error

2019-09-27 Thread Indranil mal
Respected Sir, In my linux(Ubuntu 18.04 LTS) in ssh_config, and in
sshd_config there are two line already "SendEnv LANG LC_*" "AcceptEnv LANG
LC_*" respectively. However, ssh vsli1 'echo $WIENROOT' gives nothing
(blank).   The command ssh vsli1 'pwd $WIENROOT' print "/home/vlsi" the
common home directory and
ssh vlsi1 "env"
SSH_CONNECTION=172.27.46.251 44138 172.27.46.233 22
LANG=en_IN
XDG_SESSION_ID=47
USER=niel
PWD=/home/niel
HOME=/home/niel
SSH_CLIENT=172.27.46.251 44138 22
LC_NUMERIC=POSIX
MAIL=/var/mail/niel
SHELL=/bin/bash
SHLVL=1
LANGUAGE=en_IN:en
LOGNAME=niel
DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1000/bus
XDG_RUNTIME_DIR=/run/user/1000
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games
_=/usr/bin/env

this is similar as server, and other nodes.


Sir After changing the parallel option file in $WIENROOT in server to

setenv TASKSET *"yes" from "no"*
if ( ! $?USE_REMOTE ) setenv USE_REMOTE 1
if ( ! $?MPI_REMOTE ) setenv MPI_REMOTE 1
setenv WIEN_GRANULARITY 1
setenv DELAY 0.1
setenv SLEEPY 1
setenv WIEN_MPIRUN "mpirun -np _NP_ -machinefile _HOSTS_ _EXEC_"
setenv CORES_PER_NODE 1

the error is not coming but the program is not increasing steps after lapw0
it stuck in lapw1


what should be the parallel option file in server and all client node?



On Fri, Sep 27, 2019 at 12:05 PM Peter Blaha 
wrote:

> Ok. So the problem seems to be that in your linux the ssh does not
> send/accept the "environment".
>
> What do you get with:
>
> ssh vsli2 'echo $WIENROOT'
>
> If you have root permissions, I suggest to do the following:
>
> At least on my Linux (Suse) there is a  /etc/ssh   directory, with files
>
> ssh_config and sshd_config.
>
> Edit these files and add lines:
> SendEnv *  # in ssh_config
> AcceptEnv *# in sshd_config
>
>
>
> On 9/27/19 11:20 AM, Indranil mal wrote:
> > Respected Sir, As per Your suggestion I have done the single process
> > with one iteration successfully encountered no issue in all the nodes.
> > However in parallel running facing the same  error
> >
> > grep: *scf1*: No such file or directory
> > cp: cannot stat '.in.tmp': No such file or directory
> > FERMI - Error
> > grep: *scf1*: No such file or directory
> > Parallel.scf1_1: No such file or directory.
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> >   LAPW0 END
> > hup: Command not found.
> >
> > Previously I was doing a mistake with user name and home directory now
> in all the pc the user name and the home directory is same (/home/vlsi) is
> same and the working directory is accessible from every node.
> >
> >   (ls -l $WIENROOT/lapw1c
> > -rwxr-xr-x 1 vlsi vlsi 2151824 Sep 26 02:41 /servernode01/lapw1c) this
> reflects in all the pcs.
> >
> >
> >
> >
> > On Thu, Sep 26, 2019 at 1:27 PM Peter Blaha
> > mailto:pbl...@theochem.tuwien.ac.at>>
> wrote:
> >
> > First of all, one of the errors was: lapw1c: command not found
> >
> > You showed us only the existence of "lapw1", not "lapw1c" with the ls
> > commands.
> >
> > However, since you also have:  fixerror_lapw: command not found
> >
> > I don't think that this is the problem.
> >
> > -
> > I'm more concerned about the different usernames/owners of lapw1 on
> > different computers.
> > It is not important who owns $WIENROOT/*, as long as everybody has
> r-x
> > permissions.
> >
> > However, what is your username and your home-directory on the
> different
> > machines ? It must be the same ! And do you have access to the actual
> > working directory ?
> > In what directory did you start the calculations?
> > Is it a directory called "Parallel" ? What is the full path of that
> on
> > every computer (/casenode1/Parallel ?)
> > --
> >
> > First check would be:
> >
> > On vlsi1 change into the working directory (Parallel ?) and run one
> > iteration without parallelization:   run -i 1
> >
> > then login to   ssh vsli2 (passwordless), cd into "Parallel" and do
> > another non-parallel cycle.  Does it work ?
> > ---
> >
> >
> > On 9/26/19 11:48 AM, Indranil mal wrote:
> >  > Dear developers and users
> >  >  I have 5 individual Linux
> >  > (Ubuntu) pc with intel i7 octa core processors and 16GB RAM in
> each
> >  > connected via a 1GBps LAN.  password less 

Re: [Wien] parallel ssh error

2019-09-27 Thread Peter Blaha
Ok. So the problem seems to be that in your linux the ssh does not 
send/accept the "environment".


What do you get with:

ssh vsli2 'echo $WIENROOT'

If you have root permissions, I suggest to do the following:

At least on my Linux (Suse) there is a  /etc/ssh   directory, with files

ssh_config and sshd_config.

Edit these files and add lines:
SendEnv *  # in ssh_config
AcceptEnv *# in sshd_config



On 9/27/19 11:20 AM, Indranil mal wrote:
Respected Sir, As per Your suggestion I have done the single process 
with one iteration successfully encountered no issue in all the nodes. 
However in parallel running facing the same  error


grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
  LAPW0 END
hup: Command not found.

Previously I was doing a mistake with user name and home directory now in all 
the pc the user name and the home directory is same (/home/vlsi) is same and 
the working directory is accessible from every node.

  (ls -l $WIENROOT/lapw1c
-rwxr-xr-x 1 vlsi vlsi 2151824 Sep 26 02:41 /servernode01/lapw1c) this reflects 
in all the pcs.




On Thu, Sep 26, 2019 at 1:27 PM Peter Blaha 
mailto:pbl...@theochem.tuwien.ac.at>> wrote:


First of all, one of the errors was: lapw1c: command not found

You showed us only the existence of "lapw1", not "lapw1c" with the ls
commands.

However, since you also have:  fixerror_lapw: command not found

I don't think that this is the problem.

-
I'm more concerned about the different usernames/owners of lapw1 on
different computers.
It is not important who owns $WIENROOT/*, as long as everybody has r-x
permissions.

However, what is your username and your home-directory on the different
machines ? It must be the same ! And do you have access to the actual
working directory ?
In what directory did you start the calculations?
Is it a directory called "Parallel" ? What is the full path of that on
every computer (/casenode1/Parallel ?)
--

First check would be:

On vlsi1 change into the working directory (Parallel ?) and run one
iteration without parallelization:   run -i 1

then login to   ssh vsli2 (passwordless), cd into "Parallel" and do
another non-parallel cycle.  Does it work ?
---


On 9/26/19 11:48 AM, Indranil mal wrote:
 > Dear developers and users
 >      I have 5 individual Linux
 > (Ubuntu) pc with intel i7 octa core processors and 16GB RAM in each
 > connected via a 1GBps LAN.  password less ssh working properly. I
have
 > installed WIEN2K 19 in the one machine (M1 server) in the directory
 > "/servernode1" and the case directory is "/casenode1"  and
through NFS I
 > have mounted the "servernode1", and "casenode1" in other four pcs
with
 > same name local folders ("servernode1", and "casenode1") in them.
I have
 > installed ,intel compilers, libxc, fftw, elpa in all the nodes
 > individually. I have manually edited the bash file  $WIENROOT
path and
 > case directory and the WIEN2K options file. Keep all the value
same in
 > all the client nodes as it is in the server node.
 >
 > alias cdw="cd /casenode1"
 > export OMP_NUM_THREADS=4
 > #export LD_LIBRARY_PATH=.
 > export EDITOR="emacs"
 > export SCRATCH=./
 > export WIENROOT=/servernode1
 > export W2WEB_CASE_BASEDIR=/casenode1
 > export STRUCTEDIT_PATH=$WIENROOT/SRC_structeditor/bin
 >
 > Now when I am doing parallel calculations with all the client
nodes in
 > machine file ,
 > # k-points are left, they will be distributed to the
residual-machine_name.
 > #
 > 1:vlsi1
 > 1:vlsi2
 > 1:vlsi3
 > 1:vlsi4
 >
 > granularity:1
 > extrafine:1
 > #
 >
 >
 > and getting the following error
 >
 > grep: *scf1*: No such file or directory
 > cp: cannot stat '.in.tmp': No such file or directory
 > FERMI - Error
 > grep: *scf1*: No such file or directory
 > Parallel.scf1_1: No such file or directory.
 > bash: fixerror_lapw: command not found
 > bash: lapw1c: command not found
 > bash: fixerror_lapw: command not found
 > 

Re: [Wien] parallel ssh error

2019-09-26 Thread Indranil mal
Respected Sir, As per Your suggestion I have done the single process with
one iteration successfully encountered no issue in all the nodes. However
in parallel running facing the same  error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
 LAPW0 END
hup: Command not found.

Previously I was doing a mistake with user name and home directory now
in all the pc the user name and the home directory is same
(/home/vlsi) is same and the working directory is accessible from
every node.

 (ls -l $WIENROOT/lapw1c
-rwxr-xr-x 1 vlsi vlsi 2151824 Sep 26 02:41 /servernode01/lapw1c) this
reflects in all the pcs.




On Thu, Sep 26, 2019 at 1:27 PM Peter Blaha 
wrote:

> First of all, one of the errors was: lapw1c: command not found
>
> You showed us only the existence of "lapw1", not "lapw1c" with the ls
> commands.
>
> However, since you also have:  fixerror_lapw: command not found
>
> I don't think that this is the problem.
>
> -
> I'm more concerned about the different usernames/owners of lapw1 on
> different computers.
> It is not important who owns $WIENROOT/*, as long as everybody has r-x
> permissions.
>
> However, what is your username and your home-directory on the different
> machines ? It must be the same ! And do you have access to the actual
> working directory ?
> In what directory did you start the calculations?
> Is it a directory called "Parallel" ? What is the full path of that on
> every computer (/casenode1/Parallel ?)
> --
>
> First check would be:
>
> On vlsi1 change into the working directory (Parallel ?) and run one
> iteration without parallelization:   run -i 1
>
> then login to   ssh vsli2 (passwordless), cd into "Parallel" and do
> another non-parallel cycle.  Does it work ?
> ---
>
>
> On 9/26/19 11:48 AM, Indranil mal wrote:
> > Dear developers and users
> >  I have 5 individual Linux
> > (Ubuntu) pc with intel i7 octa core processors and 16GB RAM in each
> > connected via a 1GBps LAN.  password less ssh working properly. I have
> > installed WIEN2K 19 in the one machine (M1 server) in the directory
> > "/servernode1" and the case directory is "/casenode1"  and through NFS I
> > have mounted the "servernode1", and "casenode1" in other four pcs with
> > same name local folders ("servernode1", and "casenode1") in them. I have
> > installed ,intel compilers, libxc, fftw, elpa in all the nodes
> > individually. I have manually edited the bash file  $WIENROOT path and
> > case directory and the WIEN2K options file. Keep all the value same in
> > all the client nodes as it is in the server node.
> >
> > alias cdw="cd /casenode1"
> > export OMP_NUM_THREADS=4
> > #export LD_LIBRARY_PATH=.
> > export EDITOR="emacs"
> > export SCRATCH=./
> > export WIENROOT=/servernode1
> > export W2WEB_CASE_BASEDIR=/casenode1
> > export STRUCTEDIT_PATH=$WIENROOT/SRC_structeditor/bin
> >
> > Now when I am doing parallel calculations with all the client nodes in
> > machine file ,
> > # k-points are left, they will be distributed to the
> residual-machine_name.
> > #
> > 1:vlsi1
> > 1:vlsi2
> > 1:vlsi3
> > 1:vlsi4
> >
> > granularity:1
> > extrafine:1
> > #
> >
> >
> > and getting the following error
> >
> > grep: *scf1*: No such file or directory
> > cp: cannot stat '.in.tmp': No such file or directory
> > FERMI - Error
> > grep: *scf1*: No such file or directory
> > Parallel.scf1_1: No such file or directory.
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> > bash: fixerror_lapw: command not found
> > bash: lapw1c: command not found
> >   LAPW0 END
> > hup: Command not found.
> >
> > ###Error file lapw2 error
> >   'LAPW2' - can't open unit: 30
> >   'LAPW2' -filename: Parallel.energy_1
> > **  testerror: Error in Parallel LAPW2
> >
> > I have checked the with " ls -l $WIENROOT/lapw1" as suggested in the
> > previous mailing list and got the
> > -rwxr-xr-x 1 vlsi vlsi 2139552 Sep 26 02:41 /servernode1/lapw1 for
> > server (vlsi is the user name in server)
> > -rwxr-xr-x 1 vlsi1 vlsi1 2139552 Sep 26 02:41 /servernode1/lapw1 (in
> > node1 the user name is vlsi1)
> > -rwxr-xr-x 1 vlsi2 vlsi2 2139552 Sep 26 02:41 /servernode1/lapw1 (in
> > node2 the user name is vlsi2)
> > please help

Re: [Wien] parallel ssh error

2019-09-26 Thread Peter Blaha

First of all, one of the errors was: lapw1c: command not found

You showed us only the existence of "lapw1", not "lapw1c" with the ls 
commands.


However, since you also have:  fixerror_lapw: command not found

I don't think that this is the problem.

-
I'm more concerned about the different usernames/owners of lapw1 on 
different computers.
It is not important who owns $WIENROOT/*, as long as everybody has r-x 
permissions.


However, what is your username and your home-directory on the different 
machines ? It must be the same ! And do you have access to the actual 
working directory ?

In what directory did you start the calculations?
Is it a directory called "Parallel" ? What is the full path of that on 
every computer (/casenode1/Parallel ?)

--

First check would be:

On vlsi1 change into the working directory (Parallel ?) and run one 
iteration without parallelization:   run -i 1


then login to   ssh vsli2 (passwordless), cd into "Parallel" and do 
another non-parallel cycle.  Does it work ?

---


On 9/26/19 11:48 AM, Indranil mal wrote:

Dear developers and users
     I have 5 individual Linux 
(Ubuntu) pc with intel i7 octa core processors and 16GB RAM in each 
connected via a 1GBps LAN.  password less ssh working properly. I have 
installed WIEN2K 19 in the one machine (M1 server) in the directory 
"/servernode1" and the case directory is "/casenode1"  and through NFS I 
have mounted the "servernode1", and "casenode1" in other four pcs with 
same name local folders ("servernode1", and "casenode1") in them. I have 
installed ,intel compilers, libxc, fftw, elpa in all the nodes 
individually. I have manually edited the bash file  $WIENROOT path and 
case directory and the WIEN2K options file. Keep all the value same in 
all the client nodes as it is in the server node.


alias cdw="cd /casenode1"
export OMP_NUM_THREADS=4
#export LD_LIBRARY_PATH=.
export EDITOR="emacs"
export SCRATCH=./
export WIENROOT=/servernode1
export W2WEB_CASE_BASEDIR=/casenode1
export STRUCTEDIT_PATH=$WIENROOT/SRC_structeditor/bin

Now when I am doing parallel calculations with all the client nodes in 
machine file ,

# k-points are left, they will be distributed to the residual-machine_name.
#
1:vlsi1
1:vlsi2
1:vlsi3
1:vlsi4

granularity:1
extrafine:1
#


and getting the following error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
  LAPW0 END
hup: Command not found.

###Error file lapw2 error
  'LAPW2' - can't open unit: 30
  'LAPW2' -        filename: Parallel.energy_1
**  testerror: Error in Parallel LAPW2

I have checked the with " ls -l $WIENROOT/lapw1" as suggested in the 
previous mailing list and got the
-rwxr-xr-x 1 vlsi vlsi 2139552 Sep 26 02:41 /servernode1/lapw1 for 
server (vlsi is the user name in server)
-rwxr-xr-x 1 vlsi1 vlsi1 2139552 Sep 26 02:41 /servernode1/lapw1 (in 
node1 the user name is vlsi1)
-rwxr-xr-x 1 vlsi2 vlsi2 2139552 Sep 26 02:41 /servernode1/lapw1 (in 
node2 the user name is vlsi2)

please help


thanking you
Indranil




___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html



--

  P.Blaha
--
Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
Phone: +43-1-58801-165300 FAX: +43-1-58801-165982
Email: bl...@theochem.tuwien.ac.atWIEN2k: http://www.wien2k.at
WWW:   http://www.imc.tuwien.ac.at/TC_Blaha
--
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


[Wien] parallel ssh error

2019-09-25 Thread Indranil mal
Dear developers and users
I have 5 individual Linux (Ubuntu)
pc with intel i7 octa core processors and 16GB RAM in each connected via a
1GBps LAN.  password less ssh working properly. I have installed WIEN2K 19
in the one machine (M1 server) in the directory "/servernode1" and the case
directory is "/casenode1"  and through NFS I have mounted the
"servernode1", and "casenode1" in other four pcs with same name local
folders ("servernode1", and "casenode1") in them. I have installed ,intel
compilers, libxc, fftw, elpa in all the nodes individually. I have manually
edited the bash file  $WIENROOT path and case directory and the WIEN2K
options file. Keep all the value same in all the client nodes as it is in
the server node.

alias cdw="cd /casenode1"
export OMP_NUM_THREADS=4
#export LD_LIBRARY_PATH=.
export EDITOR="emacs"
export SCRATCH=./
export WIENROOT=/servernode1
export W2WEB_CASE_BASEDIR=/casenode1
export STRUCTEDIT_PATH=$WIENROOT/SRC_structeditor/bin

Now when I am doing parallel calculations with all the client nodes in
machine file ,
# k-points are left, they will be distributed to the residual-machine_name.
#
1:vlsi1
1:vlsi2
1:vlsi3
1:vlsi4

granularity:1
extrafine:1
#


and getting the following error

grep: *scf1*: No such file or directory
cp: cannot stat '.in.tmp': No such file or directory
FERMI - Error
grep: *scf1*: No such file or directory
Parallel.scf1_1: No such file or directory.
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
bash: fixerror_lapw: command not found
bash: lapw1c: command not found
 LAPW0 END
hup: Command not found.

###Error file lapw2 error
 'LAPW2' - can't open unit: 30

 'LAPW2' -filename: Parallel.energy_1

**  testerror: Error in Parallel LAPW2

I have checked the with " ls -l $WIENROOT/lapw1" as suggested in the
previous mailing list and got the
-rwxr-xr-x 1 vlsi vlsi 2139552 Sep 26 02:41 /servernode1/lapw1 for server
(vlsi is the user name in server)
-rwxr-xr-x 1 vlsi1 vlsi1 2139552 Sep 26 02:41 /servernode1/lapw1 (in node1
the user name is vlsi1)
-rwxr-xr-x 1 vlsi2 vlsi2 2139552 Sep 26 02:41 /servernode1/lapw1 (in node2
the user name is vlsi2)
please help


thanking you
Indranil
___
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html