You don't need to do that. Running the LVM commands in one node is all you need 
to do assuming that its the same storage presented to both hosts.

Sent from my iPhone

On Nov 20, 2012, at 7:02 AM, emmanuel segura <emi2f...@gmail.com> wrote:

> Do it the same step on second server
> 
> 2012/11/20 james cyriac <jamescyria...@gmail.com>
>> Hi,
>>  
>> can you send the detials,i have to put entry in both servers?now i created
>>  
>> map disk 150G both servers
>> and created in node 1 vg03
>> then
>> mkfs.gfs2 -p lock_dlm -t sap-cluster1:gfs2 -j 8 /dev/vg03/lvol0
>>  
>> now i able to mount in first server.
>>  
>>  
>>  /dev/vg03/lvol0          /usr/sap/trans       gfs2 defaults   0   0
>> 
>> On Tue, Nov 20, 2012 at 3:07 PM, emmanuel segura <emi2f...@gmail.com> wrote:
>>> You have to use /etc/fstab with _netdev option, redhat cluster doesn't 
>>> support active/active service
>>> 
>>> 
>>> 2012/11/20 james cyriac <jamescyria...@gmail.com>
>>>> Hi all,
>>>> 
>>>> i am installing redhat cluster 6 two node cluser.the issue is i am not 
>>>> able to mount my GFS file sytem in both the node at same time..
>>>>  
>>>> please find my clustat output ..
>>>> 
>>>> 
>>>> [root@saperpprod01 ~]# clustat
>>>> Cluster Status for sap-cluster1 @ Tue Nov 20 14:51:28 2012
>>>> Member Status: Quorate
>>>>  Member Name                                                     ID   
>>>> Status
>>>>  ------ ----                                                     ---- 
>>>> ------
>>>>  saperpprod01                                                        1 
>>>> Online, Local, rgmanager
>>>>  saperpprod02                                                        2 
>>>> Online, rgmanager
>>>>  Service Name                                                     Owner 
>>>> (Last)                                                     State
>>>>  ------- ----                                                     ----- 
>>>> ------                                                     -----
>>>>  service:oracle                                                   
>>>> saperpprod01                                                     started
>>>>  service:profile-gfs                                              
>>>> saperpprod01                                                     started
>>>>  service:sap                                                      
>>>> saperpprod01                                                     started
>>>> [root@saperpprod01 ~]#
>>>> oralce and sap is fine and it is flaying in both nodes.i want mount my GFS 
>>>> vols same time at both the nodes.
>>>>  
>>>> Thanks in advacne
>>>> james
>>>>  
>>>>  
>>>> but profile-gfs is GFS file system and i want present the GFS mount point 
>>>> same time both the node.please help me this
>>>> On Mon, Nov 19, 2012 at 1:16 PM, Jacek Konieczny <jaj...@jajcus.net> wrote:
>>>>> Hi,
>>>>> 
>>>>> I am setting up a cluster using:
>>>>> 
>>>>> Linux kernel 3.6.6
>>>>> Corosync 2.1.0
>>>>> DLM 4.0.0
>>>>> CLVMD 2.02.98
>>>>> Pacemaker 1.1.8
>>>>> DRBD 8.3.13
>>>>> 
>>>>> Now I have stuck on the 'clean shutdown of a node' scenario.
>>>>> 
>>>>> It goes like that:
>>>>> - resources using the shared storage are properly stopped by Pacemaker.
>>>>> - DRBD is cleanly demoted and unconfigured by Pacemaker
>>>>> - Pacemaker cleanly exits
>>>>> - CLVMD is stopped.
>>>>> – dlm_controld is stopped
>>>>> – corosync is being stopped
>>>>> 
>>>>> and at this point the node is fenced (rebooted) by the dlm_controld on
>>>>> the other node. I would expect it continue with a clean shutdown.
>>>>> 
>>>>> Any idea how to debug/fix it?
>>>>> Is this '541 cpg_dispatch error 9' the problem?
>>>>> 
>>>>> Logs from the node being shut down (log file system mounted with the 
>>>>> 'sync'
>>>>> option, syslog shutdown delayed as much as possible):
>>>>> 
>>>>> Kernel:
>>>>> Nov 19 09:49:40 dev1n2 kernel: : [  542.049407] block drbd0: worker 
>>>>> terminated
>>>>> Nov 19 09:49:40 dev1n2 kernel: : [  542.049412] block drbd0: Terminating 
>>>>> drbd0_worker
>>>>> Nov 19 09:49:43 dev1n2 kernel: : [  544.934390] dlm: clvmd: leaving the 
>>>>> lockspace group...
>>>>> Nov 19 09:49:43 dev1n2 kernel: : [  544.937584] dlm: clvmd: group event 
>>>>> done 0 0
>>>>> Nov 19 09:49:43 dev1n2 kernel: : [  544.937897] dlm: clvmd: 
>>>>> release_lockspace final free
>>>>> Nov 19 09:49:43 dev1n2 kernel: : [  544.961407] dlm: closing connection 
>>>>> to node 2
>>>>> Nov 19 09:49:43 dev1n2 kernel: : [  544.961431] dlm: closing connection 
>>>>> to node 1
>>>>> 
>>>>> User space:
>>>>> Nov 19 09:49:41 dev1n2 pacemakerd[1267]:   notice: stop_child: Stopping 
>>>>> cib: Sent -15 to process 1279
>>>>> Nov 19 09:49:41 dev1n2 notifyd[1139]: [notice] dev1n2[2] 
>>>>> stonithd:1281:0x7fc423dfd5e0 is now disconnected from corosync
>>>>> Nov 19 09:49:41 dev1n2 cib[1279]:   notice: terminate_cs_connection: 
>>>>> Disconnecting from Corosync
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1db
>>>>> Nov 19 09:49:41 dev1n2 notifyd[1139]: [notice] dev1n2[2] 
>>>>> cib:1279:0x7fc4240008d0 is now disconnected from corosync
>>>>> Nov 19 09:49:41 dev1n2 cib[1279]:   notice: terminate_cs_connection: 
>>>>> Disconnecting from Corosync
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1dd
>>>>> Nov 19 09:49:41 dev1n2 pacemakerd[1267]:   notice: pcmk_shutdown_worker: 
>>>>> Shutdown complete
>>>>> Nov 19 09:49:41 dev1n2 notifyd[1139]: [notice] dev1n2[2] 
>>>>> pacemakerd:1267:0x7fc423bf8ed0 is now disconnected from corosync
>>>>> Nov 19 09:49:41 dev1n2 notifyd[1139]: [notice] dev1n2[2] 
>>>>> pacemakerd:1267:0x7fc423bf7660 is now disconnected from corosync
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1de
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1de
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1e1
>>>>> Nov 19 09:49:41 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1e1
>>>>> Nov 19 09:49:43 dev1n2 dlm_controld[1142]: 541 cpg_dispatch error 9
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1e7
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [TOTEM ] Retransmit List: 1e7
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Unloading all Corosync 
>>>>> service engines.
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [QB    ] withdrawing server 
>>>>> sockets
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync vote quorum service v1.0
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [QB    ] withdrawing server 
>>>>> sockets
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync configuration map access
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [QB    ] withdrawing server 
>>>>> sockets
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync configuration service
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [QB    ] withdrawing server 
>>>>> sockets
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync cluster closed process group service v1.01
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [QB    ] withdrawing server 
>>>>> sockets
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync cluster quorum service v0.1
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync profile loading service
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [WD    ] magically closing the 
>>>>> watchdog.
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [SERV  ] Service engine unloaded: 
>>>>> corosync watchdog service
>>>>> Nov 19 09:49:43 dev1n2 corosync[1130]:  [MAIN  ] Corosync Cluster Engine 
>>>>> exiting normally
>>>>> 
>>>>> 
>>>>> Logs from the surviving node:
>>>>> 
>>>>> Kernel:
>>>>> Nov 19 09:49:39 dev1n1 kernel: : [80664.615988] block drbd0: conn( 
>>>>> Unconnected -> WFConnection )
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497187] dlm: clvmd: dlm_recover 11
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497241] dlm: clvmd: 
>>>>> dlm_clear_toss 1 done
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497291] dlm: clvmd: remove member 
>>>>> 2
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497295] dlm: clvmd: 
>>>>> dlm_recover_members 1 nodes
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497298] dlm: clvmd: generation 15 
>>>>> slots 1 1:1
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497300] dlm: clvmd: 
>>>>> dlm_recover_directory
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497302] dlm: clvmd: 
>>>>> dlm_recover_directory 0 in 0 new
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497304] dlm: clvmd: 
>>>>> dlm_recover_directory 0 out 0 messages
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497307] dlm: clvmd: 
>>>>> dlm_recover_masters
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497309] dlm: clvmd: 
>>>>> dlm_recover_masters 0 of 1
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497311] dlm: clvmd: 
>>>>> dlm_recover_locks 0 out
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497313] dlm: clvmd: 
>>>>> dlm_recover_locks 0 in
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497314] dlm: clvmd: 
>>>>> dlm_recover_rsbs 1 done
>>>>> Nov 19 09:49:42 dev1n1 kernel: : [80667.497366] dlm: clvmd: dlm_recover 
>>>>> 11 generation 15 done: 0 ms
>>>>> Nov 19 09:49:43 dev1n1 kernel: : [80668.211818] dlm: closing connection 
>>>>> to node 2
>>>>> Nov 19 09:49:46 dev1n1 kernel: : [80670.779015] igb: p1p2 NIC Link is Down
>>>>> 
>>>>> User space:
>>>>> Nov 19 09:49:40 dev1n1 pengine[1078]:   notice: stage6: Scheduling Node 
>>>>> dev1n2 for shutdown
>>>>> Nov 19 09:49:40 dev1n1 pengine[1078]:    error: rsc_expand_action: 
>>>>> Couldn't expand vpbx_vg_cl_demote_0
>>>>> Nov 19 09:49:40 dev1n1 pengine[1078]:    error: rsc_expand_action: 
>>>>> Couldn't expand vpbx_vg_cl_demote_0
>>>>> Nov 19 09:49:40 dev1n1 pengine[1078]:   notice: LogActions: Stop    
>>>>> stonith-dev1n1      (dev1n2)
>>>>> Nov 19 09:49:40 dev1n1 pengine[1078]:   notice: process_pe_message: 
>>>>> Calculated Transition 17: /var/lib/pacemaker/pengine/pe-input-1035.bz2
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1d1
>>>>> Nov 19 09:49:40 dev1n1 crmd[1080]:   notice: run_graph: Transition 17 
>>>>> (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
>>>>> Source=/var/lib/pacemaker/pengine/pe-input-1035.bz2): Complete
>>>>> Nov 19 09:49:40 dev1n1 crmd[1080]:   notice: do_state_transition: State 
>>>>> transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
>>>>> cause=C_FSA_INTERNAL origin=notify_crmd ]
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1d4
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1d6 1d8
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1d6
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1d6
>>>>> Nov 19 09:49:40 dev1n1 crmd[1080]:   notice: peer_update_callback: 
>>>>> do_shutdown of dev1n2 (op 63) is complete
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1df
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1df
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1e3
>>>>> Nov 19 09:49:40 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1e3
>>>>> Nov 19 09:49:42 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1e6
>>>>> Nov 19 09:49:42 dev1n1 corosync[1004]:  [TOTEM ] Retransmit List: 1e9
>>>>> Nov 19 09:49:43 dev1n1 corosync[1004]:  [QUORUM] Members[1]: 1
>>>>> Nov 19 09:49:43 dev1n1 crmd[1080]:   notice: 
>>>>> corosync_mark_unseen_peer_dead: Node 2/dev1n2 was not seen in the 
>>>>> previous transition
>>>>> Nov 19 09:49:43 dev1n1 crmd[1080]:   notice: crm_update_peer_state: 
>>>>> corosync_mark_unseen_peer_dead: Node dev1n2[2] - state is now lost
>>>>> Nov 19 09:49:43 dev1n1 corosync[1004]:  [TOTEM ] A processor joined or 
>>>>> left the membership and a new membership (10.28.45.27:30736) was formed.
>>>>> Nov 19 09:49:43 dev1n1 corosync[1004]:  [MAIN  ] Completed service 
>>>>> synchronization, ready to provide service.
>>>>> Nov 19 09:49:43 dev1n1 dlm_controld[1014]: 80664 fence request 2 pid 
>>>>> 27225 nodedown time 1353314983 fence_all dlm_stonith
>>>>> Nov 19 09:49:43 dev1n1 notifyd[1010]: [notice] 192.168.1.2[2] 
>>>>> ip:192.168.1.2 left
>>>>> Nov 19 09:49:43 dev1n1 stonith-ng[1075]:   notice: stonith_command: 
>>>>> Client stonith-api.27225.b5ff8f4d wants to fence (reboot) '2' with device 
>>>>> '(any)'
>>>>> Nov 19 09:49:43 dev1n1 stonith-ng[1075]:   notice: 
>>>>> initiate_remote_stonith_op: Initiating remote operation reboot for 
>>>>> dev1n2: 71447261-0e53-4b20-b628-d3f026a4ae24 (0)
>>>>> Nov 19 09:49:44 dev1n1 external/ipmi[27242]: [27254]: debug: ipmitool 
>>>>> output: Chassis Power Control: Reset
>>>>> Nov 19 09:49:45 dev1n1 stonith-ng[1075]:   notice: log_operation: 
>>>>> Operation 'reboot' [27234] (call 0 from stonith-api.27225) for host 
>>>>> 'dev1n2' with device 'stonith-dev1n2' returned: 0 (OK)
>>>>> Nov 19 09:49:45 dev1n1 stonith-ng[1075]:   notice: remote_op_done: 
>>>>> Operation reboot of dev1n2 by dev1n1 for 
>>>>> stonith-api.27225@dev1n1.71447261: OK
>>>>> Nov 19 09:49:45 dev1n1 crmd[1080]:   notice: tengine_stonith_notify: Peer 
>>>>> dev1n2 was terminated (st_notify_fence) by dev1n1 for dev1n1: OK 
>>>>> (ref=71447261-0e53-4b20-b628-d3f026a4ae24) by client stonith-api.27225
>>>>> 
>>>>> Greets,
>>>>>         Jacek
>>>>> 
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster@redhat.com
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>> 
>>>> 
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster@redhat.com
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>> 
>>> 
>>> 
>>> -- 
>>> esta es mi vida e me la vivo hasta que dios quiera
>>> 
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster@redhat.com
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>> 
>> 
>> --
>> Linux-cluster mailing list
>> Linux-cluster@redhat.com
>> https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> 
> 
> -- 
> esta es mi vida e me la vivo hasta que dios quiera
> -- 
> Linux-cluster mailing list
> Linux-cluster@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
-- 
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to