Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
rb:sanconfig-upgrade [10/10/2013 15:21]
127.0.0.1 external edit
rb:sanconfig-upgrade [19/12/2018 16:25] (current)
andrew [Proposed Toplology change]
Line 5: Line 5:
 If this is met, it will be safe to portdisable all server connected ports on the switch to be upgraded prior to disabling the switch. See this link below for the status of all SAN connected hosts and if they have been tested individually. ​ If this is met, it will be safe to portdisable all server connected ports on the switch to be upgraded prior to disabling the switch. See this link below for the status of all SAN connected hosts and if they have been tested individually. ​
  
 +It's worth pointing out that during the update process, fabric redundancy will be lost, as only one fabric will be active, the other will be undergoing the update so a failure of the working fabric will result in data loss. 
 +So even if all hosts have been proved to be able to switch paths to the 2nd fabric, it's best to do this at a quiet time.
  
 ===== SAN connected servers ===== ===== SAN connected servers =====
Line 25: Line 26:
 ===== Brocade firmware location ===== ===== Brocade firmware location =====
  
-Brocade firmware can be found at:- satellite02:/​firmware/​BrocadeFirmware. This is not web or ftp available, you will need to use sftp. The archive is read for all users.+Brocade firmware can be found at:- satellite02:/​firmware/​BrocadeFirmware. This is not web or ftp enabled, you will need to use sftp. The archive is read only for all users.
  
  
Line 54: Line 55:
  
           ​           ​
-Failing sw02 will break db connectivity in DC02 on sw1 (RVB requires storage on wager02 ​off sw4) and db connectivity on sw3 (RGS links through sw01, sw02, sw03 to access ​wager02) as well as db connectivity on sw2 itself (RGS needs storage on wager02 ​off sw4).  Apart from sw3, any upgrades will disrupt the whole fabric.+Failing sw02 will break db connectivity in DC02 on sw1 (app1 requires storage on array02 ​off sw4) and db connectivity on sw3 (app2 links through sw01, sw02, sw03 to access ​array02) as well as db connectivity on sw2 itself (app2 needs storage on array02 ​off sw4).  Apart from sw3, any upgrades will disrupt the whole fabric.
  
-Sw02 has the oldest firmware and probably should be done first. But this risks the most disruption. ​+sw02 has the oldest firmware and probably should be done first. But this risks the most disruption. ​
 To do this safely, we should disable all server ports on the fabric on connected switches to control the order in which we loose paths and to check that storage connectivity to the hosts systems is maintained through the alternate fabric. To do this safely, we should disable all server ports on the fabric on connected switches to control the order in which we loose paths and to check that storage connectivity to the hosts systems is maintained through the alternate fabric.
  
 sw01 has the next oldest firmware and also is the fabric principal. The backup window may be disrupted because all the tape drives are connected to FAb1 on this switch. sw01 has the next oldest firmware and also is the fabric principal. The backup window may be disrupted because all the tape drives are connected to FAb1 on this switch.
  
-sw03 is already on a 6.x release, upgrading this only takes 2 RGS db servers off.+sw03 is already on a 6.x release, upgrading this only takes 2 app02 db servers off.
    
-sw04 is the most current release firmware, but taking this offline has an impact to ALL SAN connected servers because all the storage at present is on Wager02 ​which is only connected to sw04.+sw04 is the most current release firmware, but taking this offline has an impact to ALL SAN connected servers because all the storage at present is on array02 ​which is only connected to sw04.
 Also connected to switch4 is the MSA and dbr01. The MSA is not an active/​active storage device, so failing one fabric may force an interruption to traffic. It is probably best to stop Oracle on dbr01 as a precaution. Also connected to switch4 is the MSA and dbr01. The MSA is not an active/​active storage device, so failing one fabric may force an interruption to traffic. It is probably best to stop Oracle on dbr01 as a precaution.
  
Line 104: Line 105:
  
 One of the limits is that there are only two fibre interconnects between DC2 and DC4. The partially meshed design below is resiliant to any one link failing and  limits hops to a maximum of one except for traffic from sw3 to sw1 (in DC02), whilst not using more than the two existing links between DC02 and DC04. Failure of sw01 or sw04 will still cause disruption as these switches have the storage attached to them. One of the limits is that there are only two fibre interconnects between DC2 and DC4. The partially meshed design below is resiliant to any one link failing and  limits hops to a maximum of one except for traffic from sw3 to sw1 (in DC02), whilst not using more than the two existing links between DC02 and DC04. Failure of sw01 or sw04 will still cause disruption as these switches have the storage attached to them.
-One additional ​sft is required for sw2 on each fabric to link sw2 to sw4. A spare sft slot is available on sw2.+One additional ​sfp is required for sw2 on each fabric to link sw2 to sw4. A spare sfp slot is available on sw2.
  
  

rb/sanconfig-upgrade.1381414873.txt.gz · Last modified: 10/10/2013 15:21 by 127.0.0.1