No relevant resource is found in the selected language.

This site uses cookies. By continuing to browse the site you are agreeing to our use of cookies. Read our privacy policy>Search

Reminder

To have a better experience, please upgrade your IE browser.

upgrade

FusionStorage OBS 7.0 Parts Replacement 05

Rate and give feedback:
Huawei uses machine translation combined with human proofreading to translate this document to different languages in order to help you better understand the content of this document. Note: Even the most advanced machine translation cannot match the quality of professional translators. Huawei shall not bear any responsibility for translation accuracy and it is recommended that you refer to the English document (a link for which has been provided).
Replacing a Service Disk Module

Replacing a Service Disk Module

As storage media, service disk modules are hot-swappable.

Impact on the System

When a service disk module is replaced, system performance is compromised. Therefore, replace a service disk module during off-peak hours.

Prerequisites

  • A spare disk module is ready. The capacity of the spare disk module is equal to or larger than that of the faulty disk module, and the type of the spare disk module is the same as that of the faulty disk module.
  • The faulty disk module has been located.
  • The storage pool to which the faulty disk module belongs is in the normal state, and no data reconstruction task is running.
NOTE:

For details about the slot numbers of service disk modules, see Slot Numbers.

Precautions

  • To prevent damaging disk modules or connectors, remove or install disk modules with even force.
  • When removing a disk module, first remove it from its connector. Wait at least 30 seconds and then remove the disk module completely from the chassis.
  • To prevent disk module damage, wait at least one minute between removal and insertion actions.
  • To avoid data loss, replace only the disk module whose fault indicator is yellow on.
  • To avoid system failures, do not reuse disk modules.
  • SSDs and HDDs (including NL-SAS, SAS, and SATA disks) cannot be preserved for a long time after being powered off.
    • SSDs with no data stored can be preserved for a maximum of 12 months while they are powered off. SSDs with data stored can be preserved for a maximum of three months while they are powered off. If the maximum preservation time is exceeded, data loss or SSD failures may occur.
    • Packed HDDs can be preserved for a maximum of six months. Unpacked HDDs that are powered off can also be preserved for a maximum of six months. If the maximum preservation time is exceeded, data loss or disk failures may occur.
    • The maximum preservation time is determined based on the disk preservation specifications provided by vendors. For details, see the manuals provided by the vendors.

Tools and Materials

  • ESD gloves
  • ESD wrist straps
  • ESD bags
  • Labels

Replacement Process

Replace a service disk module by following the process shown in Figure 2-1.

Figure 2-1 Service disk module replacement process

Procedure

  1. Query the ID of the storage pool to which the faulty disk belongs.

    1. Log in to the CLI of the primary management node as user dfvmanager, and run the following command to obtain the storage pool ID. To run this command, you need to enter the user name and password of CLI super administrator account admin.

      sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryStoragePool

      Information about all storage pools is displayed in the command output. poolId in the leftmost column displays IDs of all storage pools.

      [dfvmanager@node0101 ~]$ sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryStoragePool
      [Thu Dec 20 09:44:35 CST 2018] DswareTool operation start.
      Enter User Name:admin
      Enter Password :
      Operation finish successfully. Result Code:0
      Dec 20, 2018 9:44:44 AM com.huawei.dfv.persistent.oam.client.cmd.QueryStoragePool handleSuccessResult
      INFO:
      poolId poolName totalCapacity phy(MB) usedCapacity phy(MB) freeCapacity logic(MB) thinRate thinThreshold azProperty protectMode routingMode slowIoSwitch replicationFactor poolGroupName poolGroupId
      --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
      0 pool_sata_ec 130351795 449696 86601399 0 70 inner ec online null 0 default 0
      
      When there is no dsware client or failed to query storage pool capacity, the totalCapacity/usedCapacity will be set as 0.
      [Thu Dec 20 09:44:45 CST 2018] DswareTool operation end.
    2. Run the sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryStorageNodeInfo -id poolId command to query the ID of the storage pool to which the faulty service disk belongs. To run this command, you need to enter the user name and password of CLI super administrator account admin.
      If the command output contains the storage plane IP address of the faulty node, poolId in the command is the ID of the storage pool to which the faulty service disk belongs.
      [dfvmanager@node0101 ~]$ sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryStorageNodeInfo -id 0
      [Wed Dec 19 11:36:38 CST  2018] DswareTool operation start.
      Enter User Name:admin
      Enter Password  :
      Operation finish successfully. Result Code:0
      The result as  follow:
      nodeMgrIp:192.168.124.6 poolId:0 nodeType:0 rack:2  subRack:null
      diskInfo:
      diskSn:K4KDYJRB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:0
      diskSn:K4KDL7JB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:1
      diskSn:K3GEU03B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:2
      diskSn:K4J7KZRB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:6
      diskSn:K4KEBS2B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:7
      diskSn:K4KDL46B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:8
      diskSn:K4JK3K1B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:12
      diskSn:K4J7TX2B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:13
      diskSn:K4J7X04B diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:14
      diskSn:K4JK3JJB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:18
      diskSn:K4J0DZBB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:19
      diskSn:K4J7NBNB diskType:5 diskSize:3726 diskUse:1  diskStatus:0 diskSlot:20
      cacheInfo:
      cacheEsn:031YSVFSJ7000600 cacheType:3  cacheSize:1600 usedSize:1596 cacheStatus:3
      cacheEsn:031YSVFSJ6000940 cacheType:3  cacheSize:1600 usedSize:1596 cacheStatus:3
      ......

  2. Run the following command to check the view status of partitions. To run this command, you need to enter the user name and password of CLI super administrator account admin.

    sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryRebalanceInfo -id ID of the storage pool to which the faulty service disk belongs

    If a record whose status is unok exists, the partition views are abnormal. Replace the disk after all partition views are normal.

  3. Remove the disk module.

    Correctly record the slots where disk modules reside. Install disk modules into the same slots before and after the replacement. Otherwise, services may be affected.

    1. Press the button that secures the disk module ejector lever, as shown in step 1 in Figure 2-2.

      The ejector lever automatically ejects.

      Figure 2-2 Removing a disk module
    2. Hold the ejector lever, and pull out the disk module for approximately 3 cm, as shown in step 2 in Figure 2-2.
    3. Wait at least 30 seconds until the disk stops spinning, and slowly pull out the disk module, as shown in step 3 in Figure 2-2.

  4. Place the removed disk module in an ESD bag.
  5. Take the spare disk module out of its ESD bag.
  6. Install the disk module.

    Install disk modules into the same slots before and after the replacement. Otherwise, services may be affected.

    1. Raise the ejector lever and push the disk module in along the guide rails until it does not move, as shown in step 1 in Figure 2-3.
      Figure 2-3 Installing a disk module
    2. Ensure that the ejector lever is fastened to the beam, and lower the ejector lever to completely insert the disk module into the slot, as shown in step 2 in Figure 2-3.

  7. Check whether the disk module is successfully installed.

    Wait 3 minutes and then check the activity and fault indicators on the disk to determine whether the disk module is successfully installed.
    • If the activity indicator is steady on or blinking green and the fault indicator is off or blinking yellow, the installation is successful.
    • If the fault indicator is steady yellow, the installation fails. In this case, perform the following operations:
      • Check whether the disk module has been firmly inserted. If not, wait 1 minute and then reinsert it.
      • If the fault indicator is still steady yellow after you have reinserted the disk module, the spare disk module may be faulty. In this case, replace it with another one.
      • If the fault indicator is still steady yellow after the replacement, the link of the disk slot may be faulty. In this case, contact technical support engineers.

  8. Add the spare service disk into the storage pool.

    1. Run the following command to check whether the faulty disk is removed from the storage pool. To run this command, enter the name and password of CLI super administrator account admin as prompted:

      sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryNodeProcessInfo -ip Storage plane floating IP address of the faulty node

      NodeMgrIp      ProcessName        ProcessInfo                  ProcessStatus     
      ------------------------------------------------------------------------------------------
      192.168.124.6         
                       ZK              /persistence_layer_zk_data          OK   
                       MDC                                                 OK  
                       OSD             Z1P0MNC7                            OK 
                       OSD             9WM1LYFJ                            OK     
                       OSD             WD-WCAY00520829                     OK      
                       OSD             WD-WMAY03873314                     OK    
                       OSD             Z1P0MN0J                            OK       
                       OSD             Z1P0JWTQ                            OK    

      If a record whose ProcessInfo is Esn of the faulty disk exists, the disk is not removed from the storage pool. In this case, run the following command to remove the faulty disk from the storage pool:

      sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op fastKickOutDisk -id ID of the storage pool to which the faulty service disk belongs -esn ESN of the faulty service disk -redundancy 1 -nodeMgrIp Storage plane floating IP address of the faulty node

    2. Run the following command to add the spare service disk into the storage pool. To run this command, enter the name and password of CLI super administrator account admin as prompted:

      sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op replaceDisk -id ID of the storage pool to which the faulty service disk belongs -slot Slot ID of the faulty service disk -nodeMgrIp Storage plane floating IP address of the faulty node

      If the command is executed successfully, taskId of the background disk replacement task is returned.

    3. Run the following command to query the progress of the background disk replacement task. To run this command, enter the name and password of CLI super administrator account admin as prompted:

      sh /opt/dfv/oam/oam-p/client/bin/dswareTool.sh --op queryTaskLogInfo -taskId taskId of the background disk replacement task

      If all steps are successful, the disk replacement is successful.

      NOTE:

      If the task fails at the Add osd to master mdc. step and the error message 32501012:FSA failed to communicate with the MDC is returned, the primary MDC node may be in migration. In this case, you can perform 2 again.

  9. Check the system status.

    On SmartKit, choose Home > Storage > Routine Maintenance > More > Inspection and check the system status.
    • If all inspection items pass the inspection, the inspection is successful.
    • If some inspection items fail, the inspection fails. Rectify the faults by taking recommended actions in the inspection reports. Perform inspection again after fault rectification. If the inspection still fails, contact Huawei technical support.

    For details, see the FusionStorage OBS Administrator Guide.

Follow-up Procedure

Label the replaced service disk module to facilitate subsequent operations.

Translation
Download
Updated: 2019-07-05

Document ID: EDOC1100051325

Views: 5033

Downloads: 2

Average rating:
This Document Applies to these Products
Related Documents
Related Version
Share
Previous Next