Start a Conversation

Unsolved

DG

1 Rookie

 • 

10 Posts

76

October 22nd, 2023 21:24

MD3260: Failed Thin virtual Disk with no hints what the problem is, help!

Hello,
Really hoping for some hints as we facing 230TB recovery otherwise..
We have two MD3260 enclosures with 4 thin virtual disks that are bundled in a single 4x60T LVM volume, then XFS on it - actually 3x60TB+1x58TB and 3TB available.
We approached recently 230TB and we got critical problem, with tons of hardware errors - problem is the SMcli says that the smaller Virtual Thin Disk is problematic but without any hints and all hardware components seems to be ok (disks and controller):
SMcli -n aspera_md_1 -c 'show storagearray healthStatus;'Performing syntax check...Syntax check complete.Executing script...The following failures have been found:Thin Virtual Disk FailedStorage array: aspera_md_1Disk pool: Disk_Pool_1Thin Virtual Disk: gaia_virtual_p2Status: Failed
On the OS side we have plenty of IO errors in the kernel:
Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#0 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#0 CDB: Write(16) 8a 00 00 00 00 00 00 3d e0 f8 00 00 00 08 00 00[Sun Oct 15 05:24:10 2023] blk_update_request: 60 callbacks suppressed[Sun Oct 15 05:24:10 2023] blk_update_request: critical target error, dev sdg, sector 4055288[Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#1 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#1 Sense Key : Hardware Error [current][Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#1 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:10 2023] sd 9:0:0:5: [sdg] tag#1 CDB: Write(16) 8a 00 00 00 00 00 01 23 fa f8 00 00 00 08 00 00[Sun Oct 15 05:24:10 2023] blk_update_request: critical target error, dev sdg, sector 19135224[Sun Oct 15 05:24:10 2023] blk_update_request: critical target error, dev dm-3, sector 4055288[Sun Oct 15 05:24:10 2023] blk_update_request: critical target error, dev dm-3, sector 19135224[Sun Oct 15 05:24:10 2023] XFS (dm-7): metadata I/O error in "xfs_buf_iodone_callback_error" at daddr 0x7bb2f8 len 8 error 121[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 CDB: Write(16) 8a 00 00 00 00 05 7f ff fc 01 00 00 00 01 00 00[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev sdg, sector 23622319105[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#3 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#3 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#3 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#3 CDB: Write(16) 8a 00 00 00 00 05 7f ff fc 02 00 00 00 01 00 00[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev sdg, sector 23622319106[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev dm-3, sector 23622319105[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev dm-3, sector 23622319106[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 CDB: Write(16) 8a 00 00 00 00 00 00 3d e0 f8 00 00 00 08 00 00[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev sdg, sector 4055288[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 CDB: Write(16) 8a 00 00 00 00 00 01 23 fa f8 00 00 00 08 00 00[Sun Oct 15 05:24:11 2023] blk_update_request: critical target error, dev sdg, sector 19135224[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#2 CDB: Write(16) 8a 00 00 00 00 05 7f ff fc 01 00 00 00 02 00 00[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 Sense Key : Hardware Error [current][Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 <<vendor>>ASC=0x84 ASCQ=0x0[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#0 CDB: Write(16) 8a 00 00 00 00 00 00 3d e0 f8 00 00 00 08 00 00[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s[Sun Oct 15 05:24:11 2023] sd 9:0:0:5: [sdg] tag#1 Sense Key : Hardware Error [current]
It seems that VD size id 60TB but actual size of the thin VD is 58TB..
We tried several things:
1. increasing the size of the repo associated with this thin vd, but it errors out (we have one spare that could be used for this). There are no details shown with the error (SMClient)
2. Run the test, with the error below...
SMcli -n aspera_md_1 -c 'check virtualDisk [gaia_virtual_p2] repositoryConsistency file="/tmp/p2_consistency.txt";'Performing syntax check...Syntax check complete.Executing script...Script execution complete.SMcli completed successfully.[root@gaiaftp01: disks]$ less /tmp/p2_consistency.txtThinRepo 0xc002 State:FAILED RollbackState:NONEImage 5(xc002) validation startingError: Dir2CN:0x000000440bbb(Current) Bad signatureError: Dir2CN:0x000000440bbb(Current) Bad Level Found:0x47 Expected:0x32Error: Dir2CN:0x000000440bbb(Current) Bad Location DirA:0x5a3c806b33570a1d DirB:0xdb8e591145f3a3a2Error: Dir2CN:0x000000440bbb(Current) Cluster out of bounds DataCN:0x67fa78c873cc Offset:0x30Dumping Dir Level 2 (Current) CN:x000000440bbb LBA:x0002205dd800x0000 | ...].{m8..3Gr... | 7F94AB5D 987B6D38 DF1B3347 72CCF7E0 |0x0010 | ..W3k.<Z...E.Y.. | 1D0A5733 6B803C5A A2A3F345 11598EDB |0x0020 | ...*R....}....h. | E0A4D72A 52CEB290 B87DB292 E6C9689C |0x0030 | .s.x.g{.O@]..pIk | CC73C878 FA677BCE 4F405DA5 9670496B |0x0040 | .^.n....1wY..... | BC5E076E FC8DAA10 31775915 B20308B8 |0x0050 | .7j.(i.4...yme.. | 99376AF2 28698234 84E0F879 6D65F112 |0x0060 | ..Uf].Eo.V.....e | DFED5566 5DED456F C356D485 10C19565 |0x0070 | ...Y..yK.*...i.. | F4939759 0181794B 9E2AB1E6 F069EFE9 |0x0080 | N..&........T..f | 4E8BCE26 04010A91 819DDEB8 54DF9B66 |0x0090 | (<...>...m...@.f | 283CFA09 FB3E961D 976D9CCB 1040DB66 |0x00a0 | ~........;...... | 7EDC07E8 07ECE2E9 C43BEDFE FBBFD2B1 |0x00b0 | ......&.uo5.M5m. | 8DCBEDDA F1F62692 756F35FE 4D356D8A |0x00c0 | .....G.....2.... | E5F49EE1 F347C5F5 BC9ED632 E8C5B815 |0x00d0 | H6{.lv...J..<.]. | 48367B18 6C76DD19 B84AA887 3CA05DA1 |0x00e0 | ....Bc`....5V].. | 83D7D00D 42636018 C89CAF35 565DDD1F |0x00f0 | ....X..B.F/.V..1 | 9DA4A4F5 581AFE42 F1462FB9 56A1CB31 |0x0100 | '.{"...E..).."0. | 27EC7B22 7FD3C945 1FC229A0 0322301D |0x0110 | ..CF..(...N..Wx. | 041A4346 D18D2894 CAFA4EC1 8B577817 |0x0120 | 3..5Z$.!.j#JSf.} | 33FCCA35 5A240821 9A6A234A 5366E27D |0x0130 | .......j..-..... | A49A1FBE ED0BA56A A6E32DB2 D4AD07E9 |0x0140 | .I..z;..'.,..Zr. | DD49B38A 7A3B97B7 27972CA3 0B5A7204 |0x0150 | .^t..)...6..n.'. | F45E74D4 8229180C 0A3682C3 6E0B27B2 |0x0160 | .A.%3G..*AD..;.) | 8A41B125 3347B995 2A4144F8 C23BCF29 |0x0170 | ..-.....*...N.:. | 1F8C2DBB B2078718 2AB3EBEA 4EC63AF1 |0x0180 | .N.~...-y..`.5.. | 8D4E017E 82C2852D 79DDEB60 C535D4EA |0x0190 | ,.....Q.).....V+ | 2CCBEC04 12C451A1 29C8D004 BF88562B |0x01a0 | ..K>.....j....zl | B1824B3E 7FE9D4C2 846AA094 E7157A6C |0x01b0 | ..S...Vx...K...a | 0EDB53A3 BD985678 BAF8BA4B 149E8661 |0x01c0 | .."...+.J....... | FE032203 E2B82BD6 4AAAB813 88AFDAE5 |0x01d0 | ....hi.RQ.C..." | 0DEDB0EF 6869E252 51DC43BD DC1B2220 |0x01e0 | ....0.).../+.Jv{ | F6BDFEE6 30AE29FA 8E022F2B A44A767B |0x01f0 | .^.p.G%,.~.).Su. | 155ECB70 F047252C 117E9229 905375A6 |Directory Structure - Dir Level 0 CN:x000000000005 Offset:x0Dir Level 1 CN:x000000000007 Offset:x72Dir Level 2 CN:x000000440bbbImage 5(xc002) validation failed with 4 errors 0 warningsValidation error limit reached, stopping validationThinRepo 0xc002 validation failed with 4 errors 0 warnings
Please help as we are really out of ideas. The logical try was to expand thin v-disk repo, but it errored out - also no hints why.. Is there any chance to make it back alive?
The deeper problem is that we do not know if it's really a non-reported hardware error or not as we would not like to dump the enclosures - but we cannot be sure the same situation does not resurface if we delete the failed VD (and loose all 4 VDs LVM/XFS  volume)..
Regards,
Chris

Moderator

 • 

3.1K Posts

October 23rd, 2023 03:02

Hi,

 

Probably you can try to check if the error output can lead you to some information what is the cause of the storage not accessible: https://dell.to/3tJpkSs or, you can also try check the physical disk info status, probably multiple disk failure: https://dell.to/3FtV3K3

1 Rookie

 • 

10 Posts

October 23rd, 2023 08:38

Thank you for the hints @DELL-Joey C​ !

However, what's puzzling is that there are no hw errors reported, I am attaching the excerpt from allPhysicalDisks:

SUMMARY
   Number of physical disks: 60
      Current media type(s): Physical Disk (60)
      Current interface type(s): Serial Attached SCSI (SAS) (60)

   BASIC:

      ENCLOSURE, DRAWER, SLOT  STATUS   CAPACITY      MEDIA TYPE     INTERFACE TYPE  CURRENT DATA RATE  PRODUCT ID        FIRMWARE VERSION  CAPABILITIES
      0,         0,      0     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      1     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      2     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      3     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      4     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      5     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST33000650SS      RS17
      0,         0,      6     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      7     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      8     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      9     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      10    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         0,      11    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      0     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      1     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      2     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      3     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      4     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      5     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      6     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      7     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      8     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      9     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      10    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         1,      11    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      0     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      1     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      2     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      3     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST33000650SS      RS17
      0,         2,      4     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      5     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      6     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      7     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      8     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      9     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      10    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         2,      11    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      0     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             MG03SCA300        DG09
      0,         3,      1     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      2     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      3     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      4     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      5     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      6     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      7     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      8     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      9     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      10    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         3,      11    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      0     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      1     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      2     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      3     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      4     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      5     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      6     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      7     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      8     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      9     Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      10    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15
      0,         4,      11    Optimal  2,794.519 GB  Physical Disk  SAS             6 Gbps             ST3000NM0023      GS15

Otherwise, exerpt gave the hint why the repo epansion failed, not sure if there's a newer firmware though - are you aware of any newer version that would support underlying repository thin VD expansion?

OPERATING ENVIRONMENT DATA

Client system name: gaiaftp01.unige.ch
Client OS: Linux

Application version: 11.20.0G06.0020
Storage array management version: 11.20.0G06.0020
Storage array name: aigmd00
Firmware version: 08.20.24.60
Management class: devmgr.v1120api13.Manager

**************************************************

ERROR DATA

Command sent to RAID controller module in slot: 0
Host name: 0.0.0.0
IP address: 0.0.0.0

Return code: Error 1026 - The command issued to the storage array did not complete because the current RAID controller module firmware does not support this function.
Operation when error occurred: <<Unknown Procedure>>
Timestamp: Oct 22, 2023 11:06:16 PM

STACK DATA

devmgr.v1120api13.sam.jal.ManagementOperationFailedException: Error 1026 - The command issued to the storage array did not complete because the current RAID controller module firmware does not support this function.
        at devmgr.v1120api13.sam.jal.SYMbolClient.createExceptionWithMessage(SYMbolClient.java:1030)
        at devmgr.v1120api13.sam.jal.SYMbolClient.createException(SYMbolClient.java:1009)
        at devmgr.v1120api13.sam.jal.SYMbolClient.dispatchOperation(SYMbolClient.java:606)
        at devmgr.v1120api13.sam.jal.StorageArrayFacade.issueCommand(StorageArrayFacade.java:10054)
        at devmgr.v1120api13.sam.jal.StorageArrayFacade.sendCommandCommon(StorageArrayFacade.java:9610)
        at devmgr.v1120api13.sam.jal.StorageArrayFacade.sendCommand(StorageArrayFacade.java:9507)
        at devmgr.v1120api13.sam.jal.StorageArrayFacade.getReadLinkStatusDiagnostics(StorageArrayFacade.java:9199)
        at devmgr.v1120api13.sam.scriptengine.SArrayExecutor.getReadLinkStatusDiagnostics(SArrayExecutor.java:2446)
        at devmgr.v1120api13.sam.scriptengine.SArrayExecutor.performUploadOperation(SArrayExecutor.java:1337)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.uploadStatement(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.simpleStatement(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.statement(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.statementList(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.execute(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptParser.execute(Unknown Source)
        at devmgr.v1120api13.sam.scriptengine.ScriptEngine.execute(ScriptEngine.java:87)
        at devmgr.dmv.pm.PMUtilities.dataCollection(PMUtilities.java:270)
        at devmgr.dmv.pm.PMDataCollector.run(PMDataCollector.java:133)

Regards,

Chris

Moderator

 • 

3.1K Posts

23-10-2023 10:17 AM

Hi,

 

It could be the application version, Storage Manager. Try updating it, your current version is 11.20.0G06.0020, based on the support matrix https://dell.to/46Z1WP3 it is 11.25.0306.0026. 

 

What OS are you running and what HBA are you using?

DELL-Joey C

Social Media and Communities Professional

Dell Technologies | Enterprise Support Services

#IWork4Dell

Did I answer your query? Please click on ‘Mark as Accepted Answer’. ‘Thumbs up’ the posts you like!

1 Rookie

 • 

10 Posts

October 24th, 2023 11:28

Hello @DELL-Joey C 

Thanks, will try to update the Storage Manager and retry repo expansion!

The OS is Centos 7.9, machine is R720.

HBAs, two controllers, one per MD:
Serial Attached SCSI controller: Broadcom / LSI SAS2008 PCI-Express Fusion-MPT SAS-2 [Falcon] (rev 03)
Subsystem: Dell 6Gbps SAS HBA Adapter
Serial Attached SCSI controller: Broadcom / LSI SAS2008 PCI-Express Fusion-MPT SAS-2 [Falcon] (rev 03)
Subsystem: Dell 6Gbps SAS HBA Adapter
4 paths, one per VD:
multipath -ll
mpathr (36f01faf000d65e31000006235a3409ee) dm-2 DELL ,MD32xx
size=60T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
|-+- policy='round-robin 0' prio=14 status=active
| `- 1:0:0:0 sdb 8:16 active ready running
`-+- policy='round-robin 0' prio=9 status=enabled
`- 1:0:1:0 sdd 8:48 active ready running
mpathe (36f01faf000e5bfab000009a53ee41a12) dm-4 DELL ,MD32xx
size=63T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
|-+- policy='round-robin 0' prio=14 status=active
| `- 9:0:3:4 sdh 8:112 active ready running
`-+- policy='round-robin 0' prio=9 status=enabled
`- 9:0:4:4 sdf 8:80 active ready running
mpaths (36f01faf000d65e270000073e5a3347d0) dm-3 DELL ,MD32xx
size=61T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
|-+- policy='round-robin 0' prio=14 status=active
| `- 1:0:1:1 sde 8:64 active ready running
`-+- policy='round-robin 0' prio=9 status=enabled
`- 1:0:0:1 sdc 8:32 active ready running
mpathf (36f01faf000e5bfb90000093056ccf4dd) dm-5 DELL ,MD32xx
size=63T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
|-+- policy='round-robin 0' prio=14 status=active
| `- 9:0:4:5 sdg 8:96 active ready running
`-+- policy='round-robin 0' prio=9 status=enabled
`- 9:0:3:5 sdi 8:128 active ready running
More complete status :
PROFILE FOR STORAGE ARRAY: aspera_md_1 (Tue Oct 24 12:52:58 CEST 2023)
STORAGE ARRAY------------------------------
STORAGE ARRAY INFORMATION AND SETTINGS
Storage array world-wide identifier (ID): 6F01FAF000E5BFBF0000000052CC6FE0
Chassis Serial Number: 
Event configuration data version: 139#Dell_1120_16
CACHE SETTINGS
Start demand cache flushing at: 90%
Cache block size: 32 KB
Media scan frequency: 15 days
Failover alert delay: 5 minutes
STORAGE SUMMARY
Disk pools: 1
Virtual Disks on Disk Pools: 4
Disk groups: 0
Access virtual disks: 1
Standard Virtual Disks (Used/Allowed): 2 / 512
Base: 0
Repository: 2
Thin Virtual Disks (Used/Allowed): 2 / 512
SSD CACHE
Total virtual disks using SSD cache: 0
Maximum system-wide capacity allowed: 3,725.290 GB
HOST MAPPINGS SUMMARY
Access virtual disk: LUN 31,31 (see Mappings section for details)
Default host OS: Linux (DM-MP) (Host OS index 1)
Mapped virtual disks: 3
Unmapped virtual disks: 0
HARDWARE SUMMARY
Enclosures: 1
System configured to use batteries: Yes
RAID Controller Modules: 2
Consistency mode: Duplex (dual RAID controller modules)
Physical Disks: 60
Current physical disk media types: Physical Disk (60)
Current physical disk interface type(s): Serial Attached SCSI (SAS) (60)
Total hot spare physical disks: 0
Standby: 0
In use: 0
Physical Disk security key identifier: None
FEATURES SUMMARY
Feature enable identifier: 31540000000031590000000052CC6FE1
Feature pack: SAS MD3260
Feature pack submodel ID: 185
Feature Status Used Allowed
Snapshot (Legacy) Virtual Disk Disabled N/A N/A
Virtual Disk Copy Disabled N/A N/A
Physical Disk Slots N/A 60 120
Snapshot Feature 90 day trial available
Snapshot groups (see note below) N/A N/A
Snapshot images used N/A N/A
Snapshot virtual disks N/A N/A
Consistency groups N/A N/A
Member Virtual Disks (see note below) N/A N/A
Additional feature information
Snapshot groups allowed per source virtual disk (see note below): 4
Snapshots(legacy) allowed per source virtual disk: 0
Virtual Disks allowed per storage partition: 256
Note: If a virtual disk is a member of a consistency group, that membership (member virtual disk) counts against both the overall snapshot group limit and snapshot group per source virtual disk limit.
Automatic critical event collection: Yes
Existing support data schedule: No
Suspended: NA
Schedule Settings
Start date: NA
End Date: NA
Start Time: NA
Frequency: NA
Collected On: NA
FIRMWARE INVENTORY
MD Storage Manager®
AMW Version: 11.20.0G06.0020
Report Date: Tue Oct 24 12:52:58 CEST 2023
Storage Array
Storage Array Name: aspera_md_1
Current Package Version: 08.20.24.60
Current NVSRAM Version: N26X0-820890-008
Staged Package Version: None
Staged NVSRAM Version: None
RAID Controller Modules
Location: Enclosure 0, Slot 0
Current Package Version: 08.20.24.60
Current NVSRAM Version: N26X0-820890-008
Board ID: 2660
Sub-Model ID: 185
Location: Enclosure 0, Slot 1
Current Package Version: 08.20.24.60
Current NVSRAM Version: N26X0-820890-008
Board ID: 2660
Sub-Model ID: 185
Power Supplies
Location: Enclosure 0 Top
Firmware Version: 02.47
Location: Enclosure 0 Bottom
Firmware Version: 02.47

(edited)

1 Rookie

 • 

10 Posts

October 26th, 2023 13:39

Hello,

We managed to expand the underlying repository capacity by enabling manual expansion and then adding capacity in chunks of 256GB. But it did not help, think VD is still marked as failed.

We are inclining to reset it, but it's a pity neither reason nor explanation are given for this error..

(edited)

Moderator

 • 

2.2K Posts

26-10-2023 14:17 PM

Hi, okay I just reviewed the thread. Before resetting it, I'm not sure but might be you can try to a few SMcli command to check physical or bad sectors on any drive in pool. To check status of each drive and see bad sectors, you can use show physicalDisk [physicalDiskID] Then you can also use the SMcli command set physicalDisk [physicalDiskID] mediaScanAction=repair; to attempt to repair any errors on the drive. 

DELL-Erman O

Social Media and Communities Professional

Dell Technologies | Enterprise Support Services

#IWork4Dell

Did I answer your query? Please click on ‘Mark as Accepted Answer’. ‘Thumbs up’ the posts you like!

1 Rookie

 • 

10 Posts

October 27th, 2023 14:49

@Dell-ErmanO​ - thanks for the hints, before I saw your suggestion we tried to run some Reports from SMClient to see anything unusual, and indeed we see strange error in the event log (even drives are marked as ok)

Will do the test on all the disks of this VD as well, maybe we'll see more surprises..

the details of each of the event look like
ate/Time: 10/26/23 4:11:12 PM
Sequence number: 2199718
Event type: 2046
Event category: Internal
Priority: Critical
Event needs attention: true
Event send alert: true
Event visibility: true
Description: Isolation of physical disk causing consistency mismatch
Event specific codes: 0/0/0
Component type: Physical Disk
Component location: Enclosure 0, Drawer 2, Slot 3
Logged by: RAID Controller Module in slot 1

Raw data:
4d 45 4c 48 03 00 00 00 a6 90 21 00 00 00 00 00
46 20 41 01 80 73 3a 65 1b 00 01 00 00 80 00 00
01 00 00 00 04 00 00 00 22 00 00 00 22 00 00 00
01 00 00 00 00 00 00 00 04 00 00 00 02 00 00 00
21 00 00 00 03 00 00 00 0a 00 00 00 01 00 00 00
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
01 00 00 00 00 00 01 03 44 00 00 00 20 00 20 06
5a 32 39 33 56 4c 5a 52 00 00 00 00 00 00 00 00
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
08 00 20 86 00 b1 cc 44 00 00 00 00 10 00 43 87
50 00 c5 00 42 3c 18 e7 00 00 00 00 00 00 00 00​

1 Rookie

 • 

10 Posts

October 27th, 2023 22:26

Hello @Dell-ErmanO​ 

Tried to run "set physicalDisk" command but at no avail:


SMcli -n aspera_md_1 -c   "set physicalDisk[0,0,0] mediaScanAction=repair ;"
Performing syntax check...

Encountered "mediaScanAction" at line 1, column 39.
Was expecting one of:
    "hotspare" ...
    "operationalState" ...
    "serviceAllowedIndicator" ...
    "nativeState" ...


SMcli failed.


This is with updated SMcli (still from 2016)..

Moderator

 • 

3.1K Posts

30-10-2023 01:47 AM

Hi,

 

Based on the error message, there are certain sets of command that we can run to check if it can fix the issue, but the commands are in Windows format. I'm unsure for Centos and the CLi version that you're on, is it 11.25.0306.0026, that you wanted to update it to.

 

https://dell.to/49fu9U0

 

1) Perform a parity scan on the affected volumes as follows: Page 115, check virtualDisk [VD1] consistency consistencyErrorFile="C:\Temp\ConsistencyError.txt" mediaErrorFile="C:\Temp\mediaError.txt" priority=high verbose=true;

 

2) After this scan is completed, use the Error file to perform a parity repair on the affected volumes as follows. Page 259, repair virtualDisk [VD1] consistency consistencyErrorFile="C:\Temp\ConsistencyError.txt" verbose=true;

 

If no Parity Error file is generated, then no errors are identified. During this scan, if new errors are encountered, additional events might be logged against this drive. If there are any additional errors, then replace the drive.

DELL-Joey C

Social Media and Communities Professional

Dell Technologies | Enterprise Support Services

#IWork4Dell

Did I answer your query? Please click on ‘Mark as Accepted Answer’. ‘Thumbs up’ the posts you like!

1 Rookie

 • 

10 Posts

November 1st, 2023 15:00

Hello,

We are on  Modular Disk Storage Manager  11.25.0A06.0026 which seems to be newer, it's from DELL_MDSS_Consolidated_RDVD_6_5_0_1.iso from 2018.
I do not think "check virtualDisk" command works for think VDs? - it fails also for the healthy thin VD (gaia_parition_p1) - tried also quoted VD name:




SMcli -n aspera_md_1 -c  ' check virtualDisk ["gaia_virtual_p2"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;'
Performing syntax check...

Syntax check complete.

Executing script...

check virtual disk["gaia_virtual_p2"] consistency command has started.
Unable to execute the Check Virtual Disk Consistency command on virtual disk "gaia_virtual_p2" using the command at line 1.
Error 34 - The operation cannot complete because the virtual disk specified in the request is not valid (unknown virtual disk reference). The virtual disk may have been deleted or modified by a user on another management station accessing this storage array.

The command at line 1 that caused the error is:

check virtualDisk ["gaia_virtual_p2"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;

Script execution halted due to error.

SMcli failed.

/opt/dell/mdstoragesoftware/mdstoragemanager/client/SMcli -n aspera_md_1 -c  ' check virtualDisk ["gaia_virtual_p1"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;'
Performing syntax check...

Syntax check complete.

Executing script...

check virtual disk["gaia_virtual_p1"] consistency command has started.
Unable to execute the Check Virtual Disk Consistency command on virtual disk "gaia_virtual_p1" using the command at line 1.
Error 34 - The operation cannot complete because the virtual disk specified in the request is not valid (unknown virtual disk reference). The virtual disk may have been deleted or modified by a user on another management station accessing this storage array.

The command at line 1 that caused the error is:

check virtualDisk ["gaia_virtual_p1"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;

Script execution halted due to error.

SMcli failed.

(edited)

1 Rookie

 • 

10 Posts

November 1st, 2023 15:04

(sorry, I have trouble to format here so it does not cut the lines to oblivion - either in code block or preformatted- let me know if the full comand is visible, which should be 

/SMcli -n aspera_md_1 -c  ' check virtualDisk ["gaia_virtual_p2"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;'
Performing syntax check...

Syntax check complete.

Executing script...

check virtual disk["gaia_virtual_p2"] consistency command has started.
Unable to execute the Check Virtual Disk Consistency command on virtual disk "gaia_virtual_p2" using the command at line 1.
Error 34 - The operation cannot complete because the virtual disk specified in the request is not valid (unknown virtual disk reference). The virtual disk may have been deleted or modified by a user on another management station accessing this storage array.

The command at line 1 that caused the error is:

check virtualDisk ["gaia_virtual_p2"] consistency consistencyErrorFile="/tmp/ConsistencyError.txt" mediaErrorFile="/tmp/mediaError.txt" priority=high verbose=true;

Script execution halted due to error.

SMcli failed.

Moderator

 • 

8.5K Posts

01-11-2023 15:33 PM

DPCG Gaia,
 
I would recommend that you call in to support on this issue, as there is no clear hardware error, and it is too great of a risk of data loss.  Support will be able to directly work with you and the system to determine the cause and possible resolution. 
 
 

DELL-Chris H

Social Media and Communities Professional

Dell Technologies | Enterprise Support Services

#IWork4Dell

Did I answer your query? Please click on ‘Mark as Accepted Answer’. ‘Thumbs up’ the posts you like!

1 Rookie

 • 

10 Posts

November 6th, 2023 13:30

@DELL-Chris H 

Thank you for the hint, 

This will be difficult as we do not have support contract anymore. 

I guess we must accept the data loss (well, we lost it already).

Thanks for trying to help anyways.

Regards,

Chris

No Events found!

Top