beautypg.com

33 troubleshooting failed sfs20 arrays, 1 identifying failed sfs20 arrays, 33 troubleshooting failed sfs20 arrays -54 – HP StorageWorks Scalable File Share User Manual

Page 278: 1 identifying failed sfs20 arrays -54, Troubleshooting failed sfs20 arrays (section 9.33)

background image

Troubleshooting

9–54

8.

You can directly test whether an increment is licensed, as follows.

For the

SFSMDSCAP

or

SFSMDSENT

licenses, enter the following command:

# /usr/opt/hpls/bin/hpls_license SFSMDSCAP 1
granted {}

For the

SFSOSTCAP

or

SFSOSTENT

licenses, you can specify the number of license units (terabytes of

storage) as shown in the following examples. These examples show that 100TB is licensed, whereas

101TB is not:

# /usr/opt/hpls/bin/hpls_license SFSOSTCAP 100

granted {}

# usr/opt/hpls/bin/hpls_license SFSOSTCAP 101

exceeded {Size (101TB) exceeds units available for SFSOSTCAP}

9.33 Troubleshooting failed SFS20 arrays

This section is organized as follows:

Identifying failed SFS20 arrays (Section 9.33.1)

Recovering from a temporary SFS20 array failure (Section 9.33.2)

Recovering degraded MDS or OST services (Section 9.33.3)

9.33.1 Identifying failed SFS20 arrays

When an SFS20 array fails, it impacts the system in a number of ways, as follows:

One or both servers attached to the array may hang at the same time (without reporting any events).

The

show array

command may show an

io-hanging

status. Because an array is attached to two

servers, there are actually two status values, one for each server. The following shows an example of

output from the

show array

command:

sfs> show array 2
.

.

.

Connected to HBA/Port Status
-------------- ---------- ----------
south3 scsi-1/2 io-hanging
south4 scsi-1/2 io-hanging

Events such as the following may be reported in the event log:

2004/11/05 09:51:28 storage n south3: P6C8CX7MQPS7UG: array 2: io-hanging (was
online)

You can search for such events by entering the following command:

sfs> show log facility=storage && data contains "io-hanging"

If the array contains the service LUN, one or both of the servers attached to the array may report

events indicating that they cannot access their quorum disk. One server in the pair may attempt to

reset its peer server. You can search for such events using the following command:

sfs> show log facility=local4 && ( server=south3 || server=south4 )

Events such as the following may be reported in the event log:

2004/11/05 09:51:28 storage n south3: P6C8CX7MQPS7UG:
array 2: failed (was online)

You can search for such events using the following command:

sfs> show log facility=storage && data contains "failed"