EasyManua.ls Logo

Cray ClusterStor L300 - Page 88

Cray ClusterStor L300
145 pages
Print Icon
To Next Page IconTo Next Page
To Next Page IconTo Next Page
To Previous Page IconTo Previous Page
To Previous Page IconTo Previous Page
Loading...
2. Determine the file system name. On the primary MGMT node:
[admin@sn11000n000 ~]$ cscli fs_info
Examples show a file system named sn11000:
[admin@sn11000n000 ~]$ cscli fs_info
----------------------------------------------------------------------------------------------------
OST Redundancy style: Declustered Parity (GridRAID)
Disk I/O Integrity guard (ANSI T10-PI) is not supported by hardware
----------------------------------------------------------------------------------------------------
Information about "sn11000" file system:
----------------------------------------------------------------------------------------------------
Node Role Targets Failover partner Devices
----------------------------------------------------------------------------------------------------
sn11000n000 mgmt 1 / 0 sn11000n001
sn11000n001 mgmt 1 / 0 sn11000n000
sn11000n002 mgs 0 / 1 sn11000n003 /dev/md65
sn11000n003 mds 0 / 1 sn11000n002 /dev/md66
sn11000n004 oss
0 / 1 sn11000n005 /dev/md0
sn11000n005 oss 0 / 1 sn11000n004 /dev/md1
3. If the output in the previous command indicates that the Lustre file system is not started (e.g., the Targets
column shows 0 / 1 instead of 1 / 1 for OSS nodes), start the Lustre file system:
[admin@sn11000n000 ~]$ cscli mount -f testfs
mount: MGS is starting...
mount: MGS is started!
mount: testfs is started on sn11000n002,sn11000n003!
mount: testfs is started on sn11000n004,sn11000n005!
mount: All start commands for filesystem testfs were sent.
mount: Use "cscli show_nodes" to see mount status.
[admin@sn11000n000 ~]$ cscli show_nodes
--------------------------------------------------------------------------------
---------
Hostname Role Power State Service State Targets HA Partner HA
Resources
--------------------------------------------------------------------------------
---------
sn11000n000 MGMT On Started 1 / 0 sn11000n001 Local
sn11000n001 MGMT On Started 1 / 0 sn11000n000 Local
sn11000n002 MDS,MGS On Started 1 / 1 sn11000n003 Local
sn11000n003 MDS,MGS On Started 1 / 1 sn11000n002 Local
sn11000n004 OSS On Started 1 / 1 sn11000n005 Local
sn11000n005 OSS On Started 1 / 1 sn11000n004 Local
--------------------------------------------------------------------------------
---------
4. Log in to the secondary MGMT node:
[MGMT0]$ ssh -l admin secondary_MGMT_node
5. Change to root user on the secondary MGMT node:
[MGMT1]$ sudo su -
6. Obtain the Lustre network identifier (NIDs) of each MGS node (typically n002 and n003:
[MGMT1]# ssh n002 lctl list_nids
[MGMT1]# ssh n003 lctl list_nids
Perform the First-Run Configuration
88

Table of Contents

Related product manuals