https://sites.google.com/site/rhdisk0/unix/aix/gpfs
Filesets
gpfs.base - GPFS File Manager
gpfs.msg.en_US - GPFS Server Messages - U.S. English
gpfs.docs.data - GPFS Server Manpages and Documentation
Path
GPFS commands are in a separate directory.
export PATH=$PATH:/usr/lpp/mmfs/bin
Daemons, status
# ps -ef | grep mmfs
# mmlscluster
# mmlsconfig
Log
# tail -f /var/adm/ras/mmfs.log.latest
PV migration
Collect information
Dump/Save current configuration
# gpfs.snap
List active mounts
# mmlsmount all
File system foobar is mounted on 2 nodes.
List storage pools
# mmlsfs all -P
File system attributes for /dev/foobar:
=====================================
flag value            description
---- ---------------- -----------------------------------------------------
 -P  system           Disk storage pools in file system
List disks in each filesystem
# mmlsfs all -d
File system attributes for /dev/foobar:
=====================================
flag value            description
---- ---------------- -----------------------------------------------------
 -d  mycluster00nsd    Disks in file system
List current NSDs (network shared disks)
# mmlsnsd -M
 Disk name    NSD volume ID      Device         Node name                Remarks
---------------------------------------------------------------------------------------
 mycluster00nsd 0AEC13994BFCEEF7   /dev/hdisk7    host1.mydomain.com
 mycluster00nsd 0AEC13994BFCEEF7   /dev/hdisk7    host1.mydomain.com
 mycluster00nsd 0AEC13994BFCEEF7   -              host3.mydomain.com    (not found) directly attached
mmlsnsd: 6027-1370 The following nodes could not be reached:
host3.mydomain.com
List filesystem manager node(s)
# mmlsmgr
file system      manager node
---------------- ------------------
foobar             10.111.11.111 (host2)
Cluster manager node: 10.111.11.111 (host2)
Show the state of GPFS daemons
- on the local node:
# mmgetstate
 Node number  Node name        GPFS state
------------------------------------------
       2      myhost2          active
- on all  cluster members:
# mmgetstate -a
 Node number  Node name        GPFS state
------------------------------------------
       1      myhost1          active
       2      myhost2          active
       3      somexx1          unknown
Add new disk(s)
Configure new device
# lspv
# cfgmgr
Verify new disk
# lspv
# lspath -l hdiskX
# errpt | head
Edit desc file
Disk name: hdiskX
Primary/Backup server: only for NFS
Disk usage: almost always 'dataAndMetadata'
Failure group: can be -1 when there are no other disks for failover, otherwise see 'mmdf $fs' -> failure group
Desired name: arbitrary; 'clusternameXXnsd' is suggested
Storage pool(s): see 'mmlsfs all -P'
The contents of the file is similar to this:
#DiskName:PrimaryServer:BackupServer:DiskUsage:FailureGroup:DesiredName:StoragePool
hdisk7:::dataAndMetadata:-1:foobar00nsd:system
Configure new disk(s) as NSD
# mmcrnsd -F /path/to/hdisk7.desc
Check NSDs
# mmlsnsd; mmlsnsd -F
Add disk to FS using the transformed desc file (mmcrnsd comments out the previous line and inserts a new one)
# mmadddisk foobar -F hdisk7.desc
Remove old disk(s)
(mmrestripefs?)
Task: Migrate data by deleting disks from GPFS
Example: 4x 100GB DS8100 => 1x 400GB XIV, 2x 4Gb adapters, net 250GB data = 27 minutes on an idle system
WARNING: high I/O!
# mmdeldisk foobar "gpfsXnsd;gpfsYnsd..." -r
Reconfigure tiebreakers
WARNING: GPFS must be shut down on all nodes!
# mmshutdown -a
# mmchconfig tiebreakerDisks="foobar00nsd"
# mmstartup -a
Remove NSD and physical LUN
# mmdelnsd -p [NSD volume ID]
Record old LUNs for the storage administrator (should they be removed from the zoning)
# pcmpath query essmap ...
Delete AIX devices
# rmdev -dl hdiskX
Delete GPFS filesystem
Where fs is the GPFS filesystem device, as seen in mmlsfs, for example, fs0:
Check for processes using the filesystem
# fuser -cux /mount/point
Umount filesystem on all cluster nodes
# mmumount fs0 -a
Delete filesystem
WARNING: Data is destroyed permanently!
# mmdelfs fs0
 
0 件のコメント:
コメントを投稿