Isilon OneFS Version 7.0 Administration Guide
Isilon OneFS Version 7.0 Administration Guide
Isilon OneFS Version 7.0 Administration Guide
Version 7.0
Administration Guide
Published November, 2012 Copyright 2001 - 2012 EMC Corporation. All rights reserved. EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. The information in this publication is provided as is. EMC Corporation makes no representations or warranties of any kind with respect to the information in this publication, and specifically disclaims implied warranties of merchantability or fitness for a particular purpose. Use, copying, and distribution of any EMC software described in this publication requires an applicable software license. EMC, EMC, and the EMC logo are registered trademarks or trademarks of EMC Corporation in the United States and other countries. All other trademarks used herein are the property of their respective owners. For the most up-to-date product documentation, go to the Isilon Customer Support Center. EMC Corporation Hopkinton, Massachusetts 01748-9103 1-508-435-1000 In North America 1-866-464-7381 www.EMC.com
CONTENTS
Chapter 1
15
Architecture..................................................................................................16 Isilon Node....................................................................................................16 Internal and external networks......................................................................17 Isilon cluster.................................................................................................17 Cluster administration............................................................................17 Quorum..................................................................................................17 Splitting and merging.............................................................................18 Storage pools.........................................................................................18 IP address pools.....................................................................................19 The OneFS operating system.........................................................................19 Data-access protocols............................................................................19 Identity management and access control................................................20 Structure of the file system............................................................................21 Data layout.............................................................................................21 Writing files............................................................................................21 Reading files...........................................................................................22 Metadata layout.....................................................................................22 Locks and concurrency...........................................................................22 Striping..................................................................................................22 Data protection overview...............................................................................23 N+M data protection...............................................................................24 Data mirroring........................................................................................24 The file system journal............................................................................25 Virtual hot spare.....................................................................................25 Balancing protection with storage space.................................................25 VMware integration.......................................................................................25 The iSCSI option............................................................................................25 Software modules.........................................................................................26
Chapter 2
27
Data access control.......................................................................................29 ACLs.......................................................................................................29 UNIX permissions...................................................................................30 Mixed-permission environments.............................................................30 Roles and privileges......................................................................................31 Built-in roles...........................................................................................31 OneFS privileges.....................................................................................32 Authentication..............................................................................................34 Local provider.........................................................................................34 File provider............................................................................................34 Active Directory.......................................................................................35 LDAP.......................................................................................................35 NIS.........................................................................................................35 Authentication provider features.............................................................36 Identity management....................................................................................36 Identity types..........................................................................................36 Access tokens........................................................................................37 ID mapping.............................................................................................38
OneFS 7.0 Administration Guide
3
CONTENTS
On-disk identity selection.......................................................................39 User mapping across identities...............................................................40 Configuring user mapping.......................................................................40 Well-known security identifiers...............................................................41 Access zones................................................................................................41 Home directories...........................................................................................42 Home directory creation through SMB.....................................................42 Home directory creation through SSH and FTP.........................................42 Home directory creation in mixed environments.....................................43 Default home directory settings in authentication providers....................44 Supported expansion variables..............................................................44 Managing access permissions.......................................................................45 Configure access management settings..................................................45 Modify ACL policy settings......................................................................46 Update cluster permissions....................................................................51 Managing roles.............................................................................................52 View roles...............................................................................................52 Create a custom role...............................................................................53 Modify a role...........................................................................................53 Delete a custom role...............................................................................53 Create a local user.........................................................................................53 Create a local group......................................................................................54 Managing users and groups..........................................................................55 Modify a local user.................................................................................55 Modify a local group...............................................................................55 Delete a local user..................................................................................56 Delete a local group................................................................................56 Creating file providers...................................................................................56 Create a file provider...............................................................................56 Generate a password file........................................................................57 Managing file providers.................................................................................57 Modify a file provider..............................................................................57 Delete a file provider...............................................................................58 Password file format...............................................................................58 Group file format....................................................................................59 Netgroup file format................................................................................59 Create an Active Directory provider................................................................59 Managing Active Directory providers..............................................................60 Modify an Active Directory provider.........................................................60 Delete an Active Directory provider.........................................................60 Configure Kerberos settings....................................................................60 Active Directory provider settings............................................................61 Create an LDAP provider................................................................................62 Managing LDAP providers..............................................................................65 Modify an LDAP provider.........................................................................65 Delete an LDAP provider.........................................................................65 Create a NIS provider.....................................................................................65 Managing NIS providers................................................................................66 Modify a NIS provider.............................................................................66 Delete a NIS provider..............................................................................67 Create an access zone...................................................................................67 Managing access zones................................................................................68 Modify an access zone............................................................................68 Associate an IP address pool with an access zone..................................69 Delete an access zone............................................................................69
CONTENTS
Chapter 3
File sharing
71
NFS...............................................................................................................72 SMB..............................................................................................................72 HTTP..............................................................................................................72 FTP................................................................................................................73 Mixed protocol environments........................................................................73 Write caching with SmartCache.....................................................................73 Write caching for asynchronous writes....................................................74 Write caching for synchronous writes......................................................74 Create an NFS export.....................................................................................75 Create an SMB share.....................................................................................75 Configure NFS file sharing.............................................................................76 Disable NFS file sharing..........................................................................77 NFS service settings................................................................................77 NFS export behavior settings..................................................................78 NFS performance settings.......................................................................78 NFS client compatibility settings.............................................................80 Configure SMB file sharing............................................................................80 File and directory permission settings.....................................................81 Disable SMB file sharing.........................................................................81 Snapshots settings directory..................................................................82 SMB performance settings......................................................................82 SMB security settings.............................................................................83 Configure and enable HTTP file sharing..........................................................83 Configure and enable FTP file sharing............................................................84 Managing NFS exports...................................................................................84 Modify an NFS export..............................................................................85 Delete an NFS export..............................................................................85 View and configure default NFS export settings.......................................85 Managing SMB shares...................................................................................86 Add a user or group to an SMB share......................................................86 Modify an SMB share..............................................................................86 Delete an SMB share..............................................................................87 SMB share settings.................................................................................87 View and modify SMB share settings......................................................88
Chapter 4
Snapshots
89
Data protection with SnapshotIQ...................................................................90 Snapshot disk-space usage..........................................................................90 Snapshot schedules......................................................................................91 Snapshot aliases..........................................................................................91 File and directory restoration.........................................................................91 File clones.....................................................................................................91 File clones considerations......................................................................92 iSCSI LUN clones....................................................................................93 Snapshot locks.............................................................................................93 Snapshot reserve..........................................................................................93 SnapshotIQ license functionality...................................................................93 Creating snapshots with SnapshotIQ.............................................................94 Create a SnapRevert domain...................................................................94 Create a snapshot...................................................................................95 Create a snapshot schedule....................................................................95 Snapshot naming patterns......................................................................96 Managing snapshots ....................................................................................99
OneFS 7.0 Administration Guide
5
CONTENTS
Reducing snapshot disk-space usage.....................................................99 Delete snapshots..................................................................................100 Modify a snapshot................................................................................101 Modify a snapshot alias........................................................................101 View snapshots....................................................................................101 Snapshot information...........................................................................101 Restoring snapshot data.............................................................................102 Revert a snapshot.................................................................................102 Restore a file or directory using Windows Explorer................................102 Restore a file or directory through a UNIX command line.......................103 Clone a file from a snapshot.................................................................103 Managing snapshot schedules....................................................................103 Modify a snapshot schedule.................................................................103 Delete a snapshot schedule..................................................................104 View snapshot schedules.....................................................................104 Managing with snapshot locks....................................................................104 Create a snapshot lock.........................................................................104 Modify a snapshot lock.........................................................................105 Delete a snapshot lock.........................................................................105 Snapshot lock information....................................................................105 Configure SnapshotIQ settings....................................................................106 SnapshotIQ settings.............................................................................106 Set the snapshot reserve.............................................................................107
Chapter 5
109
Replication policies and jobs......................................................................110 Source and target cluster association...................................................111 Full and differential replication.............................................................111 Controlling replication job resource consumption.................................111 Replication reports...............................................................................112 Replication snapshots.................................................................................112 Source cluster snapshots.....................................................................112 Target cluster snapshots.......................................................................113 Data failover and failback with SyncIQ.........................................................113 Data failover.........................................................................................114 Data failback........................................................................................114 Recovery times and objectives for SyncIQ....................................................114 SyncIQ license functionality........................................................................115 Creating replication policies........................................................................115 Excluding directories in replication.......................................................115 Excluding files in replication.................................................................116 File criteria options...............................................................................117 Configure default replication policy settings.........................................118 Create a replication policy....................................................................119 Create a SyncIQ domain........................................................................123 Assess a replication policy....................................................................124 Managing replication to remote clusters......................................................124 Start a replication job...........................................................................124 Pause a replication job.........................................................................124 Resume a replication job......................................................................125 Cancel a replication job........................................................................125 View active replication jobs..................................................................125 View replication performance information............................................125 Replication job information..................................................................125 Initiating data failover and failback with SyncIQ..........................................126
6
CONTENTS
Fail over data to a secondary cluster.....................................................126 Fail over SmartLock directories.............................................................127 Failover revert.......................................................................................127 Fail back data to a primary cluster........................................................128 Prepare SmartLock directories for failback............................................128 Fail back SmartLock directories............................................................129 Managing replication policies.....................................................................130 Modify a replication policy....................................................................130 Delete a replication policy....................................................................130 Enable or disable a replication policy...................................................130 View replication policies.......................................................................131 Replication policy information..............................................................131 Replication policy settings....................................................................131 Managing replication to the local cluster.....................................................133 Cancel replication to the local cluster...................................................133 Break local target association...............................................................133 View replication jobs targeting the local cluster....................................134 Remote replication policy information...................................................134 Managing replication performance rules.....................................................134 Create a network traffic rule..................................................................134 Create a file operations rule..................................................................135 Modify a performance rule....................................................................135 Delete a performance rule.....................................................................135 Enable or disable a performance rule....................................................136 View performance rules........................................................................136 Managing replication reports.......................................................................136 Configure default replication report settings.........................................136 Delete replication reports.....................................................................136 View replication reports........................................................................137 Replication report information..............................................................137 Managing failed replication jobs.................................................................138 Resolve a replication policy..................................................................138 Reset a replication policy......................................................................138 Perform a full or differential replication.................................................139
Chapter 6
141
File striping.................................................................................................142 Data protection levels.................................................................................142 FlexProtect data recovery.............................................................................142 Smartfail...............................................................................................143 Node failures........................................................................................143 Managing protection levels.........................................................................144 Data protection level information................................................................144 Data protection level disk space usage.......................................................145
Chapter 7
NDMP backup
147
NDMP two way backup................................................................................148 NDMP protocol support...............................................................................148 Supported DMAs.........................................................................................148 NDMP hardware support.............................................................................149 NDMP backup limitations............................................................................149 NDMP performance recommendations........................................................149 Excluding files and directories from NDMP backups....................................151 Configuring basic NDMP backup settings....................................................152
OneFS 7.0 Administration Guide
7
CONTENTS
Configure and enable NDMP backup.....................................................152 Disable NDMP backup..........................................................................153 View NDMP backup settings.................................................................153 NDMP backup settings..........................................................................153 Create an NDMP user account.....................................................................153 Managing NDMP user accounts...................................................................154 Modify the password of an NDMP user account.....................................154 Delete an NDMP user account...............................................................154 View NDMP user accounts....................................................................154 Managing NDMP backup devices.................................................................154 Detect NDMP backup devices...............................................................154 Modify an NDMP backup device name..................................................155 Delete a device entry for a disconnected NDMP backup device.............155 View NDMP backup devices..................................................................155 NDMP backup device settings...............................................................156 Managing NDMP backup ports....................................................................156 Modify NDMP backup port settings.......................................................156 Enable or disable an NDMP backup port...............................................156 View NDMP backup ports......................................................................157 NDMP backup port settings..................................................................157 Managing NDMP backup sessions...............................................................157 Terminate an NDMP session.................................................................157 View NDMP sessions............................................................................158 NDMP session information...................................................................158 View NDMP backup logs..............................................................................159 NDMP environment variables......................................................................159
Chapter 8
163
SmartLock operation modes........................................................................164 Enterprise mode...................................................................................164 Compliance mode.................................................................................164 Replication and backup with SmartLock......................................................165 Data replication in compliance mode....................................................165 Data replication and backup in enterprise mode...................................165 SmartLock license functionality...................................................................166 SmartLock best practices and considerations..............................................166 Set the compliance clock............................................................................167 View the compliance clock..........................................................................168 Creating a SmartLock directory....................................................................168 Retention periods.................................................................................168 Autocommit time periods.....................................................................168 Create a SmartLock directory................................................................169 Managing SmartLock directories.................................................................170 Modify a SmartLock directory................................................................170 View SmartLock directory settings........................................................170 SmartLock directory configuration settings...........................................170 Managing files in SmartLock directories......................................................171 Set a retention period through a UNIX command line............................172 Set a retention period through Windows Powershell.............................172 Commit a file to a WORM state through a UNIX command line...............172 Commit a file to a WORM state through Windows Explorer....................173 Override the retention period for all files in a SmartLock directory.........173 Delete a file committed to a WORM state .............................................174 View WORM status of a file...................................................................174
OneFS 7.0 Administration Guide
CONTENTS
Chapter 9
Protection domains
175
Protection domain considerations...............................................................176 Create a protection domain.........................................................................176 Delete a protection domain.........................................................................176 View protection domains.............................................................................177 Protection domain types.............................................................................177
Chapter 10
Cluster administration
179
User interfaces............................................................................................180 Web administration interface................................................................180 Command-line interface .......................................................................180 Node front panel...................................................................................180 OneFS Platform API...............................................................................181 Connecting to the cluster.............................................................................181 Log in to the web administration interface............................................181 Open an SSH connection to a cluster....................................................181 Restart or shut down the cluster...........................................................181 Licensing....................................................................................................182 Activating licenses................................................................................182 Activate a license through the web administration interface.................182 Activate a license through the command-line interface.........................183 View license information......................................................................183 Unconfiguring licenses.........................................................................183 Unconfigure a license...........................................................................184 General cluster settings...............................................................................184 Configuring the cluster date and time...................................................184 Set the cluster date and time................................................................185 Specify an NTP time server....................................................................185 Set the cluster name.............................................................................186 Specify contact information..................................................................186 View SNMP settings..............................................................................186 Configure SMTP email settings..............................................................186 Configuring SupportIQ..........................................................................187 Enable and configure SupportIQ...........................................................187 Disable SupportIQ................................................................................188 Enable or disable access time tracking.................................................188 Specify the cluster join mode................................................................188 Specify the cluster character encoding..................................................188 Cluster statistics.........................................................................................189 Performance monitoring..............................................................................189 Cluster monitoring.......................................................................................189 Monitor the cluster...............................................................................190 View node status..................................................................................191 Events and notifications.......................................................................191 Monitoring cluster hardware........................................................................196 View node hardware status...................................................................196 SNMP monitoring..................................................................................197 Cluster maintenance...................................................................................199 Replacing node components.................................................................199 Managing cluster nodes.......................................................................200 Remote support using SupportIQ.................................................................201 SupportIQ scripts..................................................................................202 Upgrading OneFS........................................................................................203 Cluster join modes......................................................................................204
OneFS 7.0 Administration Guide
9
CONTENTS
Event notification settings...........................................................................204 System job management.............................................................................205 Job engine overview..............................................................................205 Job performance impact........................................................................207 Job impact policies...............................................................................207 Job priorities.........................................................................................208 Managing system jobs..........................................................................208 Monitoring system jobs........................................................................211 Creating impact policies.......................................................................211 Managing impact policies.....................................................................212
Chapter 11
SmartQuotas
215
Quotas overview.........................................................................................216 Quota types..........................................................................................216 Usage accounting and limits.................................................................218 Disk-usage calculations........................................................................219 Quota notifications...............................................................................220 Quota notification rules........................................................................221 Quota reports.......................................................................................221 Creating quotas...........................................................................................222 Create an accounting quota..................................................................222 Create an enforcement quota................................................................223 Managing quotas........................................................................................223 Search for quotas.................................................................................224 Manage quotas.....................................................................................224 Export a quota configuration file...........................................................225 Import a quota configuration file...........................................................225 Managing quota notifications......................................................................226 Configure default quota notification settings........................................226 Configure custom quota notification rules.............................................227 Map an email notification rule for a quota.............................................228 Configure a custom email quota notification template..........................228 Managing quota reports..............................................................................229 Create a quota report schedule.............................................................229 Generate a quota report........................................................................229 Locate a quota report............................................................................230 Basic quota settings....................................................................................230 Advisory limit quota notification rules settings............................................231 Soft limit quota notification rules settings...................................................232 Hard limit quota notification rules settings..................................................233 Limit notification settings............................................................................233 Quota report settings..................................................................................234 Custom email notification template variable descriptions...........................235
Chapter 12
Storage pools
237
Storage pool overview.................................................................................238 Autoprovisioning.........................................................................................238 Virtual hot spare and SmartPools................................................................239 Spillover and SmartPools............................................................................239 Node pools.................................................................................................240 Add or move node pools in a tier..........................................................240 Change the name or protection level of a node pool..............................240 SSD pools...................................................................................................241 File pools with SmartPools..........................................................................241
10
CONTENTS
Tiers............................................................................................................242 Create a tier..........................................................................................242 Rename a tier.......................................................................................242 Delete a tier..........................................................................................243 File pool policies.........................................................................................243 Pool monitoring...........................................................................................243 Monitor node pools and tiers................................................................244 View unhealthy subpools......................................................................244 Creating file pool policies with SmartPools..................................................244 Managing file pool policies.........................................................................245 Configure default file pool policy settings.............................................245 Configure default file pool protection settings.......................................246 Configure default I/O optimization settings..........................................246 Modify a file pool policy........................................................................246 Copy a file pool policy...........................................................................247 Prioritize a file pool policy.....................................................................247 Use a file pool template policy..............................................................247 Delete a file pool policy........................................................................248 SmartPools settings....................................................................................248 Default file pool protection settings.............................................................250 Default file pool I/IO optimization settings..................................................252
Chapter 13
Networking
253
Cluster internal network overview................................................................254 Internal IP address ranges....................................................................254 Cluster internal network failover...........................................................254 External client network overview.................................................................254 External network settings......................................................................255 IP address pools...................................................................................255 Connection balancing with SmartConnect.............................................256 External IP failover................................................................................257 NIC aggregation....................................................................................257 VLANs...................................................................................................258 DNS name resolution............................................................................258 IPv6 support.........................................................................................258 Configuring the internal cluster network......................................................259 Modify the internal IP address range.....................................................259 Modify the internal network netmask....................................................259 Configure and enable an internal failover network................................260 Disable internal network failover..........................................................261 Configuring an external network..................................................................261 Adding a subnet...................................................................................261 Managing external client subnets.........................................................266 Managing IP address pools...................................................................268 Managing IP address pool interface members.......................................271 Configure DNS settings.........................................................................274 Managing external client connections with SmartConnect...........................275 Configure client connection balancing..................................................275 Client connection settings....................................................................276 Managing network interface provisioning rules............................................276 Create a node provisioning rule............................................................277 Modify a node provisioning rule............................................................278 Delete a node provisioning rule............................................................278
11
CONTENTS
Chapter 14
Hadoop
279
Hadoop support overview...........................................................................280 Hadoop cluster integration..........................................................................280 Managing HDFS...........................................................................................280 Configure the HDFS protocol........................................................................280 Create a local user.......................................................................................282 Enable or disable the HDFS service..............................................................282
Chapter 15
Antivirus
283
On-access scanning....................................................................................284 Antivirus policy scanning............................................................................284 Individual file scanning...............................................................................284 Antivirus scan reports.................................................................................285 ICAP servers................................................................................................285 Supported ICAP servers...............................................................................285 Anitvirus threat responses...........................................................................286 Configuring global antivirus settings...........................................................287 Exclude files from antivirus scans.........................................................287 Configure on-access scanning settings.................................................288 Configure antivirus threat response settings.........................................288 Configure antivirus report retention settings.........................................288 Enable or disable antivirus scanning....................................................289 Managing ICAP servers................................................................................289 Add and connect to an ICAP server........................................................289 Test an ICAP server connection.............................................................289 Modify ICAP connection settings...........................................................289 Temporarily disconnect from an ICAP server..........................................289 Reconnect to an ICAP server..................................................................290 Remove an ICAP server.........................................................................290 Create an antivirus policy............................................................................290 Managing antivirus policies.........................................................................291 Modify an antivirus policy.....................................................................291 Delete an antivirus policy.....................................................................291 Enable or disable an antivirus policy.....................................................291 View antivirus policies..........................................................................291 Managing antivirus scans............................................................................291 Scan a file............................................................................................291 Manually run an antivirus policy...........................................................292 Stop a running antivirus scan...............................................................292 Managing antivirus threats..........................................................................292 Manually quarantine a file....................................................................292 Rescan a file.........................................................................................292 Remove a file from quarantine..............................................................292 Manually truncate a file........................................................................292 View threats.........................................................................................293 Antivirus threat information..................................................................293 Managing antivirus reports..........................................................................293 Export an antivirus report......................................................................293 View antivirus reports...........................................................................293 View antivirus events............................................................................294
Chapter 16
iSCSI
295
CONTENTS
Using SmartConnect with iSCSI targets.................................................296 iSNS client service.......................................................................................296 Access control for iSCSI targets...................................................................297 CHAP authentication.............................................................................297 Initiator access control.........................................................................297 iSCSI considerations and limitations...........................................................297 Supported SCSI mode pages.......................................................................298 Supported iSCSI initiators...........................................................................298 Configuring the iSCSI and iSNS services......................................................298 Configure the iSCSI service...................................................................298 Configure the iSNS client service..........................................................299 View iSCSI sessions and throughput.....................................................299 Create an iSCSI target..................................................................................299 Managing iSCSI targets...............................................................................301 Modify iSCSI target settings..................................................................301 Delete an iSCSI target...........................................................................301 View iSCSI target settings.....................................................................301 Configuring iSCSI initiator access control....................................................302 Configure iSCSI initiator access control.................................................302 Control initiator access to a target........................................................303 Modify initiator name...........................................................................303 Remove an initiator from the access list................................................303 Create a CHAP secret............................................................................304 Modify a CHAP secret............................................................................304 Delete a CHAP secret............................................................................304 Enable or disable CHAP authentication.................................................305 Creating iSCSI LUNs....................................................................................305 Create an iSCSI LUN..............................................................................305 Clone an iSCSI LUN...............................................................................307 Managing iSCSI LUNs..................................................................................309 Modify an iSCSI LUN.............................................................................309 Delete an iSCSI LUN..............................................................................309 Migrate an iSCSI LUN to another target.................................................309 Import an iSCSI LUN..............................................................................310 View iSCSI LUN settings........................................................................311
Chapter 17
VMware integration
313
VASA...........................................................................................................314 Isilon VASA alarms...............................................................................314 VASA storage capabilities.....................................................................314 VAAI............................................................................................................315 VAAI support for block storage..............................................................315 VAAI support for NAS............................................................................315 Configuring VASA support...........................................................................315 Enable VASA.........................................................................................315 Download the Isilon vendor provider certificate....................................315 Add the Isilon vendor provider..............................................................316 Disable or re-enable VASA...........................................................................316
Chapter 18
317
Browse the file system................................................................................318 Create a directory........................................................................................318 Modify file and directory properties.............................................................318 View file and directory properties................................................................318
OneFS 7.0 Administration Guide
13
CONTENTS
14
The EMC Isilon scale-out NAS storage platform combines modular hardware with unified software to harness unstructured data. Powered by the distributed OneFS operating system, an EMC Isilon cluster delivers a scalable pool of storage with a global namespace. The platform's unified software provides centralized web-based and command-line administration to manage the following features:
u u u u u u u u u u u u u u u
A symmetrical cluster that runs a distributed file system Scale-out nodes that add capacity and performance Storage options that manage files, block data, and tiering Flexible data protection and high availability Software modules that control costs and optimize resources Architecture..........................................................................................................16 Isilon Node............................................................................................................16 Internal and external networks..............................................................................17 Isilon cluster.........................................................................................................17 The OneFS operating system.................................................................................19 Structure of the file system....................................................................................21 Data protection overview.......................................................................................23 VMware integration...............................................................................................25 The iSCSI option....................................................................................................25 Software modules.................................................................................................26
15
Architecture
OneFS combines the three traditional layers of storage architecturefile system, volume manager, and data protectioninto a scale-out NAS cluster. In contrast to a scale-up approach, EMC Isilon takes a scale-out approach by creating a cluster of nodes that runs a distributed file system. Each node adds resources to the cluster. Because each node contains globally coherent RAM, as a cluster becomes larger, it becomes faster. Meanwhile, the file system expands dynamically and redistributes content, which eliminates the work of partitioning disks and creating volumes. Nodes work as peers to spread data across the cluster. Segmenting and distributing data a process known as stripingnot only protects data, but also enables a user connecting to any node to take advantage of the entire cluster's performance. The use of distributed software to scale data across commodity hardware sets OneFS apart from other storage systems. No master device controls the cluster; no slaves invoke dependencies. Instead, each node helps control data requests, boosts performance, and expands the cluster's capacity.
Isilon Node
As a rack-mountable appliance, a node includes the following components in a 2U or 4U rack-mountable chassis with an LCD front panel: memory, CPUs, RAM, NVRAM, network interfaces, InfiniBand adapters, disk controllers, and storage media. An Isilon cluster comprises three or more nodes, up to 144. When you add a node to a cluster, you increase the cluster's aggregate disk, cache, CPU, RAM, and network capacity. OneFS groups RAM into a single coherent cache so that a data request on a node benefits from data that is cached anywhere. NVRAM is grouped to write data with high throughput and to protect write operations from power failures. As the cluster expands, spindles and CPU combine to increase throughput, capacity, and input-output operations per second (IOPS). EMC Isilon makes several types of nodes, all of which can be added to a cluster to balance capacity and performance with throughput or IOPS: Node S-Series X-Series NL-Series
Use Case IOPS-intensive applications High-concurrency and throughput-driven workflows Near-primary accessibility, with near-tape value
The following EMC Isilon nodes improve performance: Node Performance Accelerator Backup Accelerator
Function Independent scaling for high performance High-speed and scalable backup-and-restore solution
16
Isilon cluster
An Isilon cluster consists of three or more hardware nodes, up to 144. Each node runs the Isilon OneFS operating system, the distributed file-system software that unites the nodes into a cluster. A clusters storage capacity ranges from a minimum of 18 TB to a maximum of 15.5 PB.
Cluster administration
OneFS centralizes cluster management through a web administration interface and a command-line interface. Both interfaces provide methods to activate licenses, check the status of nodes, configure the cluster, upgrade the system, generate alerts, view client connections, track performance, and change various settings. In addition, OneFS simplifies administration by automating maintenance with a job engine. You can schedule jobs that scan for viruses, inspect disks for errors, reclaim disk space, and check the integrity of the file system. The engine manages the jobs to minimize impact on the cluster's performance. With SNMP versions 1, 2c, and 3, you can remotely monitor hardware components, CPU usage, switches, and network interfaces. EMC Isilon supplies management information bases (MIBs) and traps for the OneFS operating system. OneFS also includes a RESTful application programming interfaceknown as the Platform APIto automate access, configuration, and monitoring. For example, you can retrieve performance statistics, provision users, and tap the file system. The Platform API integrates with OneFS role-based access control to increase security. See the Isilon Platform API Reference.
Quorum
An Isilon cluster must have a quorum to work properly. A quorum prevents data conflicts for example, conflicting versions of the same filein case two groups of nodes become unsynchronized. If a cluster loses its quorum for read and write requests, you cannot access the OneFS file system. For a quorum, more than half the nodes must be available over the internal network. A seven-node cluster, for example, requires a four-node quorum. A 10-node cluster requires a six-node quorum. If a node is unreachable over the internal network, OneFS separates the node from the cluster, an action referred to as splitting. When the split node can reconnect with the cluster and resynchronize with the other nodes, the node rejoins the cluster's majority group, an action referred to as merging. Although a cluster can contain only one majority group, nodes that split from the majority side can form multiple groups. A OneFS cluster contains two quorum properties:
Internal and external networks
17
u u
By connecting to a node with SSH and running the sysctl command-line tool as root, you can view the status of both types of quorum. Here is an example for a cluster that has a quorum for both read and write operations, as the command's output indicates with a 1, for true:
sysctl efs.gmp.has_quorum efs.gmp.has_quorum: 1 sysctl efs.gmp.has_super_block_quorum efs.gmp.has_super_block_quorum: 1
The degraded states of nodessuch as smartfail, read-only, offline, and so onaffect quorum in different ways. A node in a smartfail or read-only state affects only write quorum. A node in an offline state, however, affects both read and write quorum. In a cluster, the combination of nodes in different degraded states determines whether read requests, write requests, or both work. A cluster can lose write quorum but keep read quorum. Consider a four-node cluster in which nodes 1 and 2 are working normally. Node 3 is in a read-only state, and node 4 is in a smartfail state. In such a case, read requests to the cluster succeed. Write requests, however, receive an input-output error because the states of nodes 3 and 4 break the write quorum. A cluster can also lose both its read and write quorum. If nodes 3 and 4 in a four-node cluster are in an offline state, both write requests and read requests receive an inputoutput error, and you cannot access the file system. When OneFS can reconnect with the nodes, OneFS merges them back into the cluster. Unlike a RAID system, an Isilon node can rejoin the cluster without being rebuilt and reconfigured.
Storage pools
Storage pools segment nodes and files into logical divisions to simplify the management and storage of data. A storage pool comprises node pools and tiers. Node pools group equivalent nodes to protect data and ensure reliability. Tiers combine node pools to optimize storage by need, such as a frequently used high-speed tier or a rarely accessed archive. The SmartPools module groups nodes and files into pools. By default, the basic unlicensed technology provisions node pools and creates one file pool. When you license the SmartPools module, you receive more features. You can, for example, create multiple file pools and govern them with policies. The policies move files, directories, and file
18
pools among node pools or tiers. You can also define how OneFS handles write operations when a node pool or tier is full. A virtual hot spare, which reserves space to reprotect data if a drive fails, comes with both the licensed and unlicensed technology.
IP address pools
Within a subnet, you can partition a cluster's external network interfaces into pools of IP address ranges. The pools empower you to customize your storage network to serve different groups of users. Although you must initially configure the default external IP subnet in IPv4 format, you can configure additional subnets in IPv4 or IPv6. You can associate IP address pools with a node, a group of nodes, or NIC ports. For example, you can set up one subnet for storage nodes and another subnet for accelerator nodes. Similarly, you can allocate ranges of IP addresses on a subnet to different teams, such as engineering and sales. Such options help you create a storage topology that matches the demands of your network. In addition, network provisioning rules streamline the setup of external connections. After you configure the rules with network settings, you can apply the settings to new nodes. As a standard feature, the OneFS SmartConnect module balances connections among nodes by using a round-robin policy with static IP addresses and one IP address pool for each subnet. The licensed version of SmartConnect adds features, such as defining IP address pools to support multiple DNS zones.
Supports common data-access protocols, such as SMB and NFS. Connects to multiple identity management systems, such as Active Directory and LDAP. Authenticates users and groups. Controls access to directories and files.
u u
Data-access protocols
With the OneFS operating system, you can access data with multiple file-sharing and transfer protocols. As a result, Microsoft Windows, UNIX, Linux, and Mac OS X clients can share the same directories and files. OneFS supports the following protocols. Protocol SMB
Description Server Message Block gives Windows users access to the cluster. OneFS works with SMB 1, SMB 2, and SMB 2.1. With SMB 2.1, OneFS supports client opportunity locks (oplocks) and large (1 MB) MTU sizes. The default file share is /ifs. The Network File System enables UNIX, Linux, and Mac OS X systems to remotely mount any
IP address pools
19
NFS
Protocol -
Description
subdirectory, including subdirectories created by Windows users. OneFS works with versions 2 through 4 of the Network File System protocol (NFSv2, NFSv3, NFSv4). The default export is / ifs. File Transfer Protocol lets systems with an FTP client connect to the cluster to exchange files. The Internet Small Computer System Interface protocol provides access to block storage. The Hadoop Distributed File System protocol makes it possible for a cluster to work with Apache Hadoop, a framework for data-intensive distributed applications. HDFS integration requires a separate license. Hyper Text Transfer protocol gives systems browser-based access to resources. OneFS includes limited support for WebDAV.
HTTP
Microsoft Active Directory (AD) Lightweight Directory Access Protocol (LDAP) Network Information Service (NIS) Local users and local groups A file provider for accounts in /etc/spwd.db and /etc/group files. With the file provider, you can add an authoritative third-party source of user and group information.
You can manage users with different identity management systems; OneFS maps the accounts so that Windows and UNIX identities can coexist. A Windows user account managed in Active Directory, for example, is mapped to a corresponding UNIX account in NIS or LDAP. To control access, an Isilon cluster works with both the access control lists (ACLs) of Windows systems and the POSIX mode bits of UNIX systems. When OneFS must transform a file's permissions from ACLs to mode bits or from mode bits to ACLs, OneFS merges the permissions to maintain consistent security settings. OneFS presents protocol-specific views of permissions so that NFS exports display mode bits and SMB shares show ACLs. You can, however, manage not only mode bits but also ACLs with standard UNIX tools, such as the chmod and chown commands. In addition, ACL policies enable you to configure how OneFS manages permissions for networks that mix Windows and UNIX systems.
20
Access zones OneFS includes an access zones feature. Access zones allow users from
different authentication providers, such as two untrusted Active Directory domains, to access different OneFS resources based on an incoming IP address. An access zone can contain multiple authentication providers and SMB namespaces.
RBAC for administration OneFS includes role-based access control (RBAC) for administration. In place of a root or administrator account, RBAC lets you manage administrative access by role. A role limits privileges to an area of administration. For example, you can create separate administrator roles for security, auditing, storage, and backup.
Data layout
OneFS evenly distributes data among a cluster's nodes with layout algorithms that maximize storage efficiency and performance. The system continuously reallocates data to conserve space. OneFS breaks data down into smaller sections called blocks, and then the system places the blocks in a stripe unit. By referencing either file data or erasure codes, a stripe unit helps safeguard a file from a hardware failure. The size of a stripe unit depends on the file size, the number of nodes, and the protection setting. After OneFS divides the data into stripe units, OneFS allocates, or stripes, the stripe units across nodes in the cluster. When a client connects to a node, the client's read and write operations take place on multiple nodes. For example, when a client connects to a node and requests a file, the node retrieves the data from multiple nodes and rebuilds the file. You can optimize how OneFS lays out data to match your dominant access patternconcurrent, streaming, or random.
Writing files
On a node, the input-output operations of the OneFS software stack split into two functional layers: A top layer, or initiator, and a bottom layer, or participant. In read and write operations, the initiator and the participant play different roles. When a client writes a file to a node, the initiator on the node manages the layout of the file on the cluster. First, the initiator divides the file into blocks of 8 KB each. Second, the initiator places the blocks in one or more stripe units. At 128 KB, a stripe unit consists of 16 blocks. Third, the initiator spreads the stripe units across the cluster until they span a width of the cluster, creating a stripe. The width of the stripe depends on the number of nodes and the protection setting. After dividing a file into stripe units, the initiator writes the data first to non-volatile random-access memory (NVRAM) and then to disk. NVRAM retains the information when the power is off. During the write transaction, NVRAM guards against failed nodes with journaling. If a node fails mid-transaction, the transaction restarts without the failed node. When the
Structure of the file system
21
node returns, it replays the journal from NVRAM to finish the transaction. The node also runs the AutoBalance job to check the file's on-disk striping. Meanwhile, uncommitted writes waiting in the cache are protected with mirroring. As a result, OneFS eliminates multiple points of failure.
Reading files
In a read operation, a node acts as a manager to gather data from the other nodes and present it to the requesting client. Because an Isilon cluster's coherent cache spans all the nodes, OneFS can store different data in each node's RAM. By using the internal InfiniBand network, a node can retrieve file data from another node's cache faster than from its own local disk. If a read operation requests data that is cached on any node, OneFS pulls the cached data to serve it quickly. In addition, for files with an access pattern of concurrent or streaming, OneFS pre-fetches in-demand data into a managing node's local cache to further improve sequential-read performance.
Metadata layout
OneFS protects metadata by spreading it across nodes and drives. Metadatawhich includes information about where a file is stored, how it is protected, and who can access itis stored in inodes and protected with locks in a B+ tree, a standard structure for organizing data blocks in a file system to provide instant lookups. Meanwhile, OneFS replicates a file's metadata at least to the protection level of the file. Working together as peers, all the nodes help manage metadata access and locking. If a node detects an error in metadata, the node looks up the metadata in an alternate location and then corrects the error.
Striping
In a process known as striping, OneFS segments files into units of data and then distributes the units across nodes in a cluster. Striping protects your data and improves cluster performance. To distribute a file, OneFS reduces it to blocks of data, arranges the blocks into stripe units, and then allocates the stripe units to nodes over the internal network.
22
At the same time, OneFS distributes erasure codes that protect the file. The erasure codes encode the file's data in a distributed set of symbols, adding space-efficient redundancy. With only a part of the symbol set, OneFS can recover the original file data. Taken together, the data and its redundancy form a protection group for a region of file data. OneFS places the protection groups on different drives on different nodescreating data stripes. Because OneFS stripes data across nodes that work together as peers, a user connecting to any node can take advantage of the entire cluster's performance. By default, OneFS optimizes striping for concurrent access. If your dominant access pattern is streaming--that is, lower concurrency, higher single-stream workloads, such as with video--you can change how OneFS lays out data to increase sequential-read performance. To better handle streaming access, OneFS stripes data across more drives. Streaming is most effective on clusters or subpools serving large files.
Description OneFS can send files to servers running the Internet Content Adaptation Protocol (ICAP) to scan for viruses and other threats. OneFS enables you to create clones that share blocks with other files to save space. OneFS can back up data to tape and other devices through the Network Data Management Protocol. Although OneFS supports both NDMP 3-way and 2-way backup, 2-way backup requires an Isilon Backup Accelerator node. You can apply protection domains to files and directories to prevent changes.
Protection domains
The following software modules also help protect data, but they require a separate license:
23
Description SyncIQ replicates data on another Isilon cluster and automates failover and failback operations between clusters. If a cluster becomes unusable, you can fail over to another Isilon cluster. You can protect data with a snapshota logical copy of data stored on a cluster. The SmartLock tool prevents users from modifying and deleting files. With a SmartLock license, you can commit files to a write-once, read-many state: The file can never be modified and cannot be deleted until after a set retention period. SmartLock can help you comply with Securities and Exchange Commission Rule 17a-4.
SnapshotIQ SmartLock
Data mirroring
You can protect on-disk data with mirroring, which copies data to multiple locations. OneFS supports two to eight mirrors. You can use mirroring instead of erasure codes, or you can combine erasure codes with mirroring. Mirroring, however, consumes more space than erasure codes. Mirroring data three times, for example, duplicates the data three times, which requires more space than erasure codes. As a result, mirroring suits transactions that require high performance, such as with iSCSI LUNs.
24
You can also mix erasure codes with mirroring. During a write operation, OneFS divides data into redundant protection groups. For files protected by erasure codes, a protection group consists of data blocks and their erasure codes. For mirrored files, a protection group contains all the mirrors of a set of blocks. OneFS can switch the type of protection group as it writes a file to disk. By changing the protection group dynamically, OneFS can continue writing data despite a node failure that prevents the cluster from applying erasure codes. After the node is restored, OneFS automatically converts the mirrored protection groups to erasure codes.
VMware integration
OneFS integrates with several VMware products, including vSphere, vCenter, and ESXi. For example, OneFS works with the VMware vSphere API for Storage Awareness (VASA) so that you can view information about an Isilon cluster in vSphere. OneFS also works with the VMware vSphere API for Array Integration (VAAI) to support the following features for block storage: hardware-assisted locking, full copy, and block zeroing. VAAI for NFS requires an ESXi plug-in. With the Isilon for vCenter plug-in, you can backup and restore virtual machines on an Isilon cluster. With the Isilon Storage Replication Adapter, OneFS integrates with the VMware vCenter Site Recovery Manager to recover virtual machines that are replicated between Isilon clusters.
and LAGG to aggregate bandwidth and to fail over client sessions. The iSCSI module requires a separate license.
Software modules
You can license additional EMC Isilon software modules to manage a cluster by using advanced features.
u
SyncIQ automated failover and failback SyncIQ replicates data on another Isilon
cluster and automates failover and failback between clusters. If a cluster becomes unusable, you can fail over to another Isilon cluster. Failback restores the original source data after the primary cluster becomes available again.
File clones OneFS provides provisioning of full read/write copies of files, LUNs, and other clones. OneFS also provides virtual machine linked cloning through VMware API integration. SnapshotIQ SnapshotIQ protects data with a snapshota logical copy of data stored on
a cluster. A snapshot can be restored to its top-level directory.
SmartPools SmartPools enable you to create multiple file pools governed by file-pool policies. The policies move files and directories among node pools or tiers. You can also define how OneFS handles write operations when a node pool or tier is full. SmartConnect A SmartConnect license adds advanced balancing policies to evenly distribute CPU usage, client connections, or throughput. The licensed mode also lets you define IP address pools to support multiple DNS zones in a subnet. In addition, SmartConnect supports IP failover, also known as NFS failover. InsightIQ The InsightIQ virtual appliance monitors and analyzes the performance of your Isilon cluster to help you optimize storage resources and forecast capacity. Aspera for Isilon Aspera moves large files over long distances fast. Aspera for Isilon is a cluster-aware version of Aspera technology for non-disruptive, wide-area content delivery. iSCSI OneFS supports the Internet Small Computer System Interface (iSCSI) protocol to
provide block storage for Windows, Linux, and VMware clients. The iSCSI module includes parallel LUN allocation and zero-copy support.
HDFS OneFS works with the Hadoop Distributed File System protocol to help clients
running Apache Hadoop, a framework for data-intensive distributed applications, analyze big data.
SmartQuotas The SmartQuotas module tracks disk usage with reports and enforces
storage limits with alerts.
26
OneFS supports several methods for ensuring that your cluster remains secure, through UNIX- and Windows-style data access permissions as well as configuration controls including role-based administration and access zones. OneFS is designed for a mixed environment in which both Windows Access Control Lists (ACLs) and standard UNIX permissions can be configured on the cluster file system. Windows and UNIX permissions cannot coexist on a single file or directory; however, OneFS uses identity mapping to translate between Windows and UNIX permissions as needed. Access zones enable you to partition authentication control configuration based on the IP address that a user connects to on the cluster. OneFS includes a built-in access zone named "system." By default, new authentication providers, SMB shares, and NFS exports are added to the system zone. When a new IP address pool is added to the cluster, you can select a single access zone that will be used when connecting to any IP address in that pool. Roles enable you to assign privileges to member users and groups. By default, only the "root" and "admin" users can log in to the web administration interface through HTTP, or the command-line interface (CLI) through SSH. The root and admin users can then assign other users to built-in or custom roles with login privileges and other privileges that are required to perform administrative functions. It is recommended that you assign users to roles that contain the minimum set of privileges necessary. In most situations, the default permission policy settings, system access zone, and builtin roles are sufficient; however, you can create additional access zones and custom roles and modify permission policies as necessary for your particular environment.
u u u u u u u u u u u u u u u u u u u u
Data access control...............................................................................................29 Roles and privileges..............................................................................................31 Authentication......................................................................................................34 Identity management............................................................................................36 Access zones........................................................................................................41 Home directories...................................................................................................42 Managing access permissions...............................................................................45 Managing roles.....................................................................................................52 Create a local user.................................................................................................53 Create a local group..............................................................................................54 Managing users and groups..................................................................................55 Creating file providers...........................................................................................56 Managing file providers.........................................................................................57 Create an Active Directory provider........................................................................59 Managing Active Directory providers......................................................................60 Create an LDAP provider........................................................................................62 Managing LDAP providers......................................................................................65 Create a NIS provider.............................................................................................65 Managing NIS providers........................................................................................66 Create an access zone...........................................................................................67
Authentication and access control
27
28
ACLs
In Windows environments, file and directory permissions, referred to as access rights, are defined in access control lists (ACLs). Although ACLs are more complex than mode bits, ACLs can express much more granular sets of access rules. OneFS uses the ACL processing rules commonly associated with Windows ACLs. A Windows ACL contains zero or more access control entries (ACEs), each of which represents the security identifier (SID) of a user or a group as a trustee. In OneFS, an ACL can contain ACEs with a UID, GID, or SID as the trustee. Each ACE contains a set of rights that allow or deny access to a file or folder. An ACE can optionally contain an inheritance flag to specify whether the ACE should be inherited by child folders and files. Instead of the standard three permissions available for mode bits, ACLs have 32 bits of fine-grained access rights. Of these, the upper 16 bits are general and apply to all object types. The lower 16 bits vary between files and directories but are defined in a way that allows most applications to use the same bits for files and directories. Rights can be used for granting or denying access for a given trustee. A user's access can be blocked explicitly through a deny ACE. Access can also be blocked implicitly by ensuring that the user does not directly (or indirectly through a group) appear in an ACE that grants the right in question.
29
UNIX permissions
In a UNIX environment, file and directory access is controlled by POSIX mode bits, which grant read, write, or execute permissions to the owning user, the owning group, and everyone else. OneFS supports the standard UNIX tools for changing permissions, chmod and chown. For more information, see the OneFS man pages for the chmod, chown, and ls commands. All files contain 16 permission bits, which provide information about the file or directory type and the permissions. The lower 9 bits are grouped as three 3-bit sets, called triples, which contain the read ), write (w), and execute (x) permissions for each class of users (owner, group, and other). You can set permissions flags to grant permissions to each of these classes. Assuming the user is not root, OneFS uses the class to determine whether to grant or deny access to the file. The classes are not cumulative; the first class matched is used. It is therefore common to grant permissions in decreasing order.
Mixed-permission environments
When a file operation requests an objects authorization data (for example, with the ls l command over NFS or with the Security tab of the Properties dialog box in Windows Explorer over SMB), OneFS attempts to provide that data in the requested format. In an environment that mixes UNIX and Windows systems, some translation may be required when performing create file, set security, get security, or access operations.
ACE corresponds to the file owner, group, or everyone. After all of the ACEs are generated, any that are not needed are removed before the synthetic ACL is returned.
Action Allows a user to perform a specific action on the cluster. For example the
ISI_PRIV_LOGIN_SSH privilege allows a user to log in to the cluster through an SSH client.
OneFS includes a small set of privileges that allow access to Platform API URIs, but do not allow additional configuration through the CLI or web administration interface. For example, the ISI_PRIV_EVENT privilege provides access to the /platform/1/event URI, but does not allow access to the isi events CLI command. By default, API-only privileges are part of the built-in SystemAdmin role but are hidden from the system privilege list that is viewable by running the isi auth privileges.
Built-in roles
Built-in roles include privileges to perform a set of administrative functions. The following table describes each of the built-in roles from most powerful to least powerful. The table includes the privileges and read/write access levels (if applicable) that are assigned to each role. You can assign users and groups to built-in roles as well as to roles that you create. Role SecurityAdmin
Description Administer security configuration on the cluster, including authentication providers, local users and groups, and role membership.
Read/write access N/A N/A N/A Read/write Read/write N/A N/A N/A Read-only
SystemAdmin
Administer all aspects of cluster configuration that are not specifically handled by the SecurityAdmin role.
31
Role -
Description -
Privileges
ISI_PRIV_LICENSE ISI_PRIV_NFS ISI_PRIV_QUOTA ISI_PRIV_SMB ISI_PRIV_SNAPSHOT ISI_PRIV_STATISTICS ISI_PRIV_NS_TRAVERSE ISI_PRIV_NS_IFS_ACCESS
Read/write access Read-only Read/write Read/write Read/write Read/write Read/write N/A N/A N/A N/A N/A Read-only Read-only Read-only Read-only Read-only Read-only
AuditAdmin
OneFS privileges
Privileges in OneFS are assigned through role membership; they cannot be assigned directly to users and groups.
Table 1 Login privileges
OneFS privilege -
User right -
Privilege type
Action Action
ISI_PRIV_LOGIN_CONSOLE Log in from the console. ISI_PRIV_LOGIN_PAPI Log in to the Platform API and the web administration interface. Log in using SSH.
ISI_PRIV_LOGIN_SSH
Action
32
User right Configure external authentication providers. Create new roles and assign privileges.
User right Configure the NFS server. Configure file system quotas. Configure storage pools. Configure the SMB server. Schedule, take, and view snapshots.
Access the /ifs directory Action tree through the namespace REST service.
User right View and modify system events. Activate OneFS software licenses. View file system performance statistics.
OneFS privileges
33
Authentication
OneFS supports a variety of local and remote authentication providers to verify that users attempting to access the cluster are who they claim to be. Anonymous access, which does not require authentication, is supported for protocols that allow it. To use an authentication provider, it must be added to an access zone. By default, when you create an authentication provider it is added to the built-in system zone, which already includes a local provider and a file provider. You can create multiple instances of each provider type, but it is recommended that you only use a single instance of a provider type within an access zone. For more information about creating and managing access zones, see "Access zones." OneFS supports the concurrent use of multiple authentication providers. For example, OneFS is frequently configured to authenticate Windows clients with Active Directory and to authenticate UNIX clients with LDAP. It is important that you understand their interactions before enabling multiple providers on the cluster. Authentication providers support a mix of the following features:
u
Authentication. All authentication providers support plain text authentication; some providers can also be configured to support NTLM or Kerberos authentication. Ability to manage users and groups directly on the cluster. Netgroups. Used primarily by NFS, netgroups configure access to NFS exports. UNIX-centric user and group properties such as login shell, home directory, UID, and GID. Missing information is supplemented by configuration templates or additional authentication providers. Windows-centric user and group properties such as NetBIOS domain and SID. Missing information is supplemented by configuration templates.
u u u
Local provider
The local provider provides authentication and lookup facilities for user accounts that were added by an administrator. Local users do not include system accounts such as root or admin. The local provider also maintains local group membership. Local authentication can be useful when Active Directory, LDAP, or NIS directory services are not used, or when a specific user or application needs to access the cluster. Unlike UNIX groups, local groups can include built-in groups and Active Directory groups as members. Local groups can also include users from other providers. Netgroups are not supported in the local provider. Each access zone in the cluster contains a separate instance of the local provider, which allows each access zone to have its own list of local users that can authenticate to it.
File provider
A file provider enables you to supply an authoritative third-party source of user and group information to the cluster. A third-party source is useful in UNIX environments where passwd, group, and netgroup files are synchronized across multiple UNIX servers. OneFS uses standard BSD /etc/spwd.db and /etc/group database files as the backing store for the file provider. You generate the spwd.db file by running the pwd_mkdb command-line utility. You can script updates to the database files.
34
The built-in system file provider includes services to list, manage, and authenticate against system accounts such as root, admin, and nobody. Modifying the system file provider is not recommended.
Active Directory
The Active Directory directory service is a Microsoft implementation of Lightweight Directory Access Protocol (LDAP), Kerberos, and DNS technologies that can store information about network resources. Active Directory can serve many functions, but the primary reason for joining the cluster to an Active Directory domain is to perform user and group authentication. When the cluster joins an Active Directory domain, a single Active Directory machine account is created. The machine account is used to establish a trust relationship with the domain and to enable the cluster to authenticate and authorize users in the Active Directory forest. By default, the machine account is named the same as the cluster; however, if the cluster name is more than 15 characters long, the name is hashed and displayed after joining the domain. Whenever possible, a single Active Directory instance should be used when all domains have a trust relationship. Multiple instances should be used only to grant access to multiple sets of mutually-untrusted domains.
LDAP
The Lightweight Directory Access Protocol (LDAP) is a networking protocol that enables you to define, query, and modify directory services and resources. OneFS can authenticate users and groups against an LDAP repository in order to grant them access to the cluster. The LDAP service supports the following features:
u u
Users, groups, and netgroups. Configurable LDAP schemas. For example, the ldapsam schema allows NTLM authentication over the SMB protocol for users with Windows-like attributes. Simple bind authentication (with and without SSL). Redundancy and load balancing across servers with identical directory data. Multiple LDAP provider instances for accessing servers with different user data. Encrypted passwords.
u u u u
NIS
The Network Information Service (NIS) provides authentication and identity uniformity across local area networks. OneFS includes a NIS authentication provider that enables you to integrate the cluster with your NIS infrastructure. NIS, designed by Sun Microsystems, can be used to authenticate users and groups when they access the cluster. The NIS provider exposes the passwd, group, and netgroup maps from a NIS server. Hostname lookups are also supported. Multiple servers can be specified for redundancy and load balancing. NIS is different from NIS+, which OneFS does not support.
Active Directory
35
NTLM
Kerberos
x *
x x
Identity management
There are several methods by which a user can be identified. UNIX users are represented by a user or group identifier (UID or GID); Windows users are represented by a security identifier (SID). Names can also be used as identifiers in one of a variety of formats, depending on their source (for example, SMB, NFSv3, NFSv4, or Kerberos). OneFS provides advanced identity management options to equate these different identity types and enable proper access controls.
Identity types
OneFS supports three primary identity types, each of which can be stored directly on the file system: user identifier (UID) and group identifier (GID) for UNIX and security identifier (SID) for Windows. These identity types are used when creating files, checking file ownership or group membership, and performing file access checks. In OneFS, names are classified as a secondary identifier and are used for authentication but never for authorization. UNIX and Windows identifiers are formatted as follows: A UID or GID is a 32-bit number with a maximum value of 4,294,967,295. A SID is a series of authorities and sub-authorities ending with a 32-bit relative identifier (RID). Most SIDs have the form S-1-5-21-A-B-C-<RID>, where A, B, and C are specific to a domain or computer and <RID> denotes the object in the domain. When a name is provided as an identifier, it is converted into the corresponding user or group object and the correct identity type.
u u
UNIX assumes unique case-sensitive namespaces for users and groups. For example, "Name" and "name" represent different objects. Windows provides a single, case-insensitive namespace for all objects and also specifies a prefix to target an Active Directory domain (for example, domain\name).
36
Kerberos and NFSv4 define principals, which require names to be formatted the same way as email addresses (for example, name@domain.com).
Multiple names can reference the same object. For example, given the name "support" and the domain "example.com", support, EXAMPLE\support, and support@example.com are all names for a single object in Active Directory.
Access tokens
Access tokens form the basis of who you are when performing actions on the cluster, and supply the primary owner and group identities to use during file creation. Access tokens are also compared against the ACL or mode bits during authorization checks. An access token includes all UIDs, GIDs, and SIDs for an identity as well as all OneFS privileges. OneFS exclusively uses the information in the token to determine whether a user has access to a resource. It is important that the token contains the correct list of UIDs, GIDs, and SIDs at all times. An access token is created from one of the following sources: Source Username
Authorization method SMB impersonate user Kerberized NFSv3 Kerberized NFSv4 mountd root mapping HTTP FTP
Access tokens
37
The default on-disk identity is calculated using the final token and the global setting. These identities are used for newly created files.
ID mapping
The file access protocols provided by OneFS support a limited number of identity types: UIDs, GIDs, and SIDs. When an identity is requested that does not match the stored type, a mapping is required. Administrators with advanced knowledge of UNIX and Windows identities can modify the default settings that determine how those identities are mapped in the system. Mappings are stored in a cluster-distributed database called the ID mapper. When retrieving a mapping from the database, as input the ID mapper takes a source and target identity type (UID, GID, or SID). If a mapping already exists between the specified source and the requested type, that mapping is returned; otherwise, a new mapping is created. Each mapping is stored as a one-way relationship from source to destination. Two-way mappings are presented as two complementary one-way mappings in the database. There are four types of identity mappings. The mapping type and identity source determine whether these mappings are stored persistently in the ID mapper.
u
External mappings are derived from identity sources outside OneFS. For example, Active Directory (AD) can store a UID or GID along with a SID. When retrieving the SID from AD, the UID/GID is also retrieved and used for mappings on OneFS. By default, mappings derived from AD are not persistently stored in the ID mapper, but mappings from other external identity sources including LDAP and NIS are persistently stored. Algorithmic mappings are created by adding a UID or GID to a well-known base SID, resulting in a temporary UNIX SID. (For more information, see Mapping UNIX IDs to Windows IDs.) Unlike external mappings, algorithmic mappings are not persistently stored in the ID mapper database. Manual mappings are set explicitly by running the isi auth mapping command at the command line. For command syntax and examples, see the OneFS Command Reference. Manual mappings are stored persistently in the ID mapper database. Automatic mappings are generated if no other mapping type can be found. A SID is mapped to a UID or GID out of the default range of 1,000,000-2,000,000. This range is assumed to be otherwise unused, and a check is made only to ensure there is no mapping from the given UID before it is used. After creation, these mappings are stored persistently in the ID mapper database.
4. If a lookup is necessary, follow these steps: a. By default, normalize the user or group name to lowercase.
38
b. Search all authentication providers except Active Directory for a matching user or group object by name. c. If an object is found, use the associated UID or GID to create an external mapping. 5. Allocate an automatic mapping from the configured range.
UIDs are mapped to a SID with a domain of S-1-22-1 and a resource ID (RID) matching the UID. For example, the UNIX SID for UID 600 is S-1-22-1-600. GIDs are mapped to a SID with a domain of S-1-22-2 and a RID matching the GID. For example, the UNIX SID for GID 800 is S-1-22-2-800.
The on-disk selection does not guarantee the preferred identity can always be stored on disk. On new installations, the on-disk identity is set to native, which is optimized for a mixed Windows and UNIX environment. When you upgrade from OneFS 6.0 or earlier, the on-disk identity is set to unix to match the file system behavior of those earlier versions without requiring an upgrade of all your files and directories.
The available on-disk identities and the corresponding actions taken by the system authentication daemon are described below.
u
native: Determine the identity to store on disk by checking the following ID mapping types in order. The first rule that applies is used to set the on-disk identity. 1. Algorithmic mappings: If an incoming SID matches S-1-22-1-<UID> or S-1-22-2-<GID> (also called a "UNIX SID"), convert it back to the corresponding UID or GID and set it as the on-disk identity. 2. External mappings: If an incoming UID or GID is defined in an external provider (AD, LDAP, or NIS), set it as the on-disk identity. 3. Persistent mappings (usually created with the isi auth mapping create command): If an incoming identity has a mapping that is stored persistently in the ID mapper database, store the incoming identity as the on-disk identity unless the mapping is flagged as on-disk (in which case, set the target ID as the on-disk identity). For example, if a mapping of GID:10000 -> S-1-5-32-545 exists and the --on-disk option has been set with the isi auth mapping modify command, a request for the on-disk storage of GID:10000 returns S-1-5-32-545.
On-disk identity selection
39
4. Automatic mappings: If an incoming SID is not mapped to a UID or GID, set the SID as the on-disk identity. If a UNIX identifier is later required (for example, for crossprotocol NFS or local file system access), a mapping to an auto-allocated UNIX identifier is created.
u
unix: Always store incoming UNIX identifiers on disk. For incoming SIDs, search the configured authentication providers by user name. If a match is found, the SID is mapped to either a UID or GID. If the SID does not exist on the cluster (for example, it is local to the client or part of an untrusted AD domain), a UID or GID is allocated from the ID mapper database and stored on disk, and the resulting SID-to-UID or -GID mapping is stored in the ID mapper database. sid: Store incoming SIDs on disk, with the exception of temporary UNIX SIDs, which are always converted back to their corresponding UNIX identifiers before being stored on disk. For incoming UIDs or GIDs, search the configured authentication providers. If a match is found, store the SID on disk; otherwise, store the UNIX identity.
40
SID S-1-1-0
Name Everyone
Description A system-controlled list of all users, including anonymous users and guests. If set on an ACL, this SID grants file or directory access to all users. If assigned to a role, all users are considered members of that role. A placeholder in an inheritable access control entry (ACE) for the identity of the object's creator. This well-known SID is replaced when the ACE is inherited. A placeholder in an inheritable ACE for the identity of the object creator's primary group. This well-known SID is replaced when the ACE is inherited. A group that represents the object's current owner which, when applied to an object through an ACE, instructs the system to ignore the object owner's implied READ_CONTROL and WRITE_DAC permissions. An account for users who do not have individual accounts. This account does not require a password and cannot log in to a shell. By default, the Guest account is mapped to the UNIX 'nobody' account and is disabled. A built-in group whose members can administer the cluster through Microsoft MMC RPC calls. After the initial OneFS installation, this group contains only the Administrator account. The Domain Admins group is added to this group the first time a cluster is joined to an Active Directory domain.
S-1-3-0
Creator Owner
S-1-3-1
Creator Group
S-1-3-4
Owner Rights
S-1-5-21-domain-501
Guest
S-1-5-32-544
Administrators
Access zones
Access zones provide a way to partition cluster configuration into self-contained units, allowing a subset of parameters to be configured as a virtual cluster. OneFS includes a
Well-known security identifiers
41
built-in access zone called "system." By default, all cluster IP addresses connect to the system zone, which contains all configured authentication providers, all available SMB shares, and all available NFS exports. Access zones contain all of the necessary configuration settings to support authentication and identity management services in OneFS. You can create additional access zones and configure each zone with its own set of authentication providers, user mapping rules, and SMB shares. NFS users can only be authenticated against the system zone. Multiple access zones are particularly useful for server consolidation, for example when merging multiple Windows file servers that are potentially joined to different untrusted forests. If you create access zones, it is recommended that you use them for data access only and that you use the system zone strictly for configuration access. To use an access zone, you must configure your network settings to map an IP address pool to the zone.
Home directories
When you create a local user, OneFS automatically creates a home directory for the user. OneFS also supports dynamic home directory creation for users who access the cluster by connecting to an SMB share or by logging in through FTP or SSH. Regardless of the method by which a home directory was created, you can configure access to the home directory through a combination of SMB, SSH, and FTP.
Home Directory Naming Specifies the path to use as a template for naming home
directories. The path must begin with /ifs and may contain variables, such as %U, that are expanded to generate the home directory path for the user.
Create home directories on first login Specifies whether to create a home directory the first time a user logs in, if a home directory does not already exist for the user. UNIX Shell Specifies the path to the user's login shell. This setting applies only to users
who access the file system through SSH.
42
For a user who can access the cluster both through SMB and through an SSH or FTP login, dot files are not provisioned if the user's home directory is created dynamically through an SMB connection; however, they can be manually copied from the skeleton directory of the user's access zone. You can find the location of the skeleton directory by running the isi zone zones view command through the OneFS command-line interface.
43
Description Expands to the user name, for example, user_001. This variable is typically included at the end of the path, for example, /ifs/home/ %U. Expands to the user's domain name, which varies by authentication provider:
l
%D
For Active Directory users, %D expands to the Active Directory NetBIOS name. For local users, %D expands to the cluster name in uppercase characters. For example, given a cluster named cluster1, %D expands to CLUSTER1. For users in the system file provider, %D expands to UNIX_USERS. For users in a file provider other than the system provider, %D expands to FILE_USERS. For LDAP users, %D expands to LDAP_USERS. For NIS users, %D expands to NIS_USERS.
44
Variable %Z
Description Expands to the access zone name, for example, System. If multiple zones are activated, this variable is useful for differentiating users in separate zones. For example, given the path / ifs/home/%Z/%U, a user named "admin7" in the system zone will be mapped to /ifs/ home/System/admin7.
%L %0 %1 %2
Expands to the host name of the cluster, normalized to lowercase. Expands to the first character of the user name. Expands to the second character of the user name. Expands to the third character of the user name.
If the user name includes fewer than three characters the %0, %1, and %2 variables wrap around. For example, given a user named "ab" the %2 variable maps to a; given a user named "a", all three variables map to a.
Send NTLMv2: Configures the type of NTLM response that is sent to an SMB client. Acceptable values are: yes, no (default). On-Disk Identity: Controls the preferred identity to store on-disk. If OneFS is unable to convert an identity to the preferred format, it is stored as-is. This setting does not affect identities that are currently stored on-disk. Select one of the following settings: native: Let OneFS determine the identity to store on-disk. This is the recommended setting. unix: Always store incoming UNIX identifiers (UIDs and GIDs) on-disk.
Managing access permissions
45
sid: Store incoming Windows security identifiers (SIDs) on-disk, unless the SID was generated from a UNIX identifier; in that case, convert it back to the UNIX identifier and store it on-disk.
If you change the on-disk identity selection, permission errors may occur unless you run the Repair Permissions job as described in the final step of this procedure.
l l
Workgroup: Specifies the NetBIOS workgroup. The default value is WORKGROUP. Space Replacement: For clients that have difficulty parsing spaces in user and group names, specifies a substitute character.
3. Click Save. 4. If you changed the on-disk identity, run the Repair Permissions 'Convert Permissions' task to prevent potential permission errors. a. Click Protocols > ACLs > Repair Permissions Job. b. Optional: Modify the Priority and Impact policy settings. c. For the Repair task setting, click to select Convert permissions. d. For the Path to repair setting, type or click Browse to select the path to the directory whose permissions you want to repair. e. For the Target setting, ensure the Use default system type option is selected. f. For the Access Zone setting, click to select the zone that is using the directory specified in the Path to repair setting. g. Click Start.
Balanced
46
Windows only
Causes cluster permissions to operate with Windows semantics, as opposed to UNIX semantics. Enabling this option causes the system to return an error on UNIX chmod requests.
Configure Allows you to configure the individual permissions policy permission settings available under Permission Policies. policies manually 3. If you selected the Configure permission policies manually option, configure the following settings as needed. ACL creation over SMB Specifies whether to allow or deny creation of ACLs over SMB. Select one of the following options.
l
Do not allow the creation of ACLs over Windows File Sharing (SMB): Prevents ACL creation on the cluster. Allow the creation of ACLs over SMB: Allows ACL creation on the cluster.
Inheritable ACLs on the system take precedence over this setting: If inheritable ACLs are set on a folder, any new files and folders created in that folder will inherit the folder's ACL. Disabling this setting does not remove ACLs currently set on files. If you want to clear an existing ACL, run the chmod -b <mode> <file> command to remove the ACL and set the correct permissions. chmod on Controls what happens when a chmod operation is initiated on a file files with with an ACL, either locally or over NFS. This setting controls any existing ACLs elements that set UNIX permissions, including File System Explorer. Enabling this policy setting does not change how chmod operations affect files that do not have ACLs. Select one of the following options.
l
Remove the existing ACL and set UNIX permissions instead: For chmod operations, removes any existing ACL and instead sets the chmod permissions. Select this option only if you do not need permissions to be set from Windows. Remove the existing ACL and create an ACL equivalent to the UNIX permissions: Stores the UNIX permissions in a Windows ACL. Select this option only if you want to remove Windows permissions but do not want files to have synthetic ACLs. Remove the existing ACL and create an ACL equivalent to the UNIX permissions, for all users/groups referenced in old ACL: Stores the UNIX permissions in a Windows ACL. Select this option only if you want to remove Windows permissions but do not want files to have synthetic ACLs. Merge the new permissions with the existing ACL: Causes Windows and UNIX permissions to operate smoothly in a balanced environment by merging permissions that are applied by chmod with existing ACLs. An ACE for each identity (owner, group, and everyone) is either modified or created, but all other ACEs are unmodified. Inheritable ACEs are also left unmodified to enable Windows users to continue to inherit appropriate permissions.
47
However, UNIX users can set specific permissions for each of those three standard identities.
l
Deny permission to modify the ACL: Prevents users from making NFS and local chmod operations. Enable this setting if you do not want to allow permission sets over NFS.
If you try to run the chmod command on the same permissions that are currently set on a file with an ACL, you may cause the operation to silently failThe operation appears to be successful, but if you were to examine the permissions on the cluster, you would notice that the chmod command had no effect. As a workaround, you can run the chmod command away from the current permissions and then perform a second chmod command to revert to the original permissions. For example, if your file shows 755 UNIX permissions and you want to confirm this number, you could run chmod 700
file; chmod 755 file
ACLs created On Windows systems, the access control entries for directories can on define fine-grained rules for inheritance; on UNIX, the mode bits are directories by not inherited. Making ACLs that are created on directories by the UNIX chmod chmod command inheritable is more secure for tightly controlled environments but may deny access to some Windows users who would otherwise expect access. Select one of the following options.
l l
chown on files
Changes a file or folder's owning user or group. Select one of the following options.
l
Modify the owner and/or group permissions: Causes the chown operation to perform as it does in UNIX. Enabling this setting modifies any ACEs in the ACL associated with the old and new owner or group. Do not modify the ACL: Cause the NFS chown operation to function as it does in Windows. When a file owner is changed over Windows, no permissions in the ACL are changed.
Over NFS, the chown operation changes the permissions and the owner or owning group. For example, consider a file owned by user Joe with "rwx------" (700) permissions, signifying "rwx" permissions for the owner, but no permissions for anyone else. If you run the chown command to change ownership of the file to user Bob, the owner permissions are still "rwx" but they now represent the permissions for Bob, rather than for Joe. In fact, Joe will have lost all of his permissions. This setting does not affect UNIX chown operations performed on files with UNIX permissions, and it does not affect Windows chown operations, which do not change any permissions.
48
In UNIX environments, only the file owner or superuser has the right to run a chmod or chown operation on a file. In Windows environments, you can implement this policy setting to give users the right to perform chmod operations, called the "change permissions" right, or the right to perform chown operations, called the "take ownership" right. The "take ownership" right only gives users the ability to take file ownership, not to give ownership away. Select one of the following options.
l
Allow only owners to chmod or chown: Causes chmod and chown access checks to operate with UNIX-like behavior. Allow owner and users with 'take ownership' right to chown, and owner and users with 'change permissions' right to chmod: Causes chmod and chown access checks to operate with Windows-like behavior.
4. In the Advanced Settings section, configure the following settings as needed. Treatment of "rwx" permissions In UNIX environments, "rwx" permissions signify two things: A user or group has read, write, and execute permissions; and a user or group has the maximum possible level of permissions. When you assign UNIX permissions to a file, no ACLs are stored for that file. However, a Windows system processes only ACLs; Windows does not process UNIX permissions. Therefore, when you view a file's permissions on a Windows system, the cluster must translate the UNIX permissions into an ACL. This type of ACL is called a synthetic ACL. Synthetic ACLs are not stored anywhere; instead, they are dynamically generated as needed and then they are discarded. If a file has UNIX permissions, you may notice synthetic ACLs when you run the ls file command on the cluster in order to view a files ACLs. When you generate a synthetic ACL, the cluster maps UNIX permissions to Windows rights. Windows supports a more granular permissions model than UNIX does, and it specifies rights that cannot easily be mapped from UNIX permissions. If the cluster maps "rwx" permissions to Windows rights, you must enable one of the following options. The main difference between "rwx" and "Full Control" is the broader set of permissions with "Full Control". Select one of the following options.
l
Retain 'rwx' permissions: Generates an ACE that provides only read, write, and execute permissions. Treat 'rwx' permissions as Full Control: Generates an ACE that provides the maximum Windows permissions for a user or a group by adding the "change permissions" right, the "take ownership" right, and the "delete" right.
Operating systems tend to work with group ownership and permissions in two different ways: BSD inherits the group owner from the file's parent folder; Windows and Linux inherit the group owner from the file creator's primary group. If you enable a setting
49
that causes the group owner to be inherited from the creator's primary group, it can be overridden on a per-folder basis by running the chmod command to set the set-gid bit. This inheritance applies only when the file is created. For more information, see the manual page for the chmod command. Select one of the following options.
l
When an ACL exists, use Linux and Windows semantics, otherwise use BSD semantics: Controls file behavior based on whether the new file inherits ACLs from its parent folder. If it does, the file uses the creator's primary group. If it does not, the file inherits from its parent folder. BSD semantics - Inherit group owner from the parent folder: Causes the group owner to be inherited from the file's parent folder. Linux and Windows semantics - Inherit group owner from the creator's primary group: Causes the group owner to be inherited from the file creator's primary group.
(007)
chmod (007) Specifies whether to remove ACLs when running the chmod on files with command. Select one of the following options. existing ACLs l chmod(007) does not remove existing ACL: Sets 007 UNIX permissions without removing an existing ACL.
l
chmod(007) removes existing ACL and sets 007 UNIX permissions: Removes ACLs from files over UNIX file sharing (NFS) and locally on the cluster through the chmod (007) command. If you enable this setting, be sure to run the chmod command on the file immediately after using chmod (007) to clear an ACL. In most cases, you do not want to leave 007 permissions on the file.
Owner permissions
It is impossible to represent the breadth of a Windows ACL's access rules using a set of UNIX permissions. Therefore, when a UNIX client requests UNIX permissions for a file with an ACL over NFS (an action known as a "stat"), it receives an imperfect approximation of the file's true permissions. By default, executing an ls -lcommand from a UNIX client returns a more open set of permissions than the user expects. This permissiveness compensates for applications that incorrectly inspect the UNIX permissions themselves when determining whether to attempt a file-system operation. The purpose of this policy setting is to ensure that these applications proceed with the operation to allow the file system to properly determine user access through the ACL. Select one of the following options.
l
Approximate owner mode bits using all possible owner ACEs: Makes the owner permissions appear more permissive than the actual permissions on the file. Approximate owner mode bits using only the ACE with the owner ID: Makes the owner permissions appear more accurate, in that you see only the permissions for a particular owner and not the more permissive set. However, this may cause access-denied problems for UNIX clients.
50
group permissions
Approximate group mode bits using all possible group ACEs: Makes the group permissions appear more permissive than the actual permissions on the file. Approximate group mode bits using only the ACE with the group ID: Makes the group permissions appear more accurate, in that you see only the permissions for a particular group and not the more permissive set. However, this may cause access-denied problems for UNIX clients.
No "deny" ACEs
The Windows ACL user interface cannot display an ACL if any "deny" ACEs are out of canonical ACL order. However, in order to correctly represent UNIX permissions, deny ACEs may be required to be out of canonical ACL order. Select one of the following options.
l
Remove deny ACEs from synthetic ACLs: Does not include "deny" ACEs when generating synthetic ACLs. This setting can cause ACLs to be more permissive than the equivalent mode bits. Do not modify synthetic ACLs and mode bit approximations: Specifies to not modify synthetic ACL generation; deny ACEs will be generated when necessary.
This option can lead to permissions being reordered, permanently denying access if a Windows user or an application performs an ACL get, an ACL modification, and an ACL set (known as a "roundtrip") to and from Windows. Access check You can control who can change utimes, which are the access and (utimes) modification times of a file, by selecting one of the following options.
l
Allow only owners to change utimes to client-specific times (POSIX compliant): Allows only owners to change utimes, which complies with the POSIX standardan approach that is probably familiar to administrators of UNIX systems. Allow owners and users with write access to change utimes to clientspecific times: Allows owners as well as users with write access to modify utimesa less restrictive approach that is probably familiar to administrators of Windows systems.
4. For Repair task, click to select one of the following settings: l Convert permissions: For each file and directory within the specified Path to repair setting, converts the owner, group and access control list (ACL) to the target ondisk identity. To prevent permissions issues, this task should be run whenever the on-disk identity has been changed.
l
Clone permissions: Applies the permissions settings for the specified Template Directory as-is to the directory specified in the Path to repair. Inherit permissions: Recursively applies the ACL that is used by the specified Template Directory to each file and subdirectory within the specified Path to repair directory, according to normal inheritance rules.
5. For Path to repair, type the full path beginning at /ifs to the directory whose permissions need repaired, or click Browse to navigate to the directory via File System Explorer. 6. For Template Directory (available with Clone and Inherit tasks only), type the full path beginning at /ifs to the directory whose permissions settings you want to apply, or click Browse to navigate to the directory via File System Explorer. 7. Optional: For Target (available with Convert task only), select the on-disk identity type to convert to: l Use default system type: Uses the system's default identity type. This is the default setting.
l
Use native type: If a user or group does not have an authoritative UNIX identifier (UID or GID), uses the Windows identity type (SID). Use UNIX type: Uses the UNIX identity type. Use SID (Windows) type: Uses the Windows identity type.
l l
8. Optional: For Access Zone (available with Convert task only), click to select the access zone to use for ID mapping.
Managing roles
You can view, add, or remove members of any role. Except for built-in roles, whose privileges you cannot modify, you can add or remove OneFS privileges on a role-by-role basis. Roles take both users and groups as members. If a group is added to a role, all users who are members of that group are assigned the privileges associated with the role. Similarly, members of multiple roles are assigned the combined privileges of each role.
View roles
You can view information about built-in and custom roles. For information about the commands and options used in this procedure, run the isi auth roles --help command. 1. Establish an SSH connection to any node in the cluster. 2. At the command prompt, run one of the following commands. l To view a basic list of all roles on the cluster, run:
isi auth roles list
52
To view detailed information about each role on the cluster, including member and privilege lists, run:
isi auth roles list --verbose
To view detailed information about a single role, run the following command, where <role> is the name of the role:
isi auth roles view <role>
Results After creating a role, you can add privileges and member users and groups by running the isi auth roles modify command. For more information, see "Modify a custom role" or run the isi auth roles modify --help command.
Modify a role
You can modify the description and the user or group membership of any role, including built-in roles. However, you cannot modify the name or privileges that are assigned to built-in roles. 1. Establish an SSH connection to any node in the cluster. 2. At the command prompt, run the following command, where <role> is the role name and <options> are optional parameters:
isi auth roles modify <role> [<options>]
For a complete list of the available options, see the OneFS Command Reference.
3. From the Select a provider list, select the local provider for the zone (for example, LOCAL:System). 4. Click Create a user. 5. In the Username field, type a username for the account. 6. In the Password field, type a password for the account. 7. Optional: Configure the following additional settings as needed.
l
Allow password to expire: Select this check box to specify that the password is allowed to expire. UID: If this setting is left blank, the system automatically allocates a UID for the account. This is the recommended setting. You cannot assign a UID that is in use by another local user account.
l l l l l
Full Name: Type a full name for the user. Email Address: Type an email address for the account. Primary Group: Click Select group to specify the owner group. Additional Groups: Specify any additional groups to make this user a member of. Home Directory: Type the path to the user's home directory. If you do not specify a path, a directory is automatically created at /ifs/home/<Username>. UNIX Shell: This setting applies only to users who access the file system through SSH. From the list, click the shell that you want. By default, the /bin/zsh shell is selected. Enabled: Select this check box to allow the user to authenticate against the local database for SSH, FTP, HTTP, and Windows file sharing through SMB. This setting is not used for UNIX file sharing through NFS. Account Expires: Optionally select one of the following options: Never expires: Click to specify that this account does not have an expiration date. Account expires on: Click to display the Expiration date field, and then type the date in the format mm/dd/yyyy.
Prompt password change: Select this check box to prompt for a password change the next time the user logs in.
You cannot assign a GID that is in use by another group. It is recommended that you leave this field blank to allow the system to automatically generate the GID. 7. Optional: Follow these steps for each member that you want to add the group: a. For the Members setting, click Add user. The Select a User dialog box appears. b. For the Search for setting, select either Users or Well-known SIDs. c. If you selected Users, specify values for the following fields: Username: Type all or part of a user name, or leave the field blank to return all users. Wildcard characters are accepted. Access Zone: Select the access zone that contains the authentication provider that you want to search. Provider: Select an authentication provider. d. Click Search. e. In the Search Results table, select a user and then click Select. The dialog box closes. 8. Click Create.
7. Click Close.
Users File: The full path to the spwd.db replacement file. Groups File: The full path to the group replacement file. Netgroups File: The full path to the netgroup replacement file.
5. Optional: To enable this provider to authenticate users, select the Authenticate users from this provider check box. 6. Optional: To specify a home directory naming template, in the Home Directory Naming field, type the full directory path that will contain all home directories.
56
7. Optional: To automatically create home directories for users the next time they log in, select the Create home directories on first login check box. 8. Optional: From the UNIX Shell list, select the shell that will be used when users access the file system through SSH. 9. Click Add File Provider.
If you omit the -d option, the file is created in the /etc directory. For full command usage, view the manual ("man") page by running the man pwd_mkdb command. The following command generates an spwd.db file in the /ifs directory from a password file located at /ifs/test.passwd:
pwd_mkdb -d /ifs /ifs/test.passwd
What to do next To use the spwd.db file, when creating or modifying a file provider using the web administration interface, specify its full path in the Users File setting.
57
The fields are defined below in the order in which they appear in the file. UNIX systems often define the passwd format as a subset of these fields, omitting the class, change, and expire fields. To convert a file from passwd to master.passwd format, add :0:0: between the GID field and the Gecos field.
u
Username: The users name. This field is case sensitive. OneFS does not set a limit on the length; however, many applications truncate the name to 16 characters. Password: The users encrypted password. If authentication is not required for the user, an asterisk (*) can be substituted in place of a password. The asterisk character is guaranteed to not match any password. UID: The users primary identifier. This value should be in the range of 0-4294967294. Take care when choosing a UID to ensure that it does not conflict with an existing account. For example, do not choose the reserved value 0 as the UID. There is no guarantee of compatibility if an assigned value conflicts with an existing UID.
GID: The group identifier of the users primary group. All users are a member of at least one group, which is used for access checks and can also be used when creating files. Class: This field is not supported by OneFS and should be blank. Change: Password change time. OneFS does not support changing passwords of users in the file provider. Expiry: The time at which the account expires. OneFS does not support expiration time of users in the file provider. Gecos: This field can store a variety of information. It is usually used to store the users full name. Home: The users home directory. This field should point to a directory on /ifs. Shell: The absolute path to the users shell (/bin/sh, /bin/csh, /bin/tcsh, / bin/bash, /bin/rbash, /bin/zsh, or /sbin/nologin). For example, to deny command-line access to the user, set the shell to /sbin/nologin.
u u
u u
58
The fields are defined below in the order in which they appear in the file.
u
Group name: The groups name. This field is case sensitive. Although OneFS does not set a limit on the length of the group name, many applications truncate the name to 16 characters. Password: This field is not supported by OneFS and should be set as an asterisk (*). GID: The group identifier. This value should be in the range of 0-4294967294. Be careful when choosing a GID to ensure that it does not conflict with an existing group. Group members: A comma-delimited list of user names that make up the groups members.
u u
In this sample file, if you use rootgrp you get all four hosts; if you use othergrp you get only the last two. A new line signifies a new netgroup. For long netgroup entries, you can type a backslash character (\) in the right-most position of a line to indicate line continuation.
6. Optional: In the Organizational Unit field, type the organizational unit (OU) to connect to on the Active Directory server. 7. Optional: In the Machine Account field, type the name of the machine account. Joining the domain will fail if the machine account exists but resides in a different organizational unit than the one you specified. 8. Optional: To enable Active Directory authentication for NFS, select the Enable Secure NFS check box. If you enable this setting, OneFS registers NFS service principal names (SPNs) during the domain join. 9. Optional: Click Advanced Active Directory Settings to configure advanced settings. 10. Click Join.
60
Most settings require modification only if you are using a Kerberos Key Distribution Center (KDC) other than Active Directoryfor example, if you are using an MIT KDC for NFS version 3 or version 4 authentication. 1. Establish an SSH connection to any node in the cluster. 2. Run the isi auth krb5 command with the add, modify, or delete sub-command to specify which entries to modify in the Kerberos configuration file. For usage information, see the OneFS Command Reference. 3. Propagate the changes to the Kerberos configuration file by running the isi krb5 write command.
auth
By default, changes are written to the global Kerberos configuration file, / etc/krb5.conf. To update the Kerberos configuration file for Active Directory, use the --path option to specify the /etc/likewise-krb5-ad.conf file.
Description Specifies whether to support RFC 2307 attributes for domain controllers. RFC 2307 is required for Windows UNIX Integration and Services For UNIX technologies. Enables the lookup of unqualified user names in the primary domain. If this setting is not enabled, the primary domain must be specified for each authentication operation. Ignores all trusted domains. Specifies trusted domains to include if Ignore
Offline Alerts Enhanced Privacy Home Directory Naming Create Home Directory UNIX Shell Lookup User
61
Setting Match Users with Lowercase Auto-assign UIDs Lookup Group Match Groups with Lowercase Auto-assign GIDs Make UID/GID assignments for users and groups in these specific domains
Description Normalizes Active Directory user names to lowercase before lookup. Enables UID allocation for unmapped Active Directory users. Looks up Active Directory groups in all other providers before allocating a GID. Normalizes Active Directory group names to lowercase before lookup. Enables GID allocation for unmapped Active Directory groups. Restricts user and group lookups to the specified domains.
If you do not specify a port, the default port is used (389 for LDAP; 636 for secure LDAP). If non-secure LDAP (ldap://) is specified, the bind password will be transmitted to the server in clear text. If the Load balance servers option is not selected, servers will be accessed in the order in which they are listed.
Load balance servers: Select the check box to connect to a random server, or clear the check box to connect according to the order in which the servers are listed in the Servers setting. Base Distinguished Name: Type the distinguished name (DN) of the entry at which to start LDAP searches. Base DNs may include cn (Common Name), l (Locality), dc (Domain Component), ou (Organizational Unit), or other components. For example, dc=emc,dc=com is a base DN for emc.com. Bind to: Type the distinguished name of the entry to use to bind to the LDAP server.
62
Password: Specify the password to use when binding to the LDAP server. Use of this password does not require a secure connection; if the connection is not using TLS the password will be sent in clear text. Distinguished Name: Type the distinguished name of the entry at which to start LDAP searches. Search Scope: Defines the default depth from the base DN to perform LDAP searches. Click to select one of the following values: base: Search only the entry at the base DN. onelevel: Search all entries exactly one level below the base DN. subtree: Search the base DN and all entries below it. children: Search all entries below the base DN, excluding the base DN itself. Search Timeout: Type the number of seconds after which a search will not be retried and will fail. The default value is 100. Distinguished Name: Type the distinguished name of the entry at which to start LDAP searches for users. Search Scope: Defines the depth from the base DN to perform LDAP searches for users. Click to select one of the following values: default: Use the setting defined in the default query settings. base: Search only the entry at the base DN. onelevel: Search all entries exactly one level below the base DN. subtree: Search the base DN and all entries below it. children: Search all entries below the base DN, excluding the base DN itself. Query Filter: Sets the LDAP filter for user objects. Authenticate users from this provider: Select the check box to enable the provider to respond to authentication requests, or clear the check box to prevent responding to authentication requests. Home Directory Naming: Type the full path to the location on /ifs to create home directories. Create home directories on first login: Select the check box to automatically create a home directory when a user logs in, if one does not already exist for the user. UNIX Shell: Click to select a login shell from the list. This setting applies only to users who access the file system through SSH. Distinguished Name: Type the distinguished name of the entry at which to start LDAP searches for groups. Search Scope: Defines the depth from the base DN to perform LDAP searches for groups. Click to select one of the following values: default: Use the setting defined in the default query settings. base: Search only the entry at the base DN.
Create an LDAP provider
63
6. Optional: Click Default Query Settings to configure the following additional settings.
l
7. Optional: Click User Query Settings to configure the following additional settings.
l
l l
8. Optional: Click Group Query Settings to configure the following additional settings.
l
onelevel: Search all entries exactly one level below the base DN. subtree: Search the base DN and all entries below it. children: Search all entries below the base DN, excluding the base DN itself.
l
Query Filter: Sets the LDAP filter for group objects. Distinguished Name: Type the distinguished name of the entry at which to start LDAP searches for netgroups. Search Scope: Defines the depth from the base DN to perform LDAP searches for netgroups. Click to select one of the following values: default: Use the setting defined in the default query settings. base: Search only the entry at the base DN. onelevel: Search all entries exactly one level below the base DN. subtree: Search the base DN and all entries below it. children: Search all entries below the base DN, excluding the base DN itself. Query Filter: Sets the LDAP filter for netgroup objects. Name Attribute: Specifies the LDAP attribute that contains UIDs, which are used as login names. The default value is uid. Common Name Attribute: Specifies the LDAP attribute that contains common names. The default value is cn. Email Attribute: Specifies the LDAP attribute that contains email addresses. The default value is email. GECOS Field Attribute: Specifies the LDAP attribute that contains GECOS fields. The default value is gecos. UID Attribute: Specifies the LDAP attribute that contains UID numbers. The default value is uidNumber. GID Attribute: Specifies the LDAP attribute that contains GIDs. The default value is gidNumber. Home Directory Attribute: Specifies the LDAP attribute that contains home directories. The default value is homeDirectory. UNIX Shell Attribute: Specifies the LDAP attribute that contains UNIX login shells. The default value is loginShell. Netgroup Members Attribute: Specifies the LDAP attribute that contains netgroup members. The default value is memberNisNetgroup. Netgroup Triple Attribute: Specifies the LDAP attribute that contains netgroup triples. The default value is nisNetgroupTriple. Group Members Attribute: Specifies the LDAP attribute that contains group members. The default value is memberUid. Unique Group Members Attribute: Specifies the LDAP attribute that contains unique group members. This determines what groups a user is a member of if the LDAP server is queried by the users DN instead of the users name. This setting has no default value.
9. Optional: Click Netgroup Query Settings to configure the following additional settings.
l
10. Optional: Click Advanced LDAP Settings to configure the following additional settings.
l
64
UNIX Password Attribute: Specifies the LDAP attribute that contains UNIX passwords. This setting has no default value. Windows Password Attribute: Specifies the LDAP attribute that contains Windows passwords. The default value is ntpasswdhash. Certificate Authority File: Specifies the full path to the root certificates file. Require secure connection for passwords: Specifies whether to require a TLS connection. Ignore TLS Errors: Continues over a secure connection even if identity checks fail.
l l
65
If the Load balance servers option is not selected, servers will be accessed in the order in which they are listed. 5. Optional: For the Load balance servers setting, click the check box to connect to a random server, or clear the check box to connect according to the order in which the servers are listed in the Servers setting. 6. Optional: Click Default Query Settings to configure the following additional settings.
l l
NIS Domain: Type the NIS domain name. Search Timeout: Type the number of seconds after which a search will not be retried and will fail. The default value is 100. Retry Frequency: Type the number of seconds after which a search will be retried. The default value is 5. Authenticate users from this provider: Select the check box to enable this provider to respond to authentication requests, or clear the check box to prevent responding to authentication requests. Home Directory Naming: Type the full path to the location on /ifs to create home directories. Create home directories on first login: Select the check box to automatically create a home directory when a user logs in, if one does not already exist for the user. UNIX Shell: Click to select a shell from the list. This setting applies only to users who access the file system through SSH. Resolve Hosts: Select the check box to resolve hosts, or clear the check box to specify not to resolve hosts.
7. Optional: Click User Query Settings to configure the following additional settings.
l
8. Optional: Click Host Name Query Settings to configure the following additional settings.
l
66
Manually select authentication providers: Allows you to select one or more provider instances to add to the access zone. Follow these steps for each provider instance that you want to add: a. Click Add an authentication provider. b. In the Authentication Provider Type list, select a provider type. A provider type is listed only if an instance of that type exists and is not already in use by the access zone. c. In the Authentication Provider list, select an available provider instance. d. If you are finished adding provider instances, you can change the priority in which they are called by changing the order in which they are listed. To do so, click the title bar of a provider instance and drag it up or down to a new position in the list.
5. Optional: In the User Mapping Rules list, follow these steps for each user mapping rule that you want to add: a. Click Create a user mapping rule. The User Mapping Rules table appears and displays the Create a User Mapping Rule form. b. In the Operation list, click to select one of the following operations: Append fields from a user: Modifies a token by adding specified fields to it. All appended identifiers become members of the additional groups list. Insert fields from a user: Modifies a token by adding specified fields from another token. An inserted primary user or group becomes the new primary user or group in the token and moves the old primary user or group to the additional identifiers list. Modifying the primary user leaves the tokens username unchanged. When inserting additional groups from a token, the new groups are added to the existing groups. Replace a user with a new user: Replaces a token with the token identified by another user. If another user is not specified, the token is removed from the
Delete a NIS provider
67
list and no user is inserted to replace it. If there are no tokens in the list, access is denied with a "no such user" error. Remove supplemental groups from a user: Modifies a token by removing the supplemental groups. Join two users together: Inserts the new token into the list of tokens. If the new token is the second user, it is inserted after the existing token; otherwise, it is inserted before the existing token. The insertion point is primarily relevant when the existing token is already the first in the list because the first token is used to determine the ownership of new system objects. c. Fill in the fields as needed. Available fields differ depending on the selected operation. d. Click Add Rule. e. If you are finished adding user mapping rules, you can change the priority in which they are called by changing the order in which they are listed. To do so, click the title bar of a rule and drag it up or down to a new position in the list. To ensure that each rule gets processed, it is recommended that you list replacements first and allow/deny rules last. 6. Optional: For the SMB Shares setting, select one of the following options: l Use no SMB shares: Ignores all SMB shares.
l l
Use all SMB shares: Adds each available SMB share to the access zone. Manually select SMB shares: Allows you to select the SMB shares to add to the access zone. The following additional steps are required: a. Click Add SMB shares. The Select SMB Shares dialog box appears. b. Select the check box for each SMB share that you want to add to the access zone. c. Click Select.
7. Click Create Access Zone. What to do next Before you can use an access zone, you must associate it with an IP address pool. See "Associate an IP address pool with an access zone."
68
69
70
Multi-protocol support is built into the OneFS operating system, enabling a single file or directory to be accessed through SMB for Windows file sharing, NFS for UNIX file sharing, secure shell (SSH), FTP, and HTTP. By default, only the SMB and NFS protocols are enabled. OneFS creates the /ifs directory, which is the root directory for all file system data on the cluster. The /ifs directory is configured as an SMB share and an NFS export by default. You can create additional shares and exports within the /ifs directory tree. You can set Windows- and UNIX-based permissions on OneFS files and directories. Users who have the required permissions and administrative privileges can create, modify, and read data on the cluster through one or more of the supported file sharing protocols.
u
SMB. Allows Microsoft Windows and Mac OS X clients to access files that are stored on the cluster. NFS. Allows UNIX, Linux, Mac OS X, Solaris, and other UNIX-based clients to access files that are stored on the cluster. HTTP (with optional DAV). Allows clients to access files that are stored on the cluster through a web browser. FTP. Allows any client that is equipped with an FTP client program to access files that are stored on the cluster through the FTP protocol. NFS.......................................................................................................................72 SMB......................................................................................................................72 HTTP......................................................................................................................72 FTP........................................................................................................................73 Mixed protocol environments................................................................................73 Write caching with SmartCache.............................................................................73 Create an NFS export.............................................................................................75 Create an SMB share.............................................................................................75 Configure NFS file sharing.....................................................................................76 Configure SMB file sharing....................................................................................80 Configure and enable HTTP file sharing..................................................................83 Configure and enable FTP file sharing....................................................................84 Managing NFS exports...........................................................................................84 Managing SMB shares...........................................................................................86
u u u u u u u u u u u u u u
File sharing
71
File sharing
NFS
NFS exports provide UNIX clients network access to file system resources on the cluster. OneFS includes a configurable NFS service that enables you to create and manage NFS exports. OneFS supports asynchronous and synchronous communication over NFS. The OneFS cluster supports the following authentication providers for NFS file sharing.
u
Network Information Service (NIS). NIS is a client/server directory service protocol for distributing system configuration data, such as user and host names, between computers on a network. Lightweight Directory Access Protocol (LDAP). LDAP is an application protocol for querying and modifying directory services running over TCP/IP.
SMB
SMB shares provide Windows clients network access to file system resources on the cluster. OneFS includes a configurable SMB service that enables you to create and manage SMB shares. You can grant permissions to users and groups to carry out operations such as reading, writing, and setting access permissions on SMB shares. The /ifs directory is configured as an SMB share and enabled by default. OneFS supports the "user" and "anonymous" security modes. If the "user" security mode is enabled, when you connect to a share from an SMB client you must provide a valid user name with proper credentials. The SMB protocol uses security identifiers (SIDs) exclusively for authorization data. All identities are converted to SIDs during retrieval and are converted back to their on-disk representation before storage. When a file or directory is created, OneFS checks the access control list (ACL) of its parent directory. If any inheritable access control entries (ACEs) are found, a new ACL is generated from those ACEs. If no inheritable ACEs are found, a default ACL is created from the combined file and directory create mask / create mode settings. OneFS supports the following SMB clients:
u u u u
SMB 1 in Windows 2000/ Windows XP and later. SMB 1 in Mac OS X 10.5 and later. SMB 2 in Windows Vista/ Windows Server 2008 and later. SMB 2.1 in Windows 7/ Windows Server 2008 R2 and later.
HTTP
OneFS includes a configurable HTTP service, which it uses to request files that are stored on the cluster and to interact with the web administration interface. OneFS supports the Distributed Authoring and Versioning (DAV) service to enable multiple users to manage and modify files. DAV is a set of extensions to HTTP that allows clients to read and write from the cluster through the HTTP protocol. You can enable DAV in the web administration interface. OneFS supports a form of the web-based DAV (WebDAV) protocol that enables users to modify and manage files on remote web servers. OneFS performs distributed authoring, but does not support versioning and does not perform security checks. Each node in the cluster runs an instance of the Apache HTTP Server to provide HTTP access. You can configure the HTTP service to run in different modes.
72
File sharing
FTP
The FTP service is disabled by default. You can enable the FTP service to allow any node in the cluster to respond to FTP requests through a standard user account. When configuring FTP access, make sure that the specified FTP root is the home directory of the user who logs in. For example, the FTP root for local user "jsmith" should be ifs/ home/jsmith. You can enable the transfer of files between remote FTP servers and enable anonymous FTP service on the root by creating a local user named "anonymous" or "ftp".
FTP
73
File sharing
Synchronous The stable field is set to data_sync or file_sync. The write-through flag has been applied.
Asynchronous The stable field is set to unstable. The write-through flag has not been applied.
The write-cache enabled (WCE) The WCE setting is set to true. setting is set to false.
Risk If a node fails, no data will be lost except in the unlikely event that a client of that node also crashes before it can reconnect to the cluster. In that situation, asynchronous writes that have not been committed to disk will be lost. If a node fails, asynchronous writes that have not been committed to disk will be lost.
SMB iSCSI
If a node fails, asynchronous writes that have not been committed can cause inconsistencies in any file system that is laid out on the LUN, rendering the file system unusable. It is recommended that you do not disable write caching, regardless of the protocol that you are writing with. If you are writing to the cluster with asynchronous writes, and you decide that the risks of data loss are too great, it is recommended that you configure your clients to use synchronous writes, rather than disable write caching.
File sharing
Enable mount access to subdirectories. Allow subdirectories below the path(s) to be mounted.
7. Specify User/Group mapping. If you select the Custom Default option, you can limit access by mapping root users or all users to a specific user and/or group ID. For root squash, map the root user (UID 0) to the user name "nobody". 8. Select the security type for the export: l UNIX system.
l l l
75
File sharing
Variable %D %U %Z %L %0 %1 %2
Expansion NetBIOS domain name. User name, for example user_001. Zone name, for example System. Host name of the cluster, normalized to lowercase. First character of the user name. Second character of the user name. Third character of the user name.
For example, if a user in the domain named DOMAIN and with the username of user_1, the path /ifs/home/%D/%U is interpreted as /ifs/home/DOMAIN/user_1. 6. Apply the initial Directory ACLs settings. These settings can be modified later. l To maintain the existing permissions on the shared directory, click the Do not change existing permissions option.
l
To apply a default ACL to the shared directory, click the Apply Windows default ACLs option. To maintain the existing permissions on the shared directory, click the Do not change existing permissions option.
To apply a default ACL to the shared directory, click the Apply Windows default ACLs option If the Auto-Create Directories setting is selected, an ACL with the equivalent of UNIX 700 mode bit permissions is created for any directory that is automatically created. 7. Optional: If needed, apply the home directory provisioning options. l Select the Allow Variable Expansion option to expand path variables (%U, %L, %D, and %Z) in the share directory path.
l
Select the Auto-Create Directories option to automatically create directories when users access the share for the first time. This option can only be selected if Allow Variable Expansion has been applied.
Allow Variable Expansion is required for paths to be created automatically. If Allow Variable Expansionis applied, but Auto-Create Directories is not, no paths will be automatically created, and any expansion variables in the name will not be expanded. For example: a new share named "home_share" is created with the path / ifs/%U/home. User "user_1" attempts to connect to "home_share", they will not be able to connect to the cluster and will see an error message. 8. If needed, apply the Users & Groups options. 9. If needed, apply advanced SMB share settings. 10. Click Create.
File sharing
the settings for individual NFS exports as you create them, or edit the settings for individual exports as needed. 1. Click Protocols > UNIX Sharing (NFS) > NFS Settings. 2. Enable or disable the NFS service and version support settings: l NFS Service
l l l
3. Select the Lock Protection Level setting. 4. Click the Reload Cached Configuration button. The cached NFS export settings are reloaded to ensure that changes to DNS or NIS are applied. 5. In the Users/Groups Mapping menu, click Custom Default. A box containing the settings for Map to User Credentials and Also map these user groups appears. a. To limit access by mapping root users or all users to a specific user or group, from the Root users list, click Specific username and then type the user names in the text field. A user is any user available in one of the configured authorization providers. b. To map users to groups, select the Also map these users to groups check box, click Specific user group(s), and then type the group names in the text field. 6. Select the security type. The default setting is UNIX. 7. Click Save.
Description Enables or disables the NFS service. This setting is enabled by default. Enables or disables support for NFSv2. This setting is enabled by default. Enables or disables support for NFSv3. This setting is enabled by default.
77
File sharing
Description Enables or disables support for NFSv4. This setting is disabled by default. Determines the number of node failures that can happen before a lock may be lost. The default value is +2
Setting value Permits non-root users to set file times. The default value is Yes. Overrides the general encoding settings the cluster has for the export. The default value is DEFAULT. Looks up incoming user identifiers (UIDs) in the local authentication database. The default value is No. Enables symlink support for the export. The default value is Yes. Sets the server clock granularity. The default value is 1e-9.
Encoding
Symlinks
Time delta
Description The block size reported to NFSv2+ clients. The default value is 8192. If set to yes, allows NFSv3 and NFSv4 COMMIT operations to be asynchronous. The default value is No.
Commit asynchronously
78
File sharing
Description The preferred directory read transfer size reported to NFSv3 and NFSv4 clients. The default value is 131072. The maximum read transfer size reported to NFSv3 and NFSv4 clients. The default value is 1048576. The recommended read transfer size multiple reported to NFSv3 and NFSv4 clients. The default value is 512. The preferred read transfer size reported to NFSv3 and NFSv4 clients. The number of file nodes to be prefetched on readdir. The default value is 10. If set to yes, performs set attribute operations asynchronously. The default value is No. The action to perform for DATASYNC writes. The default value is DATASYNC. The reply to send for DATASYNC writes. The default value is DATASYNC. The action to perform for FILESYNC writes. The default value is FILESYNC. The reply to send for FILESYNC writes.The default value is FILESYNC. The maximum write transfer size reported to NFSv3 and NFSv4 clients. The default value is 1048576. The recommended write transfer size reported to NFSv3 and NFSv4 clients. The default value is 512. The preferred write transfer size reported to NFSv3 and NFSv4 clients. The default value is 524288. The action to perform for UNSTABLE writes. The default value is UNSTABLE. The reply to send for UNSTABLE writes. The default value is UNSTABLE.
Setattr asynchronous
79
File sharing
Setting value Specifies the maximum file size to allow. The default value is 9223372036854776000. Enables readdirplus. The default value is yes. Returns 32-bit file IDs.
The File and Directory Permissions include Create Permissions, Create Mask (Dir), Create Mode (Dir), Create Mask (File), and Create Mode (File).
l
Performance settings.
The Performance settings include Change Notify and Oplocks. The Security settings include Impersonate Guest, Impersonate User, and NTFS ACL.
l
Security Settings.
The Security settings include Impersonate Guest, Impersonate User, and NTFS ACL. 2. Select the setting you want to modify, and then click the drop-down list next to it and select Custom default. a. A message reminding you that uninformed changes to the advanced settings could result in operational failures. Be aware of the potential consequences of changes before committing to save them appears. Click Continue. The setting properties can now be modified. 3. Make your changes to all of the settings you want to modify, and then click Save. Results The global settings for SMB have now been configured.
80
File sharing
Setting value Sets the default source permissions to apply when a file or directory is created. The default value is Default ACL. Specifies UNIX mode bits that are removed when a directory is created, restricting permissions. Mask bits are applied before mode bits are applied. Specifies UNIX mode bits that are added when a directory is created, enabling permissions. Mode bits are applied after mask bits are applied. Specifies UNIX mode bits that are removed when a file is created, restricting permissions. Mask bits are applied before mode bits are applied. Specifies UNIX mode bits that are added when a file is created, enabling permissions. Mode bits are applied after mask bits are applied.
81
File sharing
Setting value Specifies whether to make the .snapshot directory visible at the root of the share. The default value is Yes. Specifies whether to make the .snapshot directory accessible at the root of the share. The default value is Yes. Specifies whether to make the .snapshot directory visible in subdirectories of the share root. The default value is No. Specifies whether to make the .snapshot directory accessible in subdirectories of the share root. The default value is Yes.
Accessible at Root
Visible in Subdirectories
Accessible in Subdirectories
Setting value Configures notification of clients when files or directories change. This helps prevent clients from seeing stale content, but requires server resources. The default value is All. Indicates whether an opportunistic lock (oplock) request is allowed. An oplock allows clients to provide performance improvements by using locally-cached information. The default value is Yes.
Oplocks
82
File sharing
Setting value Determines guest access to a share. The default value is Never. Allows all file access to be performed as a specific user. This must be a fully qualified user name. The default value is No value. Allows ACLs to be stored and edited from SMB clients. The default value is Yes.
Impersonate User
NTFS ACL
Disable HTTP and redirect to the web interface. Allows only administrative access to the web administration interface. This is the default mode. Disable HTTP entirely. Closes the HTTP port used for file access. Users can still access the web administration interface, but they must specify the port number (8080) in the URL in order to do so.
3. In the Document root directory field, type or click Browse to navigate to an existing directory in /ifs, or click File System Explorer to create a new directory and set its permissions. The HTTP server runs as the daemon user and group. To properly enforce access controls, you must grant the daemon user or group read access to all files under the document root, and allow the HTTP server to traverse the document root. 4. In the Server hostname field, type the HTTP server name. The server hostname must be a fully-qualified, SmartConnect zone name and valid DNS name. The name must begin with a letter and contain only letters, numbers, and hyphens (-). 5. In the Administrator email address field, type an email address to display as the primary contact for issues that occur while serving files. 6. From the Active Directory Authentication list, select an authentication setting: l Off.
SMB security settings
83
File sharing
Basic Authentication Only. Enables HTTP basic authentication. User credentials are sent in plain text. Integrated Authentication Only. Enables HTTP authentication via NTLM, Kerberos, or both. Integrated and Basic Authentication. Enables both basic and integrated authentication. Basic Authentication with Access Controls. Enables HTTP authentication via NTLM and Kerberos, and enables the Apache web server to perform access checks. Integrated and Basic Auth with Access Controls. Enables HTTP basic authentication and integrated authentication, and enables access checks via the Apache web server.
7. Click the Enable DAV check box. This allows multiple users to manage and modify files collaboratively across remote web servers. 8. Click the Disable access logging check box. 9. Click Submit.
Anonymous access. This enables users with "anonymous" or "ftp" as the user name to access files and directories. With this setting enabled, authentication is not required. This setting is disabled by default. Local access. The enables local users to access files and directories with their local user name and password. Enabling this setting allows local users to upload files directly through the file system. This setting is enabled by default.
4. Click Submit.
84
File sharing
Changes to the advanced settings affect all current and future NFS exports that use default settings, and may impact the availability of the NFS file sharing service. Do not make changes to these settings unless you have experience working with NFS. It is recommended that you change the default values for individual NFS exports as you create them, or edit the settings of existing exports.
85
File sharing
Well-known SIDs Skip to step 5. 4. From the Access Zone list, select the access zone you want to search. 5. From the Provider list, select the authentication provider you want to search. Only providers that are currently configured and enabled on the cluster are listed. 6. Click Search. The results of the search appear in the Search Results box. 7. In the search results, click the user, group, or SID that you want to add to the SMB share and then click Select. 8. By default, the access rights of the new account are set to "Deny All". To enable a user or group to access the share, follow these additional steps: a. Next to the user or group account you added, click Edit. b. Select the permission level you want to assign to the user or group. The choices are Run as Root or specific permission levels: Full Control, Read-Write, or Read. 9. Click Save.
File sharing
Any changes made to these settings will only affect the settings for this share. If you need to make changes to the global default values, that can be done from the SMB Settings tab. 1. Click Protocols > Windows Sharing (SMB) > SMB Shares. 2. From the list of SMB shares, locate the share you want to modify and then click View details. 3. For each setting that you want to modify, click Edit, make the change, and then click Save. 4. To modify the settings for file and directory permissions, performance, or security, click Advanced SMB Share Settings.
Share Name. Description. Shared Directory. Home Directory Provisioning. Users and Groups.
The advanced settings for individual SMB shares are the same as the advanced settings for all SMB shares. If you need to change the global default values for the advanced settings, click the SMB Settings tab.
Delete an SMB share
87
File sharing
Uninformed changes to the advanced settings could result in operational failures. Be aware of the potential consequences of changes before committing to save them. The advanced settings for an SMB share are the following:
u
The File and Directory Permissions include Create Permissions, Create Mask (Dir), Create Mode (Dir), Create Mask (File), and Create Mode (File).
u
Performance settings.
The Performance settings include Change Notify and Oplocks. The Security settings include Impersonate Guest, Impersonate User, and NTFS ACL.
u
Security Settings.
The Security settings include Impersonate Guest, Impersonate User, and NTFS ACL.
88
CHAPTER 4 Snapshots
A OneFS snapshot is a logical pointer to data stored on a cluster at a specific point in time. A snapshot contains a directory on a cluster, and includes all data stored in the given directory and any subdirectories that the directory contains. If data contained in a snapshot is modified, the snapshot stores a physical copy of the original data, and references the copy. Snapshots are created according to users' specifications, or automatically generated by OneFS to facilitate system operations. In some cases, snapshots generated by OneFS operations are optional and can be disabled. However, some applications are unable to function without generating snapshots. To create and manage snapshots, you must configure a SnapshotIQ license on the cluster. However, some OneFS operations generate snapshots without requiring that the SnapshotIQ license be configured. If an application generates a snapshot, and a SnapshotIQ license is not configured, you can still view the snapshot. However, there are some OneFS operations that generate snapshots for internal system use. Unless a SnapshotIQ license is configured, all snapshots generated by OneFS operations are automatically deleted after they are no longer needed. You can identify and locate snapshots by the snapshot name or ID. A snapshot name is specified by a user and assigned to the subdirectory that contains the snapshot. A snapshot ID is a numerical identifier that is assigned to snapshots by the system.
u u u u u u u u u u u u u u u u
Data protection with SnapshotIQ...........................................................................90 Snapshot disk-space usage...................................................................................90 Snapshot schedules..............................................................................................91 Snapshot aliases...................................................................................................91 File and directory restoration.................................................................................91 File clones.............................................................................................................91 Snapshot locks.....................................................................................................93 Snapshot reserve..................................................................................................93 SnapshotIQ license functionality...........................................................................93 Creating snapshots with SnapshotIQ.....................................................................94 Managing snapshots ............................................................................................99 Restoring snapshot data.....................................................................................102 Managing snapshot schedules............................................................................103 Managing with snapshot locks............................................................................104 Configure SnapshotIQ settings............................................................................106 Set the snapshot reserve.....................................................................................107
Snapshots
89
Snapshots
90
Snapshots
Snapshot schedules
OneFS can automatically generate snapshots intermittently according to a snapshot schedule. With snapshot schedules, you can periodically generate snapshots of a directory, without having to manually create a snapshot every time. You can also assign an expiration period to the snapshots that are generated, causing OneFS to automatically delete each snapshot after the specified period has expired. It is often advantageous to create more than one snapshot per directory, with shorter duration periods assigned to snapshots that are generated more frequently, and longer expiration periods assigned to snapshots that are generated less frequently.
Snapshot aliases
A snapshot alias is an optional, alternative name for a snapshot. If a snapshot is assigned an alias, and that alias is later assigned to another snapshot, the alias is automatically removed from the old snapshot before it is assigned to the new snapshot. Snapshot aliases are most commonly used by snapshot schedules. When specified in a snapshot schedule, OneFS assigns the alias to each snapshot generated by the schedule. The alias is then attached only to the most recent snapshot generated based on the schedule. You can use aliases to quicky identify the most recent snapshot generated according to a schedule. OneFS also uses snapshot aliases internally to identify the most recent snapshot generated by OneFS operations.
File clones
OneFS enables you to create file clones that share blocks with existing files in order to save space on the cluster. Although you can clone files from snapshots, clones are primarily used internally by OneFS. File clones share blocks with existing files in order to save space on the cluster. The shared blocks are contained in a shadow store that is referenced by both the clone and
Snapshot schedules
91
Snapshots
the cloned file. A file clone usually consumes less space and takes less time to create than a file copy. Immediately after a clone is created, all data originally contained in the cloned file is transferred to a shadow store. A shadow store is a hidden file that is used to hold shared data for clones and cloned files. Because both files reference all blocks from the shadow store, the two files consume no more space than the original file; the clone does not take up any additional space on the cluster. However, if the cloned file or clone is modified, the file and clone will share only blocks that are common to both of them, and the modified, unshared blocks will occupy additional space on the cluster. Over time, the shared blocks contained in the shadow store might become useless if neither the file nor clone references the blocks. The blocks that are no longer needed are deleted routinely by the cluster. However, you can cause the cluster to delete unused blocks at any time by running the shadow store delete job.
Reading a cloned file might be slower than reading a copied file. Specifically, reading non-cached data from a cloned file is slower than reading non-cached data from a copied file. Reading cached data from a cloned file takes no more time than reading cached data from a copied file. When a file and its clone are replicated to another Isilon cluster or backed up to an Network Data Management Protocol (NDMP) backup device, the file and clone do not share blocks on the target Isilon cluster or backup device. Shadows stores are not transferred to the target cluster or backup device, so clones and cloned files consume the same amount of space as copies. When a file is cloned, the shadow store referenced by the clone and cloned file is assigned to the storage pool of the cloned file. If you delete the storage pool that the shadow store resides on, the shadow store is moved to a pool occupied either by the original file or a clone of the file. The protection level of a shadow store is at least as high as the most protected file or clone associated referencing the shadow store. For example, if a cloned file resides in a storage pool with +2 protection, and the clone resides in a storage pool with +3 protection, the shadow store is protected at +3. Quotas account for clones and cloned files as if they consumed both shared and unshared data; from the perspective of a quota, a clone and a copy of the same file do not consume different amounts of data. However, if the quota includes data protection overhead, the data protection overhead for the shadow store is not accounted for by the quota. Clones cannot contain alternate data streams. If you clone a file with alternate data streams (ADS), the clone will not contain the alternate data streams.
92
Snapshots
Snapshot locks
A snapshot lock prevents a snapshot from being deleted. If a snapshot has one or more locks applied to it, the snapshot cannot be deleted and is referred to as a locked snapshot. You cannot delete a locked snapshot manually. OneFS is also unable to delete a snapshot lock. If the duration period of a locked snapshot expires, the snapshot will not be deleted by the system until all locks on the snapshot have been deleted. OneFS applies snapshot locks to ensure that snapshots generated by OneFS applications are not deleted prematurely. For this reason, it is recommended that you do not delete snapshot locks or modify the duration period of snapshot locks. A limited number of locks can be applied to a snapshot at a time. If you create too many snapshot locks and the limit is reached, OneFS might be unable to apply a snapshot lock when necessary. For this reason, it is recommended that you do not create snapshot locks.
Snapshot reserve
The snapshot reserve enables you to set aside a minimum portion of the cluster-storage capacity specifically for snapshots. If specified, the percentage of cluster capacity that is reserved for snapshots is not accessible to any other OneFS operation. The snapshot reserve does not limit the amount of space that snapshots are allowed to consume on the cluster. Snapshots can consume more than the percentage of capacity specified by the snapshot reserve. It is recommended that you do not specify a snapshot reserve.
Yes
Configure SnapshotIQ settings Yes View snapshot schedules Delete snapshots Access snapshot data Yes Yes Yes
93
Snapshots
Unconfigured Yes
Yes
If you unconfigure a SnapshotIQ license, you will not be able to create new snapshots, all snapshot schedules will be disabled, and you will not be able to modify snapshots or snapshot settings. However, you will still be able to delete snapshots and access data contained in snapshots.
94
Snapshots
Create a snapshot
You can create a snapshot of a directory. 1. Click Data Protection > SnapshotIQ > Summary. 2. Click Capture a new snapshot. 3. Optional: To modify the default name of a snapshot, in the Capture a Snapshot area, in the Snapshot Name field, type a name 4. In the Directory Path field, specify the directory that you want the snapshot to contain. 5. Optional: To create an alternative name for the snapshot, specify a snapshot alias. a. Next to Create an Alias, click Yes. b. To modify the default snapshot alias name, in the Alias Name field, type an alternative name for the snapshot. 6. Optional: To assign a time that OneFS will automatically delete the snapshot, specify an expiration period. a. Next to Snapshot Expiration, click Snapshot Expires on. b. In the calendar, specify the day that you want the snapshot to be automatically deleted. 7. Click Capture.
5. In the Directory Path field, specify the directory that you want to be contained in snapshots that are generated according to this schedule. 6. Specify how often you want to generate snapshots according to the schedule. Generate snapshots every day, or skip generating snapshots for a specified number of days. From the Snapshot Frequency list, select Daily, and specify how often you want to generate snapshots.
Generate snapshots on specific days of From the Snapshot Frequency list, select the week, and optionally skip Weekly and specify how often you want to generating snapshots for a specified generate snapshots. number of weeks.
Create a snapshot
95
Snapshots
Generate snapshots on specific days of From the Snapshot Frequency list, select the month, and optionally skip Monthly and specify how often you want to generating snapshots for a specified generate snapshots. number of months. Generate snapshots on specific days of From the Snapshot Frequency list, select the year. Yearly and specify how often you want to generate snapshots. A snapshot schedule cannot span multiple days. For example, you cannot specify to begin generating snapshots at 5:00 PM Monday and end at 5:00 AM Tuesday. To continuously generate snapshots for a period greater than a day, you must create two snapshot schedules. For example, to generate snapshots from 5:00 PM Monday to 5:00 AM Tuesday, create one schedule that generates snapshots from 5:00 PM to 11:59 PM on Monday, and another schedule that generates snapshots from 12:00 AM to 5:00 AM on Tuesday. 7. Optional: To assign an alternative name to the most recent snapshot generated by the schedule, specify a snapshot alias. a. Next to Create an Alias, click Yes. b. To modify the default snapshot alias name, in the Alias Name field, type an alternative name for this snapshot. 8. Optional: To specify a length of time that snapshots generated according to the schedule exist on the cluster before they are automatically deleted by OneFS, specify an expiration period. a. Next to Snapshot Expiration, click Snapshots expire. b. Next to Snapshots expire, specify how long you want to retain the snapshots generated according to this schedule. 9. Click Create.
Description The day of the week. The abbreviated day of the week. For example, if the snapshot is generated on a Sunday, %a is replaced with "Sun". The name of the month. The abbreviated name of the month. For example, if the snapshot is generated in September, %b is replaced with "Sep".
%B %b
96
Snapshots
Variable %C
Description The first two digits of the year. For example, if the snapshot is created in 2012, %C is replaced with "20". The time and day. This variable is equivalent to specifying "%a %b %e %T %Y". The two digit day of the month. The day of the month. A single-digit day is preceded by a blank space. The date. This variable is equivalent to specifying "%Y-%m-%d" The year. This variable is equivalent to specifying "%Y". However, if the snapshot is created in a week that has less than four days in the current year, the year that contains the majority of the days of the week is displayed. The first day of the week is calculated as Monday. For example, if a snapshot is created on Sunday, January 1, 2017, %G is replaced with "2016", because only one day of that week is in 2017.
%c
%d %e %F
%G
%g
The abbreviated year. This variable is equivalent to specifying "%y". However, if the snapshot was created in a week that has less than four days in the current year, the year that contains the majority of the days of the week is displayed. The first day of the week is calculated as Monday. For example, if a snapshot is created on Sunday, January 1, 2017, %g is replaced with "16", because only one day of that week is in 2017.
%H
The hour. The hour is represented on the 24hour clock. Single-digit hours are preceded by a zero. For example, if a snapshot is created at 1:45 AM, %H is replaced with "01". The abbreviated name of the month. This variable is equivalent to specifying "%b". The hour represented on the 12-hour clock. Single-digit hours are preceded by a zero. For example, if a snapshot is created at 1:45 AM, %I is replaced with "01". The numeric day of the year. For example, if a snapshot is created on February 1, %j is replaced with "32".
%h
%I
%j
97
Snapshots
Variable %k
Description The hour represented on the 24-hour clock. Single-digit hours are preceded by a blank space. The hour represented on the 12-hour clock. Single-digit hours are preceded by a blank space. For example, if a snapshot is created at 1:45 AM, %I is replaced with "1". The two-digit minute. The two-digit month.
%l
%M %m %p %{PolicyName}
AM or PM.
The name of the replication policy that the snapshot was created for. This variable is valid only if you are specifying a snapshot naming pattern for a replication policy. The time. This variable is equivalent to specifying "%H:%M". The time. This variable is equivalent to specifying "%I:%M:%S %p". The two-digit second. The second represented in UNIX or POSIX time. The name of the source cluster of the replication policy that the snapshot was created for. This variable is valid only if you are specifying a snapshot naming pattern for a replication policy. The time. This variable is equivalent to specifying "%H:%M:%S" The two-digit numerical week of the year. Numbers range from 00 to 53. The first day of the week is calculated as Sunday.
%R
%r
%S %s %{SrcCluster}
%T
%U
%u
The numerical day of the week. Numbers range from 1 to 7. The first day of the week is calculated as Monday. For example, if a snapshot is created on Sunday, %u is replaced with "7".
%V
The two-digit numerical week of the year that the snapshot was created in. Numbers range from 01 to 53. The first day of the week is calculated as Monday. If the week of January 1 is four or more days in length, then that week is counted as the first week of the year.
98
Snapshots
Variable %v
Description The day that the snapshot was created. This variable is equivalent to specifying "%e-%b-
%Y".
%W The two-digit numerical week of the year that the snapshot was created in. Numbers range from 00 to 53. The first day of the week is calculated as Monday. %w The numerical day of the week that the snapshot was created on. Numbers range from 0 to 6. The first day of the week is calculated as Sunday. For example, if the snapshot was created on Sunday, %w is replaced with "0". %X The time that the snapshot was created. This variable is equivalent to specifying "%H:%M:
%S".
%Y %y The year that the snapshot was created in. The last two digits of the year that the snapshot was created in. For example, if the snapshot was created in 2012, %y is replaced with "12". The time zone that the snapshot was created in. The offset from coordinated universal time (UTC) of the time zone that the snapshot was created in. If preceded by a plus sign, the time zone is east of UTC. If preceded by a minus sign, the time zone is west of UTC. The time and date that the snapshot was created. This variable is equivalent to specifying "%a %b %e %X %Z %Y". Escapes a percent sign. "100%%" is replaced with "100%".
%Z %z
%+
%%
Managing snapshots
You can delete and view snapshots. You can also modify attributes of snapshots. You can modify the name, duration period, and alias of an existing snapshot. However, you cannot modify the data contained in a snapshot; the data contained in a snapshot is read-only.
Managing snapshots
99
Snapshots
The size of a snapshot is the maximum amount of data that might be freed if the snapshot is deleted. Be aware of the following considerations when attempting to reduce the capacity used by snapshots:
u
Deleting a snapshot frees only the space that is taken up exclusively by that snapshot. If two snapshots reference the same stored data, that data is not freed until both snapshots are deleted. Remember that snapshots store data contained in all subdirectories of the root directory; if snapshot_one contains /ifs/data/, and snapshot_two contains /ifs/data/dir, the two snapshots most likely share data. If you delete a directory, and then re-create it, a snapshot containing the directory stores the entire re-created directory, even if the files in that directory are never modified. Deleting multiple snapshots that contain the same directories is more likely to free data than deleting multiple snapshots that contain different directories. If multiple snapshots of the same directories exist, deleting the older snapshots is more likely to free disk-space than deleting newer snapshots. Snapshots store only data that cannot be found on the file system or another snapshot. If you delete the oldest snapshot of a directory, the amount of space known as the size of the snapshot will be freed. Snapshots that are assigned expiration dates are automatically marked for deletion by the snapshot daemon. If the daemon is disabled, snapshots will not be automatically deleted by the system.
Delete snapshots
You can delete a snapshot if you no longer want to access the data contained in the snapshot. Disk space occupied by deleted snapshots is freed when the snapshot delete job is run. Also, if you delete a snapshot that contains clones or cloned files, data in a shadow store might no longer be referenced by files on the cluster; unreferenced data in a shadow store is deleted when the shadow store delete job is run. OneFS routinely runs both the shadow store delete and snapshot delete jobs. However, you can also manually run the jobs at any time. 1. Click Data Protection > SnapshotIQ > Snapshots. 2. Specify the snapshots that you want to delete. a. For each snapshot you want to delete, in the Saved File System Snapshots table, in the row of a snapshot, select the check box. b. From the Select an action list, select Delete. c. In the confirmation dialog box, click Delete. 3. Optional: To increase the speed at which deleted snapshot data is freed on the cluster, run the snapshot delete job. a. Navigate to Cluster Management > Operations. b. In the Running Jobs area, click Start Job. c. From the Job list, select SnapshotDelete. d. Click Start. 4. Optional: To increase the speed at which deleted data shared between cloned files is freed on the cluster, run the shadow store delete job. Run the shadow store delete job only after you run the snapshot delete job.
100
Snapshots
a. Navigate to Cluster Management > Operations. b. In the Running Jobs area, click Start Job. c. From the Job list, select ShadowStoreDelete. d. Click Start.
Modify a snapshot
You can modify the name and expiration date of a snapshot. 1. Click File System Management > SnapshotIQ > Snapshots. 2. In the Saved File System Snapshots table, in the row of the snapshot that you want to modify, click View Details. 3. In the Snapshot Details area, modify snapshot attributes. 4. Next to each snapshot attribute that you modified, click Save.
View snapshots
You can view all snapshots. 1. Click Data Protection > SnapshotIQ > Snapshots. 2. In the Saved File System Snapshots table, view snapshots.
Snapshot information
You can view information about snapshots, including the total amount of space consumed by all snapshots. The following information is displayed in the Saved Snapshots area:
u u
SnapshotIQ Status Indicates whether the SnapshotIQ tool is accessible on the cluster. Total Number of Saved Snapshots Indicates the total number of snapshots that
exist on the cluster.
Modify a snapshot
101
Snapshots
Total Number of Snapshot Aliases Indicates the total number of snapshot aliases
that exist on the cluster.
Capacity Used by Saved Snapshots Indicates the total amount of space consumed
by all snapshots.
Revert a snapshot
You can revert a directory back to the state it was in when a snapshot was taken. Before you begin u Create a SnapRevert domain for the directory.
u
1. Click Cluster Management > Operations > Operations Summary. 2. In the Running Jobs area, click Start job. 3. From the Job list, select SnapRevert. 4. Optional: To specify a priority for the job, from the Priority list, select a priority. Lower values indicate a higher priority. If you do not specify a priority, the job is assigned the default snapshot revert priority. 5. Optional: To specify the amount of cluster resources the job is allowed to consume, from the Impact policy list, select an impact policy. If you do not specify a policy, the job is assigned the default snapshot revert policy. 6. In the Snapshot field, type the name or ID of the snapshot that you want to revert, and then click Start.
Snapshots
5. Restore the version of the file or directory. l To restore all files in the selected directory, click Restore.
l
To copy the selected directory to another location, click Copy and then specify a location to copy the directory to. To restore a specific file, click Open, and then copy the file into the original directory, replacing the existing copy with the snapshot version.
3. Copy the file or directory by using the cp command. For example, the following command creates a copy of file1:
cp /ifs/.snapshot/Snapshot2012Jun04/archive/file1 /ifs/archive/ file1_copy
3. Clone a file from the snapshot by running the cp command with the -c option. For example, the following command clones test.txt from Snapshot2012Jun04:
cp -c /ifs/.snapshot/Snapshot2012Jun04/archive/test.txt /ifs/ archive/test_clone.text
Snapshots
from the last snapshot that it was assigned to. Unless you manually remove the alias, the alias remains attached to the last snapshot that it was assigned to. 1. Click Data Protection > SnapshotIQ > Snapshot Schedules. 2. In the Snapshot Schedules table, in the row of the snapshot schedule you want to modify, click View details. 3. In the Snapshot Schedule Details area, modify snapshot schedule attributes. 4. Next to each snapshot schedule attribute that you modified, click Save.
It is recommended that you do not create, delete, or modify snapshots locks unless you are instructed to do so by Isilon Technical Support. Deleting a snapshot lock that was created by OneFS might result in data loss. If you delete a snapshot lock that was created by OneFS, it is possible that the corresponding snapshot might be deleted while it is still in use by OneFS. If OneFS cannot access a snapshot that is necessary for an operation, the operation will malfunction and data loss might result. Modifying the expiration period of a snapshot lock can have impacts similar to deleting and creating snapshot locks. Reducing the duration period of a snapshot lock that has been created by OneFS might cause the lock to be deleted prematurely.
Snapshots
recommended that you extend the duration period of the snapshot by modifying the snapshot. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Create a snapshot lock by running the isi snapshot locks create command. For example, the following command applies a snapshot lock to "Snapshot April2012", sets the lock to expire in one month, and adds a description of "Maintenance Lock":
isi snapshot locks create "Snapshot April2012" --expires "1M" -comment "Maintenance Lock"
It is recommended that you do not modify the expiration dates of snapshot locks. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Modify a snapshot lock by running the isi snapshot locks modify command. For example, the following command sets a snapshot lock that is applied to "Snapshot April2012" and has an ID of 1 to expire in two days:
isi snapshot locks modify "Snapshot 2012Apr16" 1 --expires "2D"
It is recommended that you do not delete snapshot locks. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Delete a snapshot lock by running the isi snapshot locks delete command. For example, the following command deletes a snapshot lock that is applied to "Snapshot April2012" and has an ID of 1:
isi snapshot locks delete "Snapshot 2012Apr16" 1
The system prompts you to confirm that you want to delete the snapshot lock. 3. Type yes and then press ENTER.
ID Numerical identification number of the snapshot lock. Comment Description of the snapshot lock. This can be any string specified by a user. Expires The date that the snapshot lock will be automatically deleted by OneFS. Count The number of times the snapshot lock is held. The file clone operation can hold a single snapshot lock multiple times. If multiple file clones are created simultaneously, the file clone operation holds the same lock multiple times, rather
Modify a snapshot lock
105
Snapshots
than creating multiple locks. If you delete a snapshot lock that is held more than once, you will delete only one of the instances that the lock is held. In order to delete a snapshot lock that is held multiple times, you must delete the snapshot lock the same number of times as displayed in the count field.
SnapshotIQ settings
SnapshotIQ settings determine how snapshots behave and can be accessed. The following SnapshotIQ settings can be configured:
u
Snapshot Scheduling Determines whether snapshots can be generated. Disabling snapshot generation might cause some OneFS operations to fail. It is recommended that you do not disable this setting.
u
Auto-create Snapshots Determines whether snapshots are automatically generated according to snapshot schedules. Auto-delete Snapshots Determines whether snapshots are automatically deleted
according to their expiration dates.
106
Snapshots
107
Snapshots
108
OneFS enables you to replicate data from one Isilon cluster to another through the SyncIQ tool. To replicate data from one Isilon cluster to another, you must configure a SyncIQ license on both of the Isilon clusters. You can specify what data you want to replicate at the directory level, with the option to exclude specific files and sub-directories from being replicated. SyncIQ creates and references snapshots to replicate a consistent point-in-time image of a root directory. Metadata such as access control lists (ACLs) and alternate data streams (ADSs) are replicated along with data. You can use data replication to retain a consistent backup copy of your data on another Isilon cluster. OneFS offers automated failover and failback capabilities that enable you to continue operations on another Isilon cluster if a primary cluster becomes unavailable.
u u u u u u u u u u u u u
Replication policies and jobs..............................................................................110 Replication snapshots.........................................................................................112 Data failover and failback with SyncIQ.................................................................113 Recovery times and objectives for SyncIQ............................................................114 SyncIQ license functionality................................................................................115 Creating replication policies................................................................................115 Managing replication to remote clusters..............................................................124 Initiating data failover and failback with SyncIQ..................................................126 Managing replication policies.............................................................................130 Managing replication to the local cluster.............................................................133 Managing replication performance rules.............................................................134 Managing replication reports...............................................................................136 Managing failed replication jobs.........................................................................138
109
110
When a replication job runs, OneFS generates workers on the source and target cluster. Workers on the source cluster send data while workers on the target cluster write data. You can modify the maximum number of workers generated per node to control the amount of resources that a replication job is allowed to consume. For example, you can increase the maximum number of workers per node to increase the speed at which data is replicated to the target cluster. OneFS generates no more than 40 workers for a replication job. You can also reduce resource consumption through file-operation rules that limit the rate at which replication policies are allowed to send files. However, it is recommended that you do not create file-operation rules unless the files you intend on replicating are predictably similar in size, and not especially large.
Replication reports
OneFS generates reports that contain detailed information about replication job operations. You cannot customize the content of replication policy reports. OneFS routinely deletes replication policy reports. You can specify the maximum number of replication reports retained by OneFS and the length of time that replication reports are retained by OneFS. If the maximum number of replication reports is exceeded on a cluster, OneFS deletes reports beginning with the oldest reports. If a replication job fails, and you run the job again, OneFS creates two reports for that replication job and consolidates those reports into a single report. This is also true if you manually cancel a replication job and then start the replication policy again. If you delete a replication policy, OneFS automatically deletes any reports that were generated for that policy.
Replication snapshots
OneFS generates snapshots to facilitate replication, failover, and failback between Isilon clusters. Snapshots generated by OneFS can also be used for archival purposes on the target cluster.
for the new job. OneFS replicates only data that has changed since the last time a snapshot was generated for the replication policy. When a replication job completes, the system deletes the previous sourcecluster snapshot and retains the most recent snapshot until the next job runs.
113
Data failover
Data failover is the process of preparing data on the secondary cluster to be modified by clients. After you fail over to a secondary cluster, you can redirect clients to modify their data on the secondary cluster. Before failover is performed, you must create and run a replication policy on the primary cluster. The failover process is initiated on the secondary cluster. Failover is performed per replication policy, meaning that if the data you want to migrate is spread across multiple replication policies, you must initiate failover for each replication policy. You can use any replication policy to failover; however, if the action of the replication policy is set to copy, any file that was deleted on the primary cluster will be present on the secondary cluster. When the client connects to the secondary cluster, all files that were deleted on the primary cluster are available to the client. If you initiate failover for a replication policy while an associated replication job is running, the replication job fails, and the failover operation succeeds. Because data might be in an inconsistent state, OneFS uses the snapshot generated by the last successful replication job to revert data on the secondary cluster to the last recovery point. If a disaster occurs on the primary cluster, any modifications to data that were made after the last successful replication job started are not reflected on the secondary cluster. When a client connects to the secondary cluster, their data appears in the same state as it was when the last successful replication job was started.
Data failback
Data failback is the process of restoring clusters to the roles they occupied before the failover was performed, with the primary cluster hosting clients and the secondary cluster being replicated to for backup. The failback process includes updating the primary cluster with all of the modifications that were made to the data on the secondary cluster, preparing the primary cluster to be accessed by clients, and resuming data replication from the primary to the secondary cluster. At the end of the failback process, you can redirect users to resume accessing their data on the primary cluster. You can fail back data with any replication policy that meets the following criteria:
u u u u
Was failed over Is a synchronization policy Does not replicate a SmartLock directory Does not exclude any files or directories from replication
two hours to complete. If a disaster occurs five hours after a replication job begins, the RPO is four hours, because it has been four hours since a completed job began replicating data. RTO is the maximum amount of time required to make backup data available to clients after a disaster. The RTO is always less than or approximately equal to the RPO, depending on the rate at which replication jobs are created for a given policy. If replication jobs run continuously, meaning that another replication job is created before the previous replication job completes, the RTO is approximately equal to the RPO. When the secondary cluster is failed over, the data on the cluster is reset to the state it was in when the last job completed; resetting the data takes an amount of time proportional to the time it took users to modify the data. If replication jobs run on an interval, meaning that there is a period of time after a replication job completes before the next replication job starts, the relationship between RTO and RPO is dependant on whether a replication job is running when the disaster occurs. If a job is in progress when a disaster occurs, the RTO is roughly equal to the RPO. However, if a job is not running when a disaster occurs, the RTO is equal to a negligible amount of time. This is because the secondary cluster was not modified since the last replication job ran, and the failover procedure is essentially instantaneous.
/ifs/data/media/music
SyncIQ license functionality
115
/ifs/data/media/movies
Excluded directories:
u u
/ifs/data/archive /ifs/data/media/music/working
In this example, the setting that excludes the /ifs/data/archive directory has no effect because the /ifs/data/archive directory is not under either of the included directories; the /ifs/data/archive directory is not replicated regardless of whether the directory is explicitly excluded. However, the setting that excludes the /ifs/data/ media/music/working directory does have an effect, because the directory would be replicated if the setting was not specified. In addition, if you exclude a directory that contains the source directory, the excludedirectory setting has no effect. For example, if the root directory of a policy is /ifs/data, explicitly excluding the /ifs directory has no effect. Any directories that you explicitly include or exclude must be contained in or under the specified root directory. For example, consider a policy in which the specified root directory is /ifs/data. In this example, you could include both the /ifs/data/media and the /ifs/data/users/ directories because they are under /ifs/data. Excluding directories from a synchronization policy does not cause the directories to be deleted on the target cluster. For example, consider a replication policy that synchronizes /ifs/data on the source cluster to /ifs/data on the target cluster. If the policy excludes /ifs/data/media from replication, and /ifs/data/media/file exists on the target cluster, running the policy does not cause /ifs/data/media/file to be deleted from the target cluster.
116
For synchronization policies, if you modify the comparison operators or comparison values of a file attribute, and a file no longer matches the specified file-matching criteria, the file is deleted from the target the next time the job is run. This rule does not apply to copy policies.
Date created Includes or excludes files based on when the file was created. This option
is available for copy policies only. You can specify a relative or specific date and time. Time settings are based on a 24-hour clock.
Date accessed Includes or excludes files based on when the file was last accessed. This
option is available for copy policies only. This setting is available only if the global accesstime-tracking option of the cluster is enabled. You can specify a relative date and time, such as "two weeks ago", or specific date and time, such as "January 1, 2012." Time settings are based on a 24-hour clock.
Date modified Includes or excludes files based on when the file was last modified. This
option is available for copy policies only. You can specify a relative or specific date and time. Time settings are based on a 24-hour clock.
File name Includes or excludes files based on the file name. You can specify to include or
exclude full or partial names that contain specific text. The following wildcards are accepted:
Alternatively, you can filter file names by using POSIX regular-expression (regex) text. Regular expressions are sets of symbols and syntax that are used to match patterns of text. These expressions can be more powerful and flexible than simple wildcard characters. Isilon clusters support IEEE Std 1003.2 (POSIX.2) regular expressions. For more information about POSIX regular expressions, see the BSD man pages. Wildcard *
Description Matches any string in place of the asterisk. For example, specifying "m*" would match "movies" and "m123"
[]
Matches any characters contained in the brackets, or a range of characters separated by a dash. For example, specifying "b[aei]t" would match "bat", "bet", and "bit" For example, specifying "1[4-7]2" would match "142", "152", "162", and "172"
117
Wildcard -
Description
You can exclude characters within brackets by following the first bracket with an exclamation mark. For example, specifying "b[!ie]" would match "bat" but not "bit" or "bet" You can match a bracket within a bracket if it is either the first or last character. For example, specifying "[[c]at" would match "cat", and "[at" You can match a dash within a bracket if it is either the first or last character. For example, specifying "car[-s]" would match "cars", and "car-"
Matches any character in place of the question mark. For example, specifying "t?p" would match "tap", "tip", and "top"
Path Includes or excludes files based on the file path. This option is available for copy
policies only. You can specify to include or exclude full or partial paths that contain specified text. You can also include the wildcard characters *, ?, and [ ].
Size Includes or excludes files based on size. File sizes are represented in multiples of 1024, not 1000.
Type Includes or excludes files based one of the following file-system object types:
u u u
Click Connect to only the nodes in the subnet and pool if the target cluster name specifies a SmartConnect zone.
3. Specify which nodes you want replication policies to connect to when the policy is run. Connect policies to all nodes on Click Run the policy on all nodes in this cluster. a source cluster.
118
a. Click Run the policy only on nodes in the specified subnet and pool. b. From the Subnet and pool list, select the subnet and pool .
Source directory Included or excluded directories File-criteria statement Target cluster name or address This applies only if you target a different cluster. If you modify the IP or domain name of a target cluster, and then modify the replication policy on the source cluster to match the new IP or domain name, a full replication is not performed. Target directory
To copy all files from the source directory to the target directory and delete any files on the target directory that are not in the source directory, click Synchronize.
6. In the Run job area, specify whether the job runs according to a schedule, or only when initiated by a user. Run the policy only Click Manually. when manually initiated by a user.
119
Run the policy a. Click Scheduled. automatically according b. Click Edit schedule. to a schedule. c. Specify a schedule. If you configure a replication policy to run more than once a day, you cannot configure the interval to span across two calendar days. For example, you can configure a replication policy to run every hour starting at 7:00 PM and ending at 11:00 PM, but you cannot configure a replication policy to run every hour starting at 7:00 PM and ending at 1:00 AM. What to do next The next step in the process of creating a replication policy is specifying source directories and files.
The host name of any node in the target cluster. The name of a SmartConnect zone in the target cluster. The IPv4 or IPv6 address of any node in the target cluster. localhost This will replicate data to another directory on the local cluster. Replication does not support dynamic pools.
2. Specify how you want the replication policy to connect to the target cluster by selecting one of the following options. l Click Connect to any nodes in the cluster.
120
Click Connect to only the nodes in the subnet and pool if the target cluster name specifies a SmartConnect zone.
3. In the Target directory field, type the absolute path of the directory on the target cluster that you want to replicate data to.
If you specify an existing directory on the target cluster, ensure that the directory is not the target of another replication policy. If this is a synchronization policy, ensure that the directory is empty. All files are deleted from the target of a synchronization policy the first time the policy is run. If the specified target directory does not already exist on the target cluster, the directory created the first time the job is run. It is recommended that you do not specify the /ifs directory. If you specify the /ifs directory, the entire target cluster is set to a read-only state, preventing you from storing any other data on the cluster. If this is a copy policy, and files exist in the target directory that are also present in the source directory, those files are overwritten when the job is run. What to do next The next step in the process of creating a replication policy is specifying policy target snapshot settings.
2. Optional: To modify the default alias of the last snapshot created according to this replication policy, in the Snapshot alias name field, type a new alias. You can specify the alias name as a snapshot naming pattern. For example, the following naming pattern is valid:
%{PolicyName}-on-%{SrcCluster}-latest
3. Optional: To modify the snapshot naming pattern, in the Snapshot naming pattern field, type a naming pattern. Each snapshot generated for this replication policy is assigned a name based on this pattern. For example, the following naming pattern is valid:
%{PolicyName}-from-%{SrcCluster}-at-%H:%M-on-%m-%d-%Y
4. Specify whether you want OneFS to automatically delete snapshots generated according to this policy.
Create a replication policy
121
l l
Click Do not delete any archival snapshots. Click Delete archival snapshots when they expire and specify an expiration period.
What to do next The next step in the process of creating a replication policy is configuring advanced policy settings.
Click Error. Click Notice. Click Network Activity. Click File Activity.
3. If you want OneFS to perform a checksum on each file data packet that is affected by the replication job, select the Validate file integrity check box. If you enable this option, and the checksum values for a file data packet do not match, OneFS retransmits the affected packet. 4. To configure shared secret authentication, in the Shared secret field, type a shared secret. To establish this type of authentication, you must configure both the source and target cluster to require the same shared secret. For more information, see the Isilon Knowledge Base. This feature does not perform any encryption. 5. To modify the length of time OneFS retains replication reports for the policy, in the Keep reports for area, specify a length of time. After the specified expiration period has passed for a report, OneFS automatically deletes the report. Some units of time are displayed differently when you view a report than how they were originally entered. Entering a number of days that is equal to a corresponding value in weeks, months, or years results in the larger unit of time being displayed. For example, if you enter a value of 7 days, 1 week appears for that report after it is created. This change occurs because OneFS internally records report retention times in seconds and then converts them into days, weeks, months, or years for display. 6. Specify which nodes you want the replication policy to connect to when the policy is run. Connect the policy to all nodes on Click Run the policy on all nodes in this cluster. the source cluster.
122
Connect the policy only to nodes contained in a specified subnet and pool.
Click Run the policy only on nodes in the specified subnet and pool. From the Subnet and pool list, select the subnet and pool .
Replication does not support dynamic pools. 7. Specify whether to record information about files that are deleted by synchronization jobs by selecting one of the following options: l Click Record when a synchronization deletes files or directories.
l
This option is applicable for synchronization policies only. What to do next The next step in the process of creating a replication policy is saving the replication policy settings.
7. Ensure that the Delete domain check box is cleared. 8. In the Domain root path field, type the path of a source directory of a replication policy, and then click Start.
Run The status of the job. Policy The name of the associated replication policy. Started The time the job started. Elapsed Indicates how much time has elapsed since the job started. Transferred The number of files that were transferred during the job run, and the total size of all transferred files. Sync Type The type of replication being performed. The possible values are Initial, which indicates that either a differential or a full replication is being performed; Upgrade, which indicates that a policy-conversion replication is being performed; and Incremental, which indicates that only modified files are being transferred to the target cluster. Source The source directory on the source cluster.
125
u u
Target The target directory on the target cluster. Actions Displays any job-related actions that you can perform.
Run Indicates the status of the job. A green icon indicates that the last job completed
successfully. A yellow icon indicates that the last job did not complete successfully, but that an earlier job did complete successfully. A red icon indicates that jobs have run, but that none of the jobs completed successfully. If no icon appears, the job was not run.
u u u u u
Policy The name of the associated replication policy. Started The time at which the job started. Ended The time at which the job finished running. Duration Indicates the total amount of time that the job ran for. Transferred The number of files that were transferred during the job run, and the total
size of all transferred files.
Sync Type The type of replication that was performed. The possible values are Initial, which indicates that either a differential or a full replication was performed; Upgrade, which indicates that a policy-conversion replication occurred after upgrading the OneFS operating system or merging policies; and Incremental, which indicates that only modified files were transferred to the target cluster. Source The source directory on the source cluster. Target The target directory on the target cluster.
u u
126
If a replication job is currently running, it is recommended that you wait until the replication job completes, and then click Allow Writes. This will maintain the association between the primary and secondary cluster. If the primary cluster became unavailable while a replication job was running, click Break. This will break the association between the primary and secondary cluster.
3. If you clicked Break, restore any files left in an inconsistent state. a. Delete all files that were not committed to a WORM state from the target directory. b. Copy all files from the failover snapshot to the target directory. Failover snapshots are named according to the following naming pattern:
SIQ-Failover-<policy-name>-<year>-<month>-<day>_<hour>-<minute><second>
4. If any SmartLock directory configuration settings, such as an autocommit time period, were specified for the source directory of the replication policy, apply those settings to the target directory. What to do next Redirect clients to begin accessing the secondary cluster.
Failover revert
You can perform a failover revert if, for example, the primary cluster becomes available before data is modified on the secondary cluster. You also might want to perform failover revert if you were failing over for testing purposes. Failover revert enables you to replicate data from the primary cluster to the secondary cluster again. Failover revert does not migrate data back to the primary cluster. If clients modified data on the secondary cluster, and you want to migrate the modified data back to the primary cluster, you must fail back to the primary cluster. Failover revert is not supported for SmartLock directories. Complete the following procedure for each replication policy you want to fail over. Before you begin Fail over a replication policy.
Fail over SmartLock directories
127
1. On the secondary Isilon cluster, click Data Protection > SyncIQ > Local Targets . 2. In the Local Targets table, in the row of a replication policy, click Disallow Writes, and then, in the confirmation dialog box, click Yes.
3. On the secondary cluster, replicate data to the primary cluster by using the mirror policies. You can replicate data either by manually starting the mirror policies or by modifying the mirror policies and specifying a schedule. 4. Disallow client access to the secondary cluster and run each mirror policy again. 5. On the primary cluster, click Data Protection > SyncIQ > Local Targets . 6. For each mirror policy, in the Local Targets table, in the row of the mirror policy, click Allow Writes. 7. On the secondary cluster, click Data Protection > SyncIQ > Policies . 8. For each mirror policy, in the Policies table, in the row of the policy, click Prepare resync. What to do next Redirect clients to begin accessing the primary cluster.
128
For example, the following command enables privileged delete functionality for / ifs/data/dir:
isi worm modify --path /ifs/data/dir --privdel on
2. Disable the autocommit time period for the directory you want to fail back to by running the isi worm modify command. For example, the following command disables the autocommit time period for /ifs/ data/dir:
isi worm modify --path /ifs/data/dir --autocommit none
The SmartLock directory you are failing back is an enterprise directory. The privileged delete functionality was not permanently disabled for the directory. During the failover process, you maintained the association between the source and target cluster.
Complete the following procedure for each replication policy you want to fail back. Before you begin Prepare SmartLock directories for failback. 1. On the secondary cluster, create a replication policy that meets the following requirements: l The source directory is the target directory of the policy you are failing back.
l
If you are failing back to an enterprise directory with the privileged delete functionality enabled, the target directory of the policy must be the source directory of the policy you are failing back. If you are failing back to a compliance directory, or an enterprise directory with the privileged delete functionality permanently disabled, the target must be an empty SmartLock directory. The directory must be of the same SmartLock type as the source directory of the policy you are failing back. For example, if the target directory is a compliance directory, the source must also be a compliance directory.
2. Optional: Replicate data to the primary cluster by running the policy you created. Continue to replicate data until a time when client access to the cluster is minimal. For example, you might wait until a weekend when client access to the cluster is reduced. 3. Disallow client access to the secondary cluster and run the policy that you created. 4. On the primary cluster, click Data Protection > SyncIQ > Local Targets . 5. In the Local Targets table, for the replication policy that you created, click Allow Writes. 6. Optional: If any SmartLock directory configuration settings, such as an autocommit time period, were specified for the source directory of the replication policy, apply those settings to the target directory. 7. Prepare the source directory of the replication policy on the secondary cluster for failback.
Fail back SmartLock directories
129
For more information, see Prepare SmartLock directories for failback. 8. Begin replicating data by enabling or replacing the replication policy that you originally failed over. What to do next Redirect clients to begin accessing the primary cluster.
Source directory Included or excluded directories File-criteria statement Target cluster This applies only if you target a different cluster. If you modify the IP or domain name of a target cluster, and then modify the replication policy on the source cluster to match the new IP or domain name, a full replication is not performed. Target directory
1. Click Data Protection > SyncIQ > Policies. 2. In the Policies table, click the name of the policy you want to modify. 3. Modify the settings of the replication policy, and then click Submit.
2. In the Policies table, in the Actions column of the policy, click either Enable or Disable. If neither Enable nor Disable is displayed in the Actions column, verify that an associated replication job is not running. If an associated replication job is not running, ensure that the SyncIQ license is configured on the cluster.
Run If a replication job is running for this policy, indicates whether the job is running or
paused. If no job is running, indicates whether the SyncIQ tool is disabled on the cluster. If no icon appears, indicates that SyncIQ is enabled and that no replication job is currently running.
Data Indicates the status of the last run of the job. A green icon indicates that the last job
completed successfully. A yellow icon indicates that the last job did not complete successfully, but that an earlier job did complete successfully. If no icon appears, a job for the policy was not run.
u u u
Policy Displays the name of the policy. Last Known Good Indicates when the last successful job ran. Schedule Indicates when the next job is scheduled to run. A value of Manual indicates
that the job can be run only manually.
u u u
Source Displays the source directory path. Target Displays the target directory path. Actions Displays any policy-related actions that you can perform.
Policy name Name of the policy. Description Optional string that describes the policy. For example, the description might explain the purpose or function of the policy. Action Describes the how the policy replicates data. All policies copy files from the
source directory to the target directory and update files in the target directory to match files on the source directory. The action dictates how deleting a file on the source directory affects the target. The following values are valid:
u
Copy If a file is deleted in the source directory, the file is not deleted in the target
directory.
Synchronize Deletes files in the target directory if they are no longer present on the
source. This ensures that an exact replica of the source directory is maintained on the target cluster.
131
Run job Specifies whether the job is run automatically according to a schedule, or only
manually when specified by a user.
Root directory The full path of the source directory. Data is replicated from the source
directory to the target directory.
Exclude directories Determines which directories are excluded from replication. Any
directories specified by this setting are not replicated.
u u
File criteria Determines which files are excluded from replication. Name or address (of target cluster) The IP address or fully qualified domain name
of the target cluster.
Target directory The full path of the target directory. Data is replicated to the target directory from the source directory. Create snapshots Determines whether archival snapshots are generated on the target
cluster.
Snapshot alias name Specifies an alias for the latest archival snapshot taken on the
target cluster.
Snapshot naming pattern Specifies how archival snapshots are named on the target
cluster.
Snapshot expiration Specifies how long archival snapshots are retained on the target
cluster before they are automatically deleted by the system.
Workers per node Specifies the number of workers per node that are generated by
OneFS to perform each replication job for the policy.
Log level Specifies the amount of information that is recorded in the logs for replication jobs. More verbose options include all information from less verbose options. The following list describes the log levels from least to most verbose:
u
Notice Includes job and process-level activity, including job starts and stops, and
worker coordination information. It is recommended that you select this option.
u u
Error Includes events related to specific types of failures. Network Activity Includes more job-level activity and work-item information,
including specific paths and snapshot names.
File Activity Includes a separate event for each action taken on a file. Do not select this option without first consulting Isilon Technical Support.
Replication logs are typically used only for debugging purposes. If necessary, you can log in to a node through the command-line interface and view the contents of the /var/log/ isi_migrate.log file on the node.
u
Check integrity Determines whether OneFS performs a checksum on each file data
packet that is affected by a replication job. If a checksum value does not match, OneFS retransmits the affected file data packet.
Shared secret Determines whether OneFS references a shared secret on the source and target cluster to prevent certain types of attacks. This feature does not perform any encryption.
132
Keep reports for Specifies how long replication reports are kept before they are
automatically deleted by OneFS.
Source node restrictions Specifies whether replication jobs connect to any nodes in
the cluster or if jobs can connect only to nodes in a specified subnet and pool.
Delete on synchronization Determines whether OneFS records when a synchronization job deletes files or directories on the target cluster.
The following replication policy fields are available only through the OneFS command-line interface.
u u
Password Specifies a password to access the target cluster. Max reports Specifies the maximum number of replication reports that are retained for
this policy.
Diff sync Determines whether full or differential replications are performed for this policy.
Full or differential replications are performed the first time a policy is run and after a policy is reset.
Rename pattern Determines whether snapshots generated for the replication policy on
the source cluster are deleted when the next replication policy is run. If specified, snapshots that are generated for the replication policy on the source cluster are retained and renamed according to the specified rename pattern. If not specified, snapshots generated on the source cluster are deleted. Specifying this setting does not require that the SnapshotIQ license be configured on the cluster.
Rename expiration If snapshots generated for the replication policy on the source
cluster are retained, specifies an expiration period for the snapshots.
Depending on the amount of data being replicated, a full or differential replication can take a very long time to complete.
Managing replication to the local cluster
133
1. Click Data Protection > SyncIQ > Local Targets. 2. In the Local Targets table, in the Actions column of the policy, click Break. 3. In the Confirm dialog box, click Yes.
Green Indicates that the last job completed successfully. Yellow Indicates that the last job did not complete successfully, but that an earlier
job did complete successfully.
Red Indicates that jobs have run, but that none of the jobs completed successfully.
A yellow or red icon might indicate that the policy is in an unrunnable state. You can view more detailed policy-status information and, if necessary, resolve the source-target association, through the web administration interface on the source cluster.
u u
Policy The name of the replication policy. Updated The time when data about the policy or job was last collected from the source
cluster.
u u u
Source The source directory on the source cluster. Target The target directory on the target cluster. Coordinator IP The IP address of the node on the source cluster that is acting as the
job coordinator.
3. In the Edit Limit dialog box, in the Limit (bits/sec) area, specify the maximum number of bits per second that replication rules are allowed to send. 4. In the Days area, select the days of the week that you want to apply the rule. 5. In the Start and End areas, specify the period of time that you want to apply the rule. 6. Optional: To add an optional description of this network traffic rule, in the Description box, type a description. 7. Click Enabled. If you do not select Enabled, the rule is disabled by default. 8. Click Submit.
135
specified for a report. Excess reports are periodically deleted by OneFS; however, you can manually delete all excess replication reports at any time. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Delete excess replication reports by running the following command:
isi sync report rotate
What to do next If OneFS did not delete the desired number of replication reports, configure the replication report settings, and then repeat this procedure.
Green Indicates that the last job completed successfully. Yellow Indicates that the last job did not complete successfully, but that an earlier job did complete successfully. Red Indicates that jobs have run, but were unsuccessful.
Policy The name of the associated policy for the job. You can view or edit settings for the policy by clicking the policy name. Started, Ended, and Duration Indicates when the job started and ended, and the
duration of the job.
Transferred The total number of files that were transferred during the job run, and the total size of all transferred files. For assessed policies, Assessment appears. Sync Type The action that was performed by the replication job.
The following actions are displayed:
u u
Initial Sync Indicates that either a differential or a full replication was performed. Incremental Sync Indicates that only modified files were transferred to the target
cluster.
Failover / Failback Allow Writes Indicates that writes were enabled on a target
directory of a replication policy.
137
Failover / Failback Resync Prep Domain Mark Indicates that a SyncIQ domain
was created for the source directory. This is the second step in the failback preparation process.
Failover / Failback Resync Prep Restore Indicates that a source directory was restored to the last recovery point. This is the third step in the failback preparation process. Failover / Failback Resync Prep Finalize Indicates that a mirror policy was
created on the target cluster. This is the last step in the failback preparation process.
u u u
Source The source directory on the source cluster. Target The target directory on the target cluster. Actions Displays any report-related actions that you can perform.
Depending on the amount of data being replicated, a full or differential replication can take a very long time to complete. Reset a replication policy only if you cannot fix the issue that caused the replication error. If you fix the issue that caused the error, resolve the policy instead of resetting the policy. 1. Click Data Protection > SyncIQ > Policies. 2. In the Policies table, in the Actions column of a policy, click Reset. 3. In the Confirm dialog box, type yes and then click Yes.
To perform a differential replication, enable the --diff_sync option. For example, the following command enables differential synchronization for newPolicy:
isi sync policy modify newPolicy --diff_sync on
3. Run the policy by running the isi sync policy run command. For example, the following command runs newPolicy:
isi sync policy run newPolicy
139
140
An Isilon cluster is designed to continuously serve data, even when one or more components simultaneously fail. OneFS ensures data availability by striping or mirroring data across the cluster. If a cluster component fails, data stored on the failed component is available on another component. After a component failure, lost data is restored on healthy components by the FlexProtect proprietary system. Data protection is specified at the file level, not the block level, enabling the system to recover data quickly. Because all data, metadata, and parity information is distributed across all nodes in the cluster, an Isilon cluster does not require a dedicated parity node or drive. This ensures that no single node limits the speed of the rebuild process.
u u u u u u
File striping.........................................................................................................142 Data protection levels.........................................................................................142 FlexProtect data recovery.....................................................................................142 Managing protection levels.................................................................................144 Data protection level information........................................................................144 Data protection level disk space usage................................................................145
141
File striping
OneFS uses the back-end network to automatically allocate and stripe data across nodes and disks in the cluster. OneFS protects data as the data is being written. No separate action is necessary to stripe data. OneFS breaks files into smaller logical chunks called stripes before writing the files to disk; the size of each file chunk is referred to as the stripe unit size. Each OneFS block is 8 KB, and a stripe unit consists of 16 blocks, for a total of 128 KB per stripe unit. During a write, OneFS breaks data into stripes and then logically places the data in a stripe unit. As OneFS stripes data across the cluster, OneFS fills the stripe unit according to the number of nodes and protection level. OneFS can continuously reallocate data and make storage space more usable and efficient. As the cluster size increases, OneFS stores large files more efficiently.
Because data is rebuilt in the free space of the cluster, the cluster does not require a dedicated hot-spare node or drive in order to recover from a component failure. Because a certain amount of free space is required to rebuild data, it is recommended that you reserve adequate free space through the virtual hot spare feature. As a cluster grows larger, data restriping operations become faster. As you add more nodes, the cluster gains more CPU, memory, and disks to use during recovery operations.
Smartfail
OneFS protects data stored on failing nodes or drives through a process called smartfailing. During the smartfail process, OneFS places a device into quarantine. Quarantined devices can be used only for read operations. While the device is quarantined, OneFS reprotects the data on the device by distributing the data to other devices. After all data migration is complete, OneFS logically removes the device from the cluster, the cluster logically changes its width to the new configuration, and the node or drive can be physically replaced. OneFS automatically smartfails devices only as a last resort. Although you can manually smartfail nodes or drives, it is recommended that you first consult Isilon Technical Support. Occasionally a device might fail before OneFS detects a problem. If a drive fails without being smartfailed, OneFS automatically starts rebuilding the data to available free space on the cluster. However, because a node might recover from a failure, if a node fails, OneFS does not start rebuilding data unless the node is logically removed from the cluster.
Node failures
Because node loss is often a temporary issue, OneFS does not automatically start reprotecting data when a node fails or goes offline. If a node reboots, the file system does not need to be rebuilt because it remains intact during the temporary failure. If an N+1 data protection level is configured, and one node fails, all of the data is still accessible from every other node in the cluster. If the node comes back online, the node rejoins the cluster automatically without requiring a full rebuild. To ensure that data remains protected, if you physically remove a node from the cluster, you must also logically remove the node from the cluster. After you logically remove a node, the node automatically reformats its own drives, and resets itself to the factory default settings. The reset occurs only after OneFS has confirmed that all data has been reprotected. You can logically remove a node using the smartfail process. It is important that you use the smartfail process only when you want to permanently remove a node from the cluster. If you remove a failed node before adding a new node, data stored on the failed node must be rebuilt in the free space in the cluster. After the new node is added, the data is then distributed to the new node. It is more efficient to add a replacement node to the cluster before failing the old node because OneFS can immediately use the replacement node to rebuild the data stored on the failed node.
Smartfail
143
Definition The cluster can absorb the failure of any single drive or the unscheduled shutdown of any single node without causing any loss in stored data. The cluster can recover from two simultaneous drive failures or one node failure without sustaining any data loss. The cluster can recover from two simultaneous drive or node failures without sustaining any data loss. The cluster can recover from three simultaneous drive failures or one node failure without sustaining any data loss. The cluster can recover from three simultaneous drive failures or one node failure without sustaining any data loss.
N+2:1
N+2
N+3:1
N+3
144
Definition The cluster can recover from four simultaneous drive or node failures without sustaining any data loss.
Nx (Data mirroring)
The cluster can recover from N - 1 node failures without sustaining data For example, 5x requires loss. For example, 5x protection a minimum of five nodes. means that the cluster can recover from four node failures
+2:1 4 + 2 (33%) 6 + 2 (25%) 8 + 2 (20%) 10 + 2 (17%) 12 + 2 (14%) 14 + 2 (12.5%) 16 + 2 (11%) 16 + 2 (11%) 16 + 2 (11%) 16 + 2 (11%)
+3:1 9 + 3 (25%) 12 + 3 (20%) 15 + 3 (17%) 15 + 3 (17%) 15 + 3 (17%) 15 + 3 (17%) 15 + 3 (17%) 15 + 3 (17%) 15 + 3 (17%)
2 +1 (33%) 3 +1 (25%) 4 +1 (20%) 5 +1 (17%) 6 +1 (14%) 7 +1 (13%) 8 +1 (11%) 9 +1 (10%) 11 +1 (8%) 13 + 1 (7%)
145
Number of +1 nodes 16 18 20 30
The parity overhead for mirrored data protection is not affected by the number of nodes in the cluster. The following table describes the parity overhead for each mirrored data protection level. 2x 50%
3x 67%
4x 75%
5x 80%
6x 83%
7x 86%
8x %88
146
OneFS enables you to back up and restore file-system data through the Network Data Management Protocol (NDMP). From a backup server, you can direct backup and recovery processes between an Isilon cluster and backup devices such as tape devices, media servers, and virtual tape libraries (VTLs). OneFS supports both NDMP three-way backup and NDMP two-way backup. During a three-way NDMP backup operation, a data management application (DMA) on a backup server instructs the cluster to start backing up data to a tape media server that is either attached to the LAN or directly attached to the DMA. During a two-way NDMP backup, a DMA on a backup server instructs a Backup Accelerator node on the cluster to start backing up data to a tape media server that is attached to the Backup Accelerator node. Two-way NDMP backup is the most efficient method in terms of cluster resource consumption; however, two-way NDMP backup requires that one or more Backup Accelerator nodes be attached to the cluster. In both the two-way and three-way NDMP backup models, file history data is transferred from the cluster to a backup server. Before a backup begins, OneFS creates a snapshot of the targeted directory. OneFS then backs up the snapshot, which ensures that the backup image represents a specific point in time. After the backup is completed or canceled, OneFS automatically deletes the snapshot. You do not need to configure a SnapshotIQ license on the cluster to perform NDMP backups. However, if a SnapshotIQ license is configured on the cluster, you can generate a snapshot through the SnapshotIQ tool, and then back up the snapshot. If you back up a snapshot that you generated, OneFS does not create another snapshot for the backup. If you are backing up SmartLock compliance directories, it is recommended that you do not specify autocommit time periods for the SmartLock directories.
u u u u u u u u u u u u u u u
NDMP two way backup........................................................................................148 NDMP protocol support.......................................................................................148 Supported DMAs.................................................................................................148 NDMP hardware support.....................................................................................149 NDMP backup limitations....................................................................................149 NDMP performance recommendations................................................................149 Excluding files and directories from NDMP backups............................................151 Configuring basic NDMP backup settings............................................................152 Create an NDMP user account.............................................................................153 Managing NDMP user accounts...........................................................................154 Managing NDMP backup devices.........................................................................154 Managing NDMP backup ports............................................................................156 Managing NDMP backup sessions.......................................................................157 View NDMP backup logs......................................................................................159 NDMP environment variables..............................................................................159
NDMP backup
147
NDMP backup
Full (0) NDMP backups Level-based (1-9) NDMP incremental backups Token-based NDMP backups NDMP TAR and dump types If you specify the NDMP dump backup type, the backup will be stored on the backup device in TAR format.
u u u u u u u
Path-based and dir/node file history format. Direct Access Restore (DAR) Directory DAR (DDAR) Including and excluding specific files and directories from backup Backup of file attributes Backup of Access Control Lists (ACLs) Backup of Alternate Data Streams (ADSs)
Supported DMAs
NDMP backups are coordinated by a data management application (DMA) that runs on a backup server. OneFS supports the following DMAs:
148
NDMP backup
u u u u u u u
Symantec NetBackup EMC Networker Symantec Backup Exec IBM Tivoli Storage Manager Quest Software NetVault CommVault Simpana Atempo Time Navigator
LTO-3 LTO-4 LTO-5 FalconStor VTL 5.20 Data Domain VTL 5.1.04 or later
Point-to-point (also known as direct) topologies are not supported for VTLs. VTLs must include a Fibre Channel switch.
OneFS does not back up file system configuration data, such as file protection level policies and quotas. OneFS does not support multiplexing across multiple streams. OneFS does not support shared storage options to shared media drives. OneFS does not support restoring data from another file system other than OneFS. However, you can migrate data from a NetApp storage system to OneFS. Backup Accelerator nodes cannot interact with more than 1024 device paths, including paths of tape and media changer devices. For example, if each device has four paths, you can connect 256 devices to a Backup Accelerator node. If each device has two paths, you can connect 512 devices. OneFS does not support more than 64 concurrent NDMP sessions.
u u u
If you are backing up multiple directories that contain small files, set up a separate schedule for each directory. If you are performing three-way NDMP backups, run multiple NDMP sessions on multiple nodes.
NDMP hardware support
149
NDMP backup
Do not perform both three-way and two-way NDMP backup operations on the same cluster. Install the latest patches from Isilon and your data management application (DMA) vendor when available. Restore files through Direct Access Restore (DAR) and Directory DAR (DDAR). This is especially recommended if you restore files frequently. However, it is recommended that you do not use DAR to restore a full backup or a large number of files. Use the largest tape record size available for your version of OneFS. The largest tape record size for OneFS versions 6.5.5 and later is 256 k. The largest tape record size for versions of OneFS earlier than 6.5.5 is 128 k. If possible, do not include or exclude files from backup. Including or excluding files can affect backup performance, due to filtering overhead during tree walks. Limit the depth of nested subdirectories in your file system. Limit the number of files in a directory. Distribute files across multiple directories instead of including a large number of files in a single directory.
u u
Networking recommendations The following best practices are recommended for configuring the connection between a cluster and NDMP backup devices:
u u
Assign static IP addresses to Backup Accelerator nodes. Configure SmartConnect zones that are dedicated to NDMP backup activity. It is recommended that you connect NDMP sessions only through SmartConnect zones that are exclusively used for NDMP backup.
Configure multiple policies when scheduling backup operations, with each policy capturing a portion of the file system. Do not attempt to back up the entire file system through a single policy.
Backup Accelerator recommendations The following best practices are recommended if you are performing NDMP two-way backups:
u
Run four concurrent streams per Backup Accelerator node. This is recommended only if you are backing up a significant amount of data. Running four concurrent streams might not be possible or necessary for smaller backups.
Attach more Backup Accelerator nodes to larger clusters. The recommended number of Backup Accelerator nodes depends on the type of nodes that are included in the cluster. The following table lists the recommended number of Backup Accelerator nodes to include in a cluster. Node type i-Series X-Series NL-Series S-Series Recommended number of nodes per Backup Accelerator node 5 3 3 3
150
NDMP backup
DMA-specific recommendations If possible, configure your DMA according to the following best practices:
u
If you perform backup operations through Symantec NetBackup, it is recommended that you increase throughput by specifying an NDMP buffer size of 256 k. It is recommended that you use path-based file history instead of dir/node file history. Enable multistreaming, which enables OneFS to back up data to multiple tape devices concurrently.
archive*
[]
user_?
user\ 1
Although you can specify both anchored and unanchored patterns, it is recommended that you do not specify unanchored patterns. Unanchored patterns target a string of text that might belong to several files or directories, such as home or user1. An anchored pattern targets a file path, such as /home/user1. Specifying unanchored patterns can degrade the performance of backups and result in empty directories being backed up. For example, you might back up the /ifs/data directory that contains the following files and directories:
u u
/ifs/data/home/user1/file.txt /ifs/data/home/user2/user1/file.txt
Excluding files and directories from NDMP backups
151
NDMP backup
u u
/ifs/data/home/user3/other/file.txt /ifs/data/home/user4/emptydirectory
If you include the home directory, you will back up the following files and directories:
u u u u
The empty directory /ifs/data/home/user4/home/emptydirectory would be backed up. If you specify both include and exclude patterns, any excluded files or directories under included directories are not backed up. If the excluded directories are not contained in any of the included directories, the exclude specification has no effect. For example, assume that you are backing up the /ifs/data directory, and you include the following directories:
u u
/ifs/data/media/music /ifs/data/media/movies
/ifs/data/media/music/working /ifs/data/archive
In this example, the setting that excludes the /ifs/data/archive directory has no effect because the /ifs/data/archive directory is not contained under either of the included directories. However, the setting that excludes the /ifs/data/media/music/ working directory does have an effect because the /ifs/data/media/music/working directory is contained under the /ifs/data/media/music directory. If the setting had not been specified, the /ifs/data/media/music/working would have been backed up.
To modify the DMA vendor, from the DMA vendor list, select the name of the DMA vendor you are coordinating backup operations through.
152
NDMP backup
If your DMA vendor is not included in the list, select generic. However, any vendors not included in the list are not supported. 4. Optional: To add an NDMP user account through which a DMA can access the cluster, click Add administrator. a. In the Add Administrator dialog box, in the Name field, type a name for the account. b. In the Password and Confirm password fields, type a password for the account. c. Click Submit.
Port number The number of the port through which data management applications (DMAs) can connect to the cluster. DMA vendor The DMA vendor that the cluster is configured to interact with.
153
NDMP backup
NDMP backup
4. Optional: To scan only a specific port for NDMP devices, from the Ports list, select a port. If you specify a port and a node, only the specified port on the node is scanned. However, if you specify only a port, the specified port will be scanned on all nodes. 5. Optional: To remove entries for devices or paths that have become inaccessible, select the Delete inaccessible paths or devices check box. 6. Click Submit. Results For each device that is detected, an entry is added to either the Tape Devices or Media Changers tables.
155
NDMP backup
Name Specifies a unique device name assigned by OneFS. State Indicates whether data is currently being backed up to or restored from the device. If the device is in use, Read/Write appears. If the device is not in use, Closed appears. WWN Specifies the world wide node name (WWNN) of the device. Product Specifies the name of the device vendor, and the model name or number of the
device.
u u
u u
Serial Number Specifies the serial number of the device. Paths Specifies the name of the Backup Accelerator node that the device is attached to,
and the port number or numbers to which the device is connected.
u u
LUN Specifies the logical unit number (LUN) of the device. Port ID Specifies the port ID of the device that binds the logical device to the physical
device.
WWPN Specifies the world wide port name (WWPN) of the port on the tape or media
changer device.
Actions Displays actions you can perform on the device. The following action is available.
u
156
NDMP backup
Port Specifies the name of the Backup Accelerator node, and the number of the port. Topology Indicates the type of Fibre Channel topology that the port is configured to
support. The following settings might appear:
u
Point to Point Indicates that the port is configured to support a point-to-point topology, with one backup device or Fibre Channel switch directly connected to the port. Loop Indicates that the port is configured to support an arbitrated loop topology, with
multiple backup devices connected to a single port in a circular formation.
Auto Indicates that the port is configured to detect the topology automatically. This is
the recommended setting. If you are using a fabric topology, specify this setting.
WWNN Indicates the world wide node name (WWNN) of the port. This name is the same for
each port on a given node.
WWPN Indicates the world wide port name (WWPN) of the port. This name is unique to the
port.
Rate Indicates the rate at which data is sent through the port. Valid values are 1 Gb/s, 2
Gb/s, 4 Gb/s, and Auto. If set to Auto, the port automatically detects the rate of data.
Actions Displays actions that you can perform on the port. The following actions might be
available:
u u
Enable Enables this port. If Enable appears, the port is currently disabled. Disable Disables this port. If Disable appears, the port is currently enabled.
NDMP backup
Session The unique identification number that OneFS assigned to the session. Elapsed Specifies how much time has elapsed since the session started. Transferred Specifies the amount of data that has been transferred during the session. Throughput Specifies the average throughput of the session over the past five minutes. Client/Remote Specifies the IP address of the backup server that the data management
application (DMA) is running on. If a three-way NDMP backup or restore operation is currently running, the IP address of the remote tape media server also appears.
Mover/Data Describes the current state of the data mover and the data server. The first
word describes the activity of the data mover. The second word describes the activity of the data server. The data mover and data server send data to and receive data from each other during backup and restore operations. The data mover is a component of the backup server that receives data during backups and sends data during restore operations. The data server is a component of OneFS that sends data during backups and receives information during restore operations. The following states might appear:
u u
Active The data mover or data server is currently sending or receiving data. Paused The data mover is in the process of receiving data during a backup operation,
but is temporarily unable to continue. While the data mover is paused, the data server cannot send data to the data mover. The data server cannot be paused.
u u
Idle The data mover or data server is not sending or receiving data. Listen The data mover or data server is ready to send or receive data, but is waiting to connect to the data server or data mover.
Backup Indicates that data is currently being backed up to a media server. The level of NDMP backup (0-9). Restore Indicates that data is currently being restored from a media server.
Source/Destination If an operation is currently in progress, specifies the /ifs directories that are affected by the operation. If a backup is in progress, the path of the source directory that is being backed up is displayed. If a restore operation is in progress, the path of the directory that is being restored is displayed along with the destination directory to which
158
NDMP backup
the tape media server is restoring data. If you are restoring data to the same location that you backed up your data from, the same path appears twice.
u
Device Specifies the name of the tape or media changer device that is communicating
with the cluster.
Mode Indicates how OneFS is interacting with data on the backup media server. OneFS interacts with data in one of the following ways:
u u u
Read/Write OneFS is reading and writing data during a backup operation. Read OneFS is reading data during a restore operation. Raw Indicates that the DMA has access to tape drives, but the drives are not
necessarily attached to tape media.
Actions Displays session-related actions that you can perform on the session. You can
perform the following action:
u
0 Performs a full NDMP backup. 1 - 9 Performs an incremental backup at the level specified.
The default value is 0. BASE_DATE If this variable is specified, a token-based incremental backup is performed. Also, if this variable is specified, the dump dates file will not be updated, regardless of the setting of the UPDATE variable. UPDATE {Y | N}
View NDMP backup logs
159
NDMP backup
Determines whether OneFS updates the dump dates file. The following values are valid:
u u
Y OneFS updates the dump dates file. N OneFS updates the dump dates file.
TYPE <backup-format> Specifies the format for the backup. The following values are valid:
u u
tar Backups are in TAR format. dump Backups are in dump format. If you specify dump, the backup will still be stored on the backup device in TAR format.
The default value is tar. HIST <file-history-format> Specifies the file history format. The following values are valid:
u u u u
D Specifies dir/node file history. F Specifies path-based file history. Y Specifies the default file history format determined by your DMA. N Disables file history.
The default value is Y. DIRECT {Y | N} Enables or disables Direct Access Restore (DAR) and Directory DAR (DDAR). The following values are valid:
u u
The default value is Y. FILES <file-matching-pattern> If this option is specified, OneFS includes only the files and directories that meet the specified pattern in backup operations. Separate multiple patterns with a space. There is no default value. EXCLUDE <file-matching-pattern> If this option is specified, OneFS does not back up files that meet the specified pattern. Separate multiple patterns with a space. There is no default value. ENCODING <encoding-type> Encodes file-selection or file-history information according to the specified encoding type. The following values are valid:
u u u u
160
NDMP backup
u u u u u u u u u u u u u u u u u u
EUC_KR CP932 CP949 CP1252 ISO_8859_1 ISO_8859_2 ISO_8859_3 ISO_8859_4 ISO_8859_5 ISO_8859_6 ISO_8859_7 ISO_8859_8 ISO_8859_9 ISO_8859_10 ISO_8859_13 ISO_8859_14 ISO_8859_15 ISO_8859_16
The default value is UTF8. RESTORE_HARDLINK_BY_TABLE {Y | N} Determines whether OneFS recovers hard links by building a hard-link table during restore operations. Specify this option if hard links have been incorrectly backed up, and restore operations are failing. If a restore operation fails because hard links have been incorrectly backed up, the following message appears in the NDMP backup logs:
Bad hardlink path for <path>
FH_REPORT_FULL_DIRENTS {Y | N} If you are using node-based file history, specifying Y causes entries for backed up directories up to be reported. You must enable this option if you are performing an incremental backup with node-based file history through NDMP v4.
161
NDMP backup
162
OneFS enables you to prevent users from modifying and deleting files on an Isilon cluster through the SmartLock tool. To use the SmartLock tool, you must configure a SmartLock license on a cluster. With the SmartLock tool, you can create SmartLock directories. Within SmartLock directories, you can commit files to a write once read many (WORM) state. A file committed to a WORM state is non-erasable and non-rewritable. The file can never be modified, and cannot be deleted until after a specified retention period has passed.
u u u u u u u u u
SmartLock operation modes................................................................................164 Replication and backup with SmartLock..............................................................165 SmartLock license functionality...........................................................................166 SmartLock best practices and considerations......................................................166 Set the compliance clock....................................................................................167 View the compliance clock..................................................................................168 Creating a SmartLock directory............................................................................168 Managing SmartLock directories.........................................................................170 Managing files in SmartLock directories..............................................................171
163
Enterprise mode
You can use SmartLock enterprise mode if you want to protect files from accidental modification or deletion, but are not required by law to do so. SmartLock enterprise mode is the default SmartLock operation mode. If a file is committed to a WORM state in a SmartLock enterprise directory, the file can be deleted by the root user through the privileged delete feature. SmartLock enterprise directories reference the system clock to facilitate time-dependent operations, including file retention. Before you can create SmartLock enterprise directories, you must configure the SmartLock enterprise license. Isilon clusters operating in SmartLock enterprise mode cannot be made compliant with the regulations defined by U.S. Securities and Exchange Commission rule 17a-4.
Compliance mode
SmartLock compliance mode enables you to protect your data in compliance with the regulations defined by U.S. Securities and Exchange Commission rule 17a-4. If you set a cluster to SmartLock compliance mode, you will not be able to log in to that cluster through the root user account. Instead, you can log in to the cluster through the compliance administrator account. You must configure the compliance administrator account during the initial cluster configuration process. If you are logged in through the compliance administrator account, you can perform administrative tasks through the sudo command. In SmartLock compliance mode, you can create SmartLock compliance directories. In a SmartLock enterprise directory, a file can be committed to a WORM state either manually or automatically by the system. A file that has been committed to a WORM state in a compliance directory cannot be modified or deleted before the specified retention period has expired. You cannot delete committed files, even if you are logged in to the compliance administrator account. The privileged delete feature is not available in SmartLock compliance mode. Before you can create SmartLock compliance directories, you must set the SmartLock compliance clock. SmartLock compliance directories reference the SmartLock compliance clock to facilitate time-dependent operations, including file retention. You can set the compliance clock only once. After you set the compliance clock, you cannot modify the compliance-clock time. In addition to creating SmartLock compliance directories, you can also create SmartLock enterprise directories on SmartLock compliance clusters.
164
SmartLock compliance mode is not compatible with Isilon for vCenter, VMware vSphere API for Storage Awareness (VASA), or the vSphere API for Array Integration (VAAI) NAS Plug-In for Isilon.
the target cluster. However, if you replicate data to a non-SmartLock directory, all metadata relating to the retention date and commit status will be lost. If you backup data to an NDMP device, all SmartLock metadata relating to the retention date and commit status will be transferred to the NDMP device. When you restore the data on an Isilon cluster, if the directory that you restore to is not a SmartLock directory, the metadata will be lost. However, if you restore to a SmartLock directory, the metadata will persist on the cluster. You can restore to a SmartLock directory only if the directory is empty.
You cannot move or rename a directory that contains a SmartLock directory. A SmartLock directory can be renamed only if the directory is empty. You cannot move a file that has been committed to a WORM state, even after the retention period for the file has expired. SmartLock compliance directories reference the compliance clock, which is controlled by the compliance clock daemon. Because a user can disable the compliance clock daemon, it is possible to increase the retention period of WORM committed files in SmartLock compliance mode. However, it is not possible to decrease the retention period of a WORM committed file. It is recommended that you create files outside of SmartLock directories and then transfer them into a SmartLock directory after you are finished working with the files. If you are uploading files to a cluster, it is recommended that you upload the files to a non-SmartLock directory, and then later transfer the files to a SmartLock directory. If a file is committed to a WORM state while it is being uploaded, the file will become trapped in an inconsistent state. Files can be committed to a WORM state even if they are not closed. If you specify an autocommit time period for a directory, the autocommit time period is calculated according to the length of time since you last modified the file, not when the file was last closed. If you do not close a file, and then delay writing to the file for more than the autocommit time period, the file will be committed to a WORM state the next time you attempt to write to it.
In order to commit a file to a WORM state, you must remove all existing read-write permissions from the file. However, if the file is already in a read-only state, attempting to remove the read-write permissions from that file will not cause the file
166
to be committed to a WORM state. You must successfully remove at least one readwrite permission in order for the file to be committed. In order to commit a file that is currently set to a read-only state, you must first enable write permissions for the file, and then remove read-write permissions for that file.
u
If the autocommit time period expires for a file, and the file is accessed by a user, the file is committed to a WORM state. However, the read-write permissions of the file are not modified. The file is still committed to a WORM state, and can never be modified and cannot be deleted until the specified retention period expires. However, the WORM state is not indicated by the read-write permissions. If you are replicating SmartLock directories to another Isilon cluster, it is recommended that you do not enable autocommit for the target directories. If you enable autocommit on target directories, it is possible that files will become committed on the target before they are committed on the source. If this happens, replication jobs will fail, and OneFS will not be able to replicate data to the target cluster. If you run the touch command on a file in a SmartLock directory without specifying a date to release the file from a SmartLock state, and you commit the file, the retention period is automatically set to the minimum retention period specified for the SmartLock directory. If you have not specified a minimum retention period for the SmartLock directory, the file is assigned a retention period of zero seconds. It is recommended that you specify a minimum retention period for all SmartLock directories. It is recommended that you set SmartLock configuration settings only once and do not modify the settings after files have been added to the SmartLock directory. If an autocommit time period is specified for the directory, modifying SmartLock configuration settings can affect the retention period of files, even if the autocommit time period of the files has already expired.
167
Retention periods
When a file is committed to a WORM state in a SmartLock directory, that file is retained for a specified retention period. If you manually commit a file by removing the read-write privileges of the file, you can optionally specify an expiration date that the retention period of a files set to expire on. You can configure minimum and maximum retention periods for a SmartLock directory that prevent files from being retained for too long or short a time period. For example, assume that you have a SmartLock cluster with a minimum retention period of two days. At 1:00 PM on Monday, you commit the file to a WORM state, and specify the file to expire on Tuesday at 3:00 PM. The retention period still expires two days later on Wednesday at 1:00 PM. You can also configure a default retention period that is assigned when a client manually commits a file, but does not specify an expiration date. For example, assume that you have a SmartLock cluster with a default retention period of two days. At 1:00 PM on Monday, you commit the file to a WORM state. The retention period expires two days later on Wednesday at 1:00 PM.
decrease the autocommit time period to two hours, the file is set to be removed from a WORM state at 5:00 PM on Tuesday, instead of 3:00 PM on Wednesday. Modifying the minimum, maximum, or default retention period of a SmartLock directory can modify the retention period of files, even after the autocommit time period of a file expires. For example, assume that you have a SmartLock directory with an autocommit time period of two days, and a default expiration period of one day. You then copy a file into the SmartLock directory on Monday, at 3:00 PM. If, at 4:00 PM on Wednesday, the file was not accessed by a user, and you decrease the default retention period to two hours, the file is set to be removed from a WORM state at 5:00 PM on Tuesday, instead of 3:00 PM on Wednesday. If you specify an autocommit time period along with a minimum, maximum, or default retention period, the retention period is calculated according to the time that the autocommit period expires. For example, assume that you have a SmartLock cluster with a minimum retention period of two days and an autocommit time period of one day. At 1:00 PM on Monday, you modify a file; then, at 1:00 PM on Tuesday, you access a file, causing the file to be committed to a WORM state. The retention period expires on Thursday at 1:00 PM, two days after the autocommit time period for the file expired.
You cannot create a SmartLock directory as a subdirectory of an existing SmartLock directory. Hard links cannot cross SmartLock directory boundaries. Creating a SmartLock directory causes a corresponding SmartLock domain to be created for that directory.
worm mkdir command to create a SmartLock directory. worm mkdir command cannot be the path of an existing
u u
1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Run the isi The path specified in the isi directory.
For example, the following command creates a compliance directory with a default retention period of four years, a minimum retention period of three years, and an maximum retention period of five years:
sudo isi worm mkdir --path /ifs/data/dir --compliance --default 4y --min 3y --max 5y
For example, the following command creates an enterprise directory with an autocommit time period of thirty minutes and a minimum retention period of three months:
isi worm mkdir --path /ifs/data/dir --autocommit 30n --min 3m
169
ID The numerical ID of the corresponding SmartLock domain. Root path The path of the directory. Type The type of directory. Enterprise directories display SmartLock. Compliance directories display Compliance. Override date The override retention date for the directory. Files committed to a WORM state are not released from a WORM state until after the specified date, regardless of the maximum retention period for this directory or whether a user specifies a retention period expiration date. Default retention period The default retention period for the directory. If a retention
period expiration date is not explicitly assigned by a user, the default retention period is assigned to the file when it is committed to a WORM state. Times are expressed in the format "<integer><time>", where <time> is one of the following values:
u u u u
170
Minimum retention period The minimum retention period for the directory. Files are
retained in a WORM state for at least the specified amount of time, even if a user specifies a retention period expiration date that equates to a shorter period of time. Times are expressed in the format "<integer><time>", where <time> is one of the following values:
u u u u
Maximum retention period The maximum retention period for the directory. Files are
retained in a WORM state for longer than the specified amount of time, even if a user specifies a retention period expiration date that equates to a longer period of time. Times are expressed in the format "<integer><time>", where <time> is one of the following values:
u u u u
Autocommit period The autocommit time period for the directory. After a file exists in this SmartLock directory without being modified for the specified time period, the file is automatically committed the next time the file is accessed by a user. Times are expressed in the format "<integer><time>", where <time> is one of the following values:
u u u u u u
y Specifies years m Specifies months w Specifies weeks d Specifies days h Specifies hours n Specifies minutes
Privileged delete The state of the privileged delete functionality for the directory.
The following values are valid:
u
On A root user can delete files committed to a WORM state by running the isi worm
filedelete command.
Off WORM committed files cannot be deleted, even through the isi worm filedelete command. Disabled (Permanently) WORM committed files cannot be deleted, even through
the isi worm filedelete command. After this setting is set for a SmartLock directory, the setting cannot be modified.
you are accessing the cluster through the root user, you can delete files that are committed to a WORM state. If you need to retain all currently committed files until a specified date, you can override the retention date for all files in a SmartLock directory. An override retention date extends the retention period of all files scheduled to expire earlier than the specified date. However, it does not decrease the retention period of files that are scheduled to expire later than the specified date. The retention period expiration date is set by modifying the access time of a file. In the UNIX command line, the access time can be modified through the touch command. Although there is no method of modifying the access time through Windows explorer, the access time can be modified through Windows Powershell. Accessing a file does not set the retention period expiration date.
3. Specify the name of the file you want to set a retention period for by creating an object. The file must exist in a SmartLock directory. For example, the following command creates an object for /smartlock/file.txt:
$file = Get-Item "\\cluster.ip.address.com\ifs\smartlock\file.txt"
4. Specify the retention period by setting the last access time for the file. For example, the following command sets an expiration date of July 1, 2012 at 1:00 PM:
$file.LastAccessTime = Get-Date "2012/7/1 1:00 pm"
172
until the retention period expires. Additionally, you cannot change the path of a file that is committed to a WORM state. To commit a file to a WORM state, you must remove all write privileges from the file. However, if you attempt to remove write privileges of a file that is already set to a readonly state, the file is not committed to a WORM state. In that case, you must add write privileges to the file, and then return the file to a read-only state. 1. Open a connection to the cluster through a UNIX command line and log in. 2. Remove write privileges from a file by running the chmod command. For example, the following command removes write privileges of /ifs/data/ smartlock/file.txt:
chmod ugo-w /ifs/data/smartlock/file.txt
173
3. Delete the WORM committed file by running the isi worm filedelete command. For example, the following command deletes /ifs/worm/enterprise/file:
isi worm filedelete /ifs/worm/enterprise/file
174
Protection domains are markers that OneFS uses to prevent modifications to files and directories. If a domain is applied to a directory, the domain is also applied to all of the files and subdirectories under the directory. You can specify domains manually; however, domains are usually created automatically by OneFS. There are three types of domains: SyncIQ, SmartLock, and SnapRevert. SyncIQ domains can be assigned to source and target directories of replication policies. OneFS automatically creates SyncIQ domains for target directories of replication policies the first time that replication policies are run. OneFS also automatically creates SyncIQ domains for source directories of replication policies during the failback process. You can manually create SyncIQ domains for source directories before you initiate the failback process, but you cannot delete SyncIQ domains that mark target directories of replication policies. SmartLock domains are assigned to SmartLock directories to prevent committed files from being modified or deleted. SmartLock domains are automatically created when a SmartLock directory is created. You cannot delete SmartLock domains. However, if you delete a SmartLock directory, OneFS automatically deletes the SmartLock domain associated with the directory. SnapRevert domains are assigned to directories that are contained in snapshots to prevent files and directories from being modified while a snapshot is being reverted. SnapRevert domains are not created automatically by OneFS. You cannot revert a snapshot until you create a SnapRevert domain for the directory that the snapshot contains. You can create SnapRevert domains for subdirectories of directories that already have SnapRevert domains. For example, you could create SnapRevert domains for both /ifs/data and /ifs/data/archive. A SnapRevert domain can be deleted if you no longer want to revert snapshots of a directory.
u u u u u
Protection domain considerations.......................................................................176 Create a protection domain.................................................................................176 Delete a protection domain.................................................................................176 View protection domains.....................................................................................177 Protection domain types.....................................................................................177
Protection domains
175
Protection domains
Copying a large number of files into a protection domain might take a very long time, because each file must be marked individually as belonging to the protection domain. You cannot move directories in or out of protection domains. However, you can move a directory contained in a protection domain to another location within the same protection domain. Creating a protection domain for a directory that contains a large number of files will take more time than creating a protection domain for a directory with fewer files. Because of this, it is recommended that you create protection domains for directories while the directories are empty, and then add files to the directory. If a domain is currently preventing the modification or deletion of a file, you cannot create a protection domain for a directory that contains that file. For example, if / ifs/data/smartlock/file.txt is set to a WORM state by a SmartLock domain, you cannot create a SnapRevert domain for /ifs/data/.
Protection domains
4. Optional: To specify a priority for the job, from the Priority list, select a priority. Lower values indicate a higher priority. If you do not specify a priority, the job is assigned the default domain mark priority. 5. Optional: To specify the amount of cluster resources the job is allowed to consume, from the Impact policy list, select an impact policy. If you do not specify a policy, the job is assigned the default domain mark policy. 6. From the Domain type list, select the type of domain you want to delete. 7. Select the Delete domain check box. 8. In the Domain root path field, type the path of the directory that is associated with the domain that you want to delete. 9. Click Start.
SmartLock SmartLock domain of an enterprise directory. Compliance SmartLock domain of a compliance directory. SyncIQ SyncIQ domain that prevents users from modifying files and directories. SyncIQ, Writable SyncIQ domain that allows users to modify files and directories. SnapRevert SnapRevert domain that prevents users from modifying files and directories while a snapshot is being reverted. Writable, SnapRevert SnapRevert domain that allows users to modify files and
directories.
If Incomplete is appended to a domain, OneFS is in the process of creating the domain. An incomplete domain does not prevent files from being modified or deleted.
177
Protection domains
178
The OneFS cluster can be managed through both the web administration interface and the command-line interface. General cluster settings can be configured and module licenses can be managed through either the command-line or web administration interface. Using either the web administration or command-line interface, you can view cluster status details for node pools, tiers, and file pool policies. Tiers and file pool policies can be managed, and you can configure events to generate email notifications and SNMP traps. OneFS cluster settings and module licenses can be configured and managed, and you can use the web administration interface to configure and graph real-time and historical cluster performance.
u u u u u u u u u u u u u u
User interfaces....................................................................................................180 Connecting to the cluster.....................................................................................181 Licensing............................................................................................................182 General cluster settings.......................................................................................184 Cluster statistics.................................................................................................189 Performance monitoring......................................................................................189 Cluster monitoring...............................................................................................189 Monitoring cluster hardware................................................................................196 Cluster maintenance...........................................................................................199 Remote support using SupportIQ.........................................................................201 Upgrading OneFS................................................................................................203 Cluster join modes..............................................................................................204 Event notification settings...................................................................................204 System job management.....................................................................................205
Cluster administration
179
Cluster administration
User interfaces
Depending on your preference, location, or the task at hand, OneFS provides four different interfaces for managing the cluster.
u
Web administration interface The browser-based OneFS web administration interface provides secure access with OneFS-supported browsers. You can use this interface to view robust graphical monitoring displays and to perform cluster-management tasks. Command-line interface Cluster-administration tasks can be performed using the
command-line interface (CLI). Although most tasks can be performed from either the CLI or the web administration interface, a few tasks can be accomplished only from the CLI.
u u
Node front panel You can monitor node and cluster details from a node front panel. OneFS Platform API OneFS includes a RESTful services application programmatic
interface (API). Through this interface, cluster administrators can develop clients and software to automate the management and monitoring of their Isilon storage systems.
Command-line interface
The OneFS command-line interface can be used to manage the OneFS cluster. The FreeBSD-based interface provides an extended standard UNIX command set for managing all aspects of the cluster. With the OneFS command-line interface, you can run OneFS isi commands to configure, monitor, and manage the cluster. You can access the command-line interface by opening a secure shell (SSH) connection or serial connection to any node in the cluster.
Node status Events Cluster details, capacity, IP and MAC addresses Throughput Drive status
180
Cluster administration
Consult the node-specific installation documentation for a complete list of monitoring activities that can be performed by using the node front panel.
Cluster administration
2. Optional: In the Shut Down or Reboot This Cluster area, select the action that you want to perform. l To shut down the cluster, click Shut down and then click Submit.
l
To stop the cluster and then restart it, click Reboot and then click Submit.
Licensing
Advanced cluster features are available when you license OneFS software modules. Each optional OneFS software module requires a separate license. For more information about the following optional software modules, contact your EMC Isilon sales representative.
u u u u u u u u u u
HDFS InsightIQ Isilon for vCenter SmartConnect Advanced SmartLock SmartPools SmartQuotas SnapshotIQ SyncIQ iSCSI
Activating licenses
Optional OneFS modules, which provide advanced cluster features, are activated by a valid license key. To activate a licensed OneFS module, you must obtain a license key and then enter the key through either the OneFS web administration interface or the command-line interface. To obtain a module license, contact your EMC Isilon Storage Division sales representative.
Cluster administration
Results You can manage the added features in both the OneFS command-line and web administration interfaces, excluding the few features that appear in only one interface.
Unconfiguring licenses
You can unconfigure a OneFS licensed module, but removing a licensed feature may have system-wide implications. You may want to unconfigure a license for a OneFS software module if, for example, you enabled an evaluation version of a module but later decided not to purchase a permanent license. Unconfiguring a module license disables recurring jobs or scheduled operations for the module, but it does not deactivate the license. You can unconfigure module licenses only through the command-line interface. You cannot unconfigure a module license through the web administration interface. The results of unconfiguring a license are different for each module.
u u u
HDFS No system impact. InsightIQ No system impact. Isilon for vCenter If you unconfigure this license, you cannot manage vSphere
machine backup and restore operations.
SmartPools If you unconfigure a SmartPools license, all file pool policies (except the
default file pool policy) are deleted.
183
Cluster administration
SmartLock If you unconfigure a SmartLock license, you cannot create new SmartLock
directories or modify SmartLock directory configuration settings for existing directories. You can commit files to a write once read many (WORM) state, even after the SmartLock license is unconfigured, but you cannot delete WORM-committed files from enterprise directories.
u u
SyncIQ If you unconfigure a SyncIQ license, the system disables SyncIQ policies and jobs. iSCSI If you unconfigure an iSCSI license, iSCSI initiators can no longer establish iSCSI connections to the cluster.
Unconfigure a license
You can unconfigure a licensed module only through the command-line interface. You must have root user privileges on your OneFS cluster to unconfigure a module license. Unconfiguring a license does not deactivate the license. 1. Open a secure shell (SSH) connection with any node in the cluster. 2. At the OneFS command prompt, log in to the cluster as root. 3. At the command prompt, run the following command, replacing <module name> with the name of the module: isi license unconfigure -m <module name> If you do not know the module name, run the command isi prompt for a list of OneFS modules and their status.
license at the command
OnesFS returns a confirmation message, similar to the following text: The <module name> module has been unconfigured. The license is unconfigured, and any processes enabled for the module are disabled.
Cluster name Cluster date and time, NTP settings Character encoding Email settings SNMP monitoring SupportIQ settings
Cluster administration
and time reported by the cluster can be set manually. The NTP service can be configured to ensure that all nodes in a cluster are synchronized to the same time source. Windows domains provide a mechanism to synchronize members of the domain to a master clock running on the domain controllers. OneFS uses a service to adjust the cluster time to that of Active Directory. Whenever a cluster is joined to an Active Directory domain and an external NTP server is not configured, the cluster is set automatically to Active Directory time, which is synchronized by a job that runs every 6 hours. When the cluster and domain time become out sync by more than 4 minutes, OneFS generates an event notification. If the cluster and Active Directory become out of sync by more than 5 minutes, authentication will not work. To summarize:
u
If no NTP server is configured but the cluster is joined to an Active Directory domain, the cluster synchronizes with Active Directory every 6 hours. If an NTP server is configured, the cluster synchronizes the time with the NTP server.
185
Cluster administration
Cluster administration
Configuring SupportIQ
OneFS logs contain data that Isilon Technical Support personnel can securely upload, with your permission, and then analyze to troubleshoot cluster problems. The SupportIQ technology must be enabled and configured. When SupportIQ is enabled, Isilon Technical Support personnel can request logs via scripts that gather cluster data and then upload the data to a secure location. You must enable and configure the SupportIQ module before SupportIQ can run scripts to gather data. The feature may have been enabled when the cluster was first set up. As an option, you can also enable remote access, which allows Isilon Technical Support personnel to troubleshoot your cluster remotely and run additional data-gathering scripts. Remote access is disabled by default. To enable Isilon to remotely access your cluster using SSH, you must provide the cluster password to a Technical Support engineer.
Send alerts via SupportIQ agent (HTTPS) SupportIQ delivers notifications to Isilon only through the SupportIQ agent over HTTPS.
4. Optional: Enable HTTPS proxy support for SupportIQ. a. Select the HTTPS proxy for SupportIQ check box. b. In the Proxy host field, type the IP address or fully qualified domain name (FQDN) of the HTTP proxy server. c. In the Proxy port field, type the number of the port on which the HTTP proxy server receives requests. d. Optional: In the Username field, type the user name for the proxy server. e. Optional: In the Password field, type the password for the proxy server. 5. Optional: Enable remote access to the cluster. a. Select the Enable remote access to cluster via SSH and web interface check box. The remote-access end user license agreement (EULA) appears. b. Review the EULA and, if you agree to the terms and conditions, select the I have read and agree to... check box. 6. Click Submit. A successful configuration is indicated by a message similar to SupportIQ settings have been updated.
Configuring SupportIQ
187
Cluster administration
Disable SupportIQ
You can disable SupportIQ, so that the SupportIQ agent does not run scripts to gather and upload data about your OneFS cluster. 1. Click Cluster Management > General Settings > SupportIQ. 2. Clear the Enable SupportIQ check box. 3. Click Submit. The SupportIQ agent is deactivated.
3. Click Submit.
3. Click Submit.
188
Cluster administration
Character encoding is typically established during installation of the cluster. Modifying the character encoding after installation may render files unreadable if done incorrectly. Modify settings only if necessary after consultation with Isilon Technical Support. 1. Click File System Management > File System Settings > Character Encoding. 2. Optional: From the Character encoding list, select the character-encoding set that you want to use. 3. Click Submit, and then click Yes to acknowledge that the encoding change becomes effective after the cluster is restarted. 4. Restart the cluster. Results After the cluster restarts, the web administration interface reflects your change. What to do next
Cluster statistics
Command-line options provide the ability to view performance, historical, and in-depth usage statistics for your cluster. The isi statistics and isi status command-line tools include options for querying and filtering the display of OneFS cluster performance and usage statistics. You can use generic and type-specific options to control filtering, aggregation, and reporting for each mode of statistics reporting. You can access these modes of operation by typing subcommands in the isi statistics and isi status tools. For more information about the statistics and status options and descriptions of the subcommands, see the OneFS 7.0 Command Reference.
Performance monitoring
Cluster throughput can be monitored through either the web administration interface or the command-line interface. You can view cluster throughput graphically and numerically for average and maximum usage. Performance can be monitored through the web administration interface or the command-line interface by using the isi statistics command options. You can view details about the input and output traffic to and from the cluster's file system and also monitor throughput distribution across the cluster. Advanced performance monitoring and analytics are available through the InsightIQ module, which requires a separate license. For more information about optional software modules, contact your EMC Isilon Storage Division sales representative.
Cluster monitoring
Cluster health, performance, and status can be monitored from both the web administration interface and from the command-line interface. In addition, real-time and historical performance can be graphed in the web administration interface. The condition and status of the OneFS hardware can be monitored in the OneFS dashboard through the web administration interface. You can monitor information about the health and performance of the cluster, including the following.
Cluster statistics
189
Cluster administration
Node status Health and performance statistics for each node in the cluster, including
hard disk drive (HDD) and solid-state drive (SSD) usage.
u u
Client connections Number of clients connected per node. New events List of event notifications generated by system events, including the severity,
unique instance ID, start time, alert message, and scope of the event.
Cluster size Current view: Used and available HDD and SSD space and space reserved for the virtual hot spare (VHS). Historical view: Total used space and cluster size for a one-year
period.
Cluster throughput (file system) Current view: Average inbound and outbound traffic volume passing through the nodes in the cluster for the past hour. Historical view: Average inbound and outbound traffic volume passing through the nodes in the cluster for the past two weeks. CPU usage Current view: Average system, user, and total percentages of CPU usage for the past hour. Historical view: displays CPU usage for the past two weeks.
Information is accessible for individual nodes, including node-specific network traffic, internal and external network interfaces, and details about node pools, tiers and overall cluster health. Using the OneFS dashboard, you can monitor the status and health of the oneFS system hardware. In addition, SNMP can be used to remotely monitor hardware components, such as fans, hardware sensors, power supplies, and disks.
Client connection summary: To view a list of current connections, click Dashboard > Cluster Overview > Client Connections Status. New events: To view more information about an event, click View details in the Actions column. Cluster size: To switch between current and historical views, click Historical or Current near the Monitoring section heading. In historical view, click Used or Cluster size to change the display. Cluster throughput (file system): To switch between current and historical views, click Historical or Current next to the Monitoring section heading. To view throughput statistics for a specific period within the past two weeks, click Dashboard > Cluster Overview > Throughput Distribution. You can hide or show inbound or outbound throughput by clicking Inbound or Outbound in the chart legend. To view maximum throughput, next to Show, select Maximum.
CPU usage: To switch between current and historical views, click Historical or Current near the Monitoring section heading.
190
Cluster administration
You can hide or show a plot by clicking System, User, or Total in the chart legend. To view maximum usage, next to Show, select Maximum.
Client connections: To view current clients connected to this node, review the list in this area. Chassis and drive status: To view the state of drives in this node, review this area. To view details about a drive, click the name link of the drive; for example, Bay1. Node size: To switch between current and historical views, click Historical or Current next to the Monitoring area heading. In historical view, click Used or Cluster size to change the display accordingly. Node throughput (file system): To switch between current and historical views, click Historical or Current next to the Monitoring area heading. To view throughput statistics for a period within the past two weeks, click Dashboard > Cluster Overview > Throughput Distribution. You can hide or show inbound or outbound throughput by clicking Inbound or Outbound in the chart legend. To view maximum throughput, next to Show, select Maximum.
CPU usage: To switch between current and historical views, click Historical or Current next to the Monitoring area heading. You can hide or show a plot by clicking System, User, or Total in the chart legend. To view maximum usage, next to Show, select Maximum.
Cluster administration
methods to notify recipients: Email, SupoprtIQ, and SNMP trap. Each event notification method can be configured through the web administration interface or the command-line interface.
u
Email If you configure email event notifications, you designate recipients and specify
SMTP, authorization, and security settings. You can specify batch email settings and the email notification template that you want to use when email notifications are sent.
SupportIQ If you enable SupportIQ, you can specify the protocol that you prefer to use
for notifications: HTTPS, SMTP, or both.
SNMP trap If you configure the OneFS cluster for SNMP monitoring, you select events to send SNMP traps to one or more network monitoring stations, or trap receivers. Each event can generate one or more SNMP traps. The ISILON-TRAP-MIB describes the traps that the cluster can generate, and the ISILON-MIB describes the associated varbinds that accompany the traps. You can download both management information base files (MIBs) from the cluster. You must configure an event notification rule to generate SNMP traps.
Coalesced events
Multiple related or duplicate event occurrences are grouped, or coalesced, into one logical event by the OneFS system. For example, if the CPU fan crosses the speed threshold more than 10 times in an hour, the system coalesces this sequence of identical but discrete occurrences into one event. You can view coalesced events and details through the web administration interface or the command-line interface. This message is representative of coalesced event output.
# isi events show 24.924 ID: 24.924 Type: 199990001 Severity: critical Value: 0.0 Message: Disk Errors detected (Bay 1) Node: 21 Lifetime: Sun Jun 17 23:29:29 2012 - Now Quieted: Not quieted Specifiers: disk: 35 val: 0.0 devid: 24 drive_serial: 'XXXXXXXXXXXXX' lba: 1953520064L lnn: 21 drive_type: 'HDD' device: 'da1' bay: 1 unit: 805306368 Coalesced by: -Coalescer Type: Group Coalesced events: ID STARTED ENDED SEV LNN MESSAGE 24.911 06/17 23:29 -- I 21 Disk stall: Bay 1, Type Disk ... 24.912 06/17 23:29 -- I 21 Sector error: da1 block 24.913 06/17 23:29 -- I 21 Sector error: da1 block 24.914 06/17 23:29 -- I 21 Sector error: da1 block 24.915 06/17 23:29 -- I 21 Sector error: da1 block 24.916 06/17 23:29 -- I 21 Sector error: da1 block 24.917 06/17 23:29 -- I 21 Sector error: da1 block 24.918 06/17 23:29 -- I 21 Sector error: da1 block 24.919 06/17 23:29 -- I 21 Sector error: da1 block 24.920 06/17 23:29 -- I 21 Sector error: da1 block 24.921 06/17 23:29 -- I 21 Sector error: da1 block
192
HDD, LNUM 35. 1953520064 2202232 2202120 2202104 2202616 2202168 2202106 2202105 1048670 223
Cluster administration
21
Quiet Acknowledges an event, which removes it from list of new events and adds it to the
list of quieted events.
If a new event of the same event type is triggered, it is a separate new event and must be quieted.
u
Unquiet Returns a quieted event to an unacknowledged state in the list of new events
and removes it from the list of quieted events.
Most events are canceled automatically by the system when they reach the end of their duration. They remain in the system until you manually acknowledge, or quiet them, however. You can acknowledge events through either the web administration interface or the command-line interface. For information about managing events through the command-line interface, see the OneFS Command Reference.
193
Cluster administration
Responding to events
You can view event details and respond to cluster events through the web administration interface or the command-line interface. In the web administration interface, you can view new events on a cluster. You can view and manage new events, open events, and recently ended events. You can also view coalesced events and additional, more detailed information about specific events, and you can quiet or cancel events.
To add the settings of this event to an existing event notification rule, in the Create Rule area, select Add to an existing notification rule, select the existing event notification rule from the list, and then click Submit.
Manage an event
You can change the status of an event by quieting, unquieting, or canceling it. 1. Click Dashboard > Events > Summary. The New Events page appears and displays a list of all new, or unquieted, events. 2. Perform the following actions as needed. l To view additional information about an event, in the Actions column for that event, click View details.
l l l
To acknowledge an event, click Quiet. To restore an event to an unacknowledged state, click Unquiet. To permanently remove an occurrence of an event, click Cancel.
194
Cluster administration
195
Cluster administration
In the Set custom notification template field, type the path and file name of the template file that you want to use.
5. Click Submit.
Cluster administration
3. In the Chassis and drive status area, click Platform. Information about the hardware appears, including status, a list of monitored components, the system partitions, and the hardware log.
SNMP monitoring
You can use SNMP to remotely monitor the OneFS cluster hardware components, such as fans, hardware sensors, power supplies, and disks. The default Linux SNMP tools or a GUI-based SNMP tool of your choice can be used for this purpose. You can enable SNMP monitoring on individual nodes on your cluster, and you can also monitor cluster information from any node. Generated SNMP traps are sent to your SNMP network. You can configure an event notification rule that specifies the network station where you want to send SNMP traps for specific events, so that when an event occurs. the cluster send the trap to that server. OneFS supports SNMP in read-only mode. SNMP v1 and v2c is the default, but you can configure settings for SNMP v3 alone or SNMP v1, v2c, and v3. When SNMP v3 is used, OneFS requires AuthNoPriv as the default. AuthPriv is not supported. Elements in an SNMP hierarchy are arranged in a tree structure, similar to a directory tree. As with directories, identifiers move from general to specific as the string progresses from left to right. Unlike a file hierarchy, however, each element is not only named, but also numbered. For example, the SNMP entity .iso.org.dod.internet.private.enterprises.isilon.oneFSss.ssLocalN odeId.0 maps to .1.3.6.1.4.1.12124.3.2.0. The part of the name that refers to the OneFS SNMP namespace is the 12124 element. Anything further to the right of that number is related to OneFS-specific monitoring. Management Information Base (MIB) documents define human-readable names for managed objects and specify their data types and other properties. You can download MIBs that are created for SNMP-monitoring of a OneFS cluster from the webadministration interface or manage them using the command-line interface. MIBs are stored in /usr/local/share/snmp/mibs/ on a OneFS node. The OneFS ISILON-MIBs serve two purposes:
u u
Augment the information available in standard MIBs Provide OneFS-specific information that is unavailable in standard MIBs ISILON-MIB Defines a group of SNMP agents that respond to queries from a network monitoring system (NMS) called OneFS Statistics Snapshot agents. As the name implies, these agents snapshot the state of the OneFS file system at the time that it receives a request and reports this information back to the NMS. ISILON-TRAP-MIB Generates SNMP traps to send to an SNMP monitoring station when
the circumstances occur that are defined in the trap protocol data units (PDUs).
ISILON-MIB is a registered enterprise MIB. OneFS clusters have two separate MIBs:
u
The OneFS MIB files map the OneFS-specific object IDs with descriptions. Download or copy MIB files to a directory where your SNMP tool can find them, such as /usr/share/ snmp/mibs/ or /usr/local/share/snmp/mibs, depending on the tool that you use. To have Net-SNMP tools read the MIBs to provide automatic name-to-OID mapping, add m All to the command, as in the following example. snmpwalk -v2c -c public -m All <node IP> isilon
SNMP monitoring
197
Cluster administration
If the MIB files are not in the default Net-SNMP MIB directory, you may need to specify the full path, as in the following example. Note that all three lines are one command.
snmpwalk -m /usr/local/share/snmp/mibs/ONEFS-MIB.txt:/usr/local/ share/snmp/mibs/ONEFS-SNAPSHOT-MIB.txt:/usr/local/share/snmp/mibs /ONEFS-TRAP-MIB.txt \ -v2c -C c -c public <node IP> enterprises.onefs
The examples are from running the snmpwalk command on a cluster. Your SNMP version may require different arguments.
6. Navigate back to the SNMP Monitoring page. 7. Configure General Settings. a. In the Settings area, configure protocol access by selecting the version that you want. OneFS does not support writable OIDs; therefore, no write-only community string setting is available. b. In the System location field, type the system name. This setting is the value that the node reports when responding to queries. Type a name that helps to identify the location of the node. c. Type the contact email address in the System contact field.
198
Cluster administration
8. Optional: If you selected SNMP v1/v2 as your protocol, in the SNMP v1/v2c Settings section, in the Read-only community field, type the community name 9. Configure SNMP v3 Settings. a. In the Read-only user field, type the SNMP v3 security name to change the name of the user with read-only privileges. The default read-only user is general. The password must contain at least eight characters and must not contain any space characters. b. in the SNMP v3 password field, type the new password for the read-only user to set a new SNMP v3 authentication password. The default password is password. c. Type the new password in the Confirm password field to confirm the new password. 10. Click Submit. Results SNMP monitoring is configured for remote monitoring of the cluster hardware components.
Cluster maintenance
Isilon nodes contain components that can be replaced or upgraded in the field by trained service personnel. Isilon Technical Support can assist you with replacing node components or upgrading components to increase performance.
Cluster administration
u u u u u u u u u u
SATA/SAS Drive memory (DIMM) fan front panel intrusion switch network interface card (NIC) IB/NVRAM card SAS controller NVRAM battery power supply
If your cluster is configured to send alerts to Isilon Technical Support, you will be contacted when a component needs to be replaced. If your cluster is not configured to send alerts to Isilon, you will need to instigate a service request on your own.
The following actions can be taken to manage the health and performance of a cluster: Add a node to the cluster. Expand a cluster by adding another node.
u
Remove a node from the cluster. Take a node out of the cluster. Shut down or restart the cluster. Shutdown or restart the cluster to perform maintenance.
Before you begin For a node to be added to a cluster, an internal IP address must be available. Before you add new nodes, add IP addresses as necessary. For information on how to add IP addresses, see "Managing the internal cluster network." 1. Navigate to Cluster Management > Hardware Configuration > Add Nodes. 2. In the Available Nodes table, click Add for the node you want to add to the cluster.
Cluster administration
2. In the Remove Node area, specify the node you want to remove. 3. Click Submit. If you remove a storage node, the Cluster Status area displays smartfail progress. If you remove a non-storage accelerator node, it is immediately removed from the cluster.
201
Cluster administration
SupportIQ scripts
When SupportIQ is enabled, Isilon Technical Support personnel can request logs with scripts that gather cluster data and then upload the data. The SupportIQ scripts are located in the /usr/local/SupportIQ/Scripts/ directory on each node. Data-gathering scripts The following table lists all of the data-gathering scripts that SupportIQ can run. These scripts can be run automatically, at the request of an Isilon Technical Support representative, to collect information about your cluster's configuration settings and operations. The SupportIQ agent then uploads the information to a secure Isilon FTP site, so that it is available for Isilon Technical Support personnel to analyze. The SupportIQ scripts do not affect cluster services or the availability of your data. Script name Clean watch folder Get application data Generate dashboard file daily Generate dashboard file sequence Get ABR data (as built record) Get ATA control and GMirror status Get cluster data Get cluster events Get cluster status Get contact info Get contents (var/crash) Get job status Get domain data Get file system data
Description Clears the contents of /var/crash. Collects and uploads information about OneFS application programs. Generates daily dashboard information. Generates dashboard information in the sequence that it occurred. Collects as-built information about hardware. Collects system output and invokes a script when it receives an event that corresponds to a predetermined eventid. Collects and uploads information about overall cluster configuration and operations. Gets the output of existing critical events and uploads the information. Collects and uploads cluster status details. Extracts contact information and uploads a text file that contains it. Uploads the contents of /var/crash. Collects and uploads details on a job that is being monitored. Collects and uploads information about the clusters Active Directory Services (ADS) domain membership. Collects and uploads information about the state and health of the OneFS /ifs/ file system. Collects and uploads information about the configuration and operation of the InfiniBand back-end network. Collects and uploads only the most recent cluster log information.
202
Cluster administration
Script name Get messages Get network data Get NFS clients Get node data Get protocol data Get Pcap client stats Get readonly status Get usage data
Description Collects and uploads active /var/log/messages files. Collects and uploads information about cluster-wide and nodespecific network configuration settings and operations. Runs a command to check if nodes are being used as NFS clients. Collects and uploads node-specific configuration, status, and operational information. Collects and uploads network status information and configuration settings for the NFS, SMB, FTP, and HTTP protocols. Collects and uploads client statistics. Warns if the chassis is open and uploads a text file of the event information. Collects and uploads current and historical information about node performance and resource usage. Collects and uploads all recent cluster log information. Collects and uploads changes to cluster log information that have occurred since the most recent full operation. Collects and uploads details for a single node. Prompts you for the node number. Collects and uploads changes to cluster log information that have occurred since the most recent full operation. Prompts you for the node number.
isi_gather_info isi_gather_info --
incremental
isi_gather_info --
node
Uploads dashboard information to the secure Isilon Technical Support FTP site.
Upgrading OneFS
Two options are available for upgrading the OneFS operating system: a rolling upgrade or a simultaneous upgrade. Before upgrading the OneFS 6.0 or 6.5.x system, a pre-upgrade check must be performed. A rolling upgrade individually upgrades and restarts each node in the cluster sequentially. During a rolling upgrade, the cluster remains online and continues serving clients with no interruption in service, although some connection resets may occur on SMB clients. Rolling upgrades are performed sequentially by node number, so a rolling upgrade takes longer to complete than a simultaneous upgrade. The final node in the upgrade process is the node that you used to start the upgrade process. Rolling upgrades are not available for all clusters. For instructions on how to upgrade the cluster operating system, see the OneFS Release Notes. A simultaneous upgrade installs the new operating system and restarts all nodes in the cluster at the same time. Simultaneous upgrades are faster than rolling upgrades but require a temporary interruption of service during the upgrade process. Your data is inaccessible during the time that it takes to complete the upgrade process.
Upgrading OneFS
203
Cluster administration
Before beginning either a simultaneous or rolling upgrade, OneFS compares the current cluster and operating system with the new version to ensure that the cluster meets certain criteria, such as configuration compatibility (SMB, LDAP, SmartPools), disk availability, and the absence of critical cluster events. If upgrading puts the cluster at risk, OneFS warns you, provides information about the risks, and prompts you to confirm whether to continue the upgrade. If the cluster does not meet the pre-upgrade criteria, the upgrade does not proceed, and the unsupported statuses are listed.
Description Configures OneFS to join new nodes to the cluster in a separate manual process, allowing the addition of a node without requiring authorization
Notes -
Secure
Requires authorization of every If you use the secure join node added to the cluster mode, you cannot use the serial console wizard option [2]
Description Generates a single email for each event notification. Generates an email that contains aggregated notifications for each event of
204
Cluster administration
Setting -
Option -
Description
the same severity, regardless of event category. Generates an email an email that contains aggregated notifications for event of the same category, regardless of severity. Generates one email per event. Sends the email notification in the default OneFS notification template format. Sends the email notifications in the format that you defined in your custom template file.
Batch by category
No batching Custom notification template No custom notification template is set Set custom notification template
Job An application built into the System Job Engine. Phase Jobs are broken down into phases. Some jobs have as little as one phase, others have as many as seven. If an error is detected while a job is running, the job will not progress past its current phase unless it is determined safe to do so. Task Phases are broken down into tasks. A phase includes at least one task, but may
employ several. A task is the actual action that is taken on the system.
Item Items are the targets of tasks, the file system components that are being operated on
by a task.
The Job Engine tracks interaction between a task and its target items. u If an error is detected, the job will continue as long as the error does not affect the overall goal of the job, otherwise the job is cancelled.
u
If the task is slowing the performance of a node, the task will be asked to slow down and consume less resources.
The System Job Engine accumulates task and item results in logs that provide administrators insight into the maintenance of a cluster.
System job management
205
Cluster administration
System jobs
OneFS offers a number of jobs to assist with cluster administration and maintenance. AutoBalance Balances free space within the cluster. AutoBalance is most efficient in clusters containing only hard disk drives (HDDs). AutoBalanceLin Balances free space within the cluster. AutoBalanceLin is most efficient in clusters where file system metadata is stored on solid state drives (SSDs). AVScan Performs an antivirus scan on all files. Collect Reclaims free space that could not be freed earlier due to a node or drive being unavailable. DomainMark Associates a path and its contents with a domain. FlexProtect Performs a protection pass on the file system. FlexProtect is most efficient in clusters containing only HDDs. FlexProtectLin Performs a protection pass on the file system. FlexProtectLin is most efficient in clusters where file system metadata is stored on SSDs. FSAnalyze Gathers file system analytics. IntegrityScan Verifies file system integrity. MediaScan Removes media-level errors from disks. MultiScan Runs the AutoBalance and Collect jobs together. PermissionRepair Corrects file and directory permissions in the /ifs directory. QuotaScan Updates quota accounting for domains created on an existing file tree. SetProtectPlus Applies a default file policy across the cluster. This is used only if SmartPools is not licensed.
206
Cluster administration
ShadowStoreDelete Creates free space associated with a shadow store. SmartPools Enforces SmartPools file policies. This is used only if SmartPools is licensed. SnapRevert Reverts an entire snapshot back to head. SnapshotDelete Creates free space that is associated with deleted snapshots. TreeDelete Deletes a file path in the /ifs directory.
Allowed to run Any time of day. Any time of day. Any time of day. Outside of business hours.
These default policies cannot be deleted or modified. However, administrators can tailor policies to a specific workflow by creating new policies. New policies are created by copying and modifying the default policies.
207
Cluster administration
Job priorities
Every job in the System Job Engine is assigned a priority. Priorities determine which job will yield when two jobs attempt to run at the same time. When jobs are scheduled to run at the same time, the higher priority job will run first. The highest priority jobs have a priority of 1. Higher priority jobs will always interrupt lower priority jobs. If a low priority job is interrupted, it will be inserted back into the priority queue. When the job reaches the front of the priority queue again, it will resume from where it left off. If two jobs at the same priority level attempt to run, the job that entered the queue first will run. The following list contains Priority values for each job. These priorities can be adjusted by a system administrator. Job Autobalance Autobalance Lin AV Scan Collect SmartPools MultiScan Flex Protect FS Analyze Integrity Scan Media Scan Repair Quota Scan Set Protect Plus Snapshot Delete Tree Delete Upgrade Priority 4 4 6 4 6 4 1 6 1 8 5 6 6 2 4 3
208
Cluster administration
u u u u
Resume a job. Continue a paused job. Cancel a job. Discontinue a job that is currently active. Retry a job. Restart a job that was previously interrupted by the system. Modify job settings. Change a job's priority level or impact policy.
Start a job
You can start a job manually at any time. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Running Jobs area, clickStart job. 3. From the Job list, select the job that you want to run. 4. From the Priority list, select the priority level for the job. If no priority is selected, the job runs at its default priority. 5. From the Impact policy list, select the impact policy for the job. If no impact policy is selected, the job runs with the default impact policy for that job type. 6. Click Start. The job displays in the Running Jobs table.
Pause a job
You can pause an in-progress job. Pausing a job allows you to temporarily free cluster resources without losing progress made by the job. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Policies table, click Pause for the job you want to pause. The job moves from the Running Jobs table to the Paused and Waiting Jobs table. 3. In the Paused and Waiting Jobs table, click Resume for the job you want to resume. The job moves from the Paused and Waiting Jobs table to the Running Jobs table.
Update a job
You can change the priority and impact policy of a running, waiting, or paused job. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Running Jobs or Paused and Waiting Jobs table, click Update for the job you want to update. 3. Adjust the priority of the job by selecting a new priority from the Priority list. 4. Adjust the default impact policy of the job by selecting a new impact policy from the Impact policy list. 5. Click Update to save the new settings. If you update a running job, the job will automatically resume. If you update a paused or waiting job, the job will return to that status. Results Only the current instance of the job will run with the updated settings. The next instance of the job will return to the default settings for that job. To permanently modify job settings, click on a job name, then click Modify job defaults in the Job Details area.
209
Cluster administration
Resume a job
You can resume a paused job. The job will continue from the phase in which it was paused. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Paused and Waiting Jobs table, click Resume for the job you want to continue. The job displays in the Running Jobs table.
Cancel a job
You can discontinue a running, paused, or waiting job. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Running Jobs table, click Cancel for the job you want to cancel. This action can also be performed on a job in the Paused and Waiting Jobs table. The job displays in the Recent Job History table as User Cancelled.
Retry a job
If a job fails, you can manually restart the job without waiting for the next scheduled run time. 1. Navigate to Cluster Management > Operations > Operations Summary. 2. In the Failed Jobs table, click Retry for the job you want to run again. The job moves from the Failed Jobs table to the Running Jobs table.
Cluster administration
Copy an impact policy. Use an existing impact policy as the foundation for a new policy.
Cluster administration
Include information specific to the impact policy such as unique schedule parameters, or logistical requirements that make the new impact policy necessary. 5. Click Submit. 6. In the Impact Schedule area, modify the schedule of the impact policy by adding, editing, or deleting impact intervals. The default impact schedule for a new policy is to run at any time with an impact setting of Low. 7. In the Policy Details area, click Submit. The new impact policy appears in the Impact Policies table, and is now available to assign to jobs.
Delete an impact policy. Remove an impact policy from the System Job Engine. Modify the impact schedule of a policy. Establish a period of time where specific impact limitations are placed on a job. For example, a job will run with low impact during business hours, but run with high impact during non-business hours. View impact policy settings. Open a full list of the available impact policies to review current settings.
212
Cluster administration
Cluster administration
c. Click Edit for the copy you want to modify. The Impact Schedule area appears at the bottom of the page. Modify an impact policy a. Navigate to Cluster Management > Operations > Jobs and Impact Policies. b. In the Impact Policies table, click Edit for the policy you want to modify. The Impact Schedule area appears at the bottom of the page.
The Impact Schedule area displays the schedule for the policy in a table. Every row in the table contains an impact interval. An impact interval is a window of time where the impact level of a job is raised or lowered, or the job is paused. 2. You can modify the impact schedule of a policy by performing the following actions: Add an impact interval a. Click Add impact interval. b. From the Impact list, select the impact level for the new interval. c. Specify the day and time to begin the impact interval. d. Specify the day and time to end the impact interval. e. Click Submit. The new impact interval appears as part of the impact schedule for the policy. Create additional impact intervals as necessary. New impact intervals will overwrite existing intervals. Modify an impact interval a. In the Impact Interval table, click Edit for the interval you want to modify. b. Adjust the impact level, start time, or end time. c. Click Submit. The modified impact interval appears as part of the impact schedule for the policy. Delete an impact interval a. In the Impact Interval table, click Delete for the interval you want to delete. b. Confirm that you would like to delete the impact interval. The impact interval is removed from the impact schedule for the policy.
214
CHAPTER 11 SmartQuotas
The SmartQuotas module is an optional quota-management tool that monitors and enforces administrator-defined storage limits. Using accounting and enforcement quota limits, reporting capabilities, and automated notifications, SmartQuotas manages storage use, monitors disk storage, and issues alerts when disk-storage limits are exceeded. Quotas help you manage storage usage according to criteria that you define. Quotas are used as a method of trackingand sometimes limitingthe amount of storage that a user, group, or project consumes. Quotas are a useful way of ensuring that a user or department does not infringe on the storage that is allocated to other users or departments. In some quota implementations, writes beyond the defined space are denied, and in other cases, a simple notification is sent. The SmartQuotas module requires a separate license. For additional information about the SmartQuotas module or to activate the module, contact your EMC Isilon sales representative.
u u u u u u u u u u u u
Quotas overview.................................................................................................216 Creating quotas...................................................................................................222 Managing quotas................................................................................................223 Managing quota notifications..............................................................................226 Managing quota reports......................................................................................229 Basic quota settings............................................................................................230 Advisory limit quota notification rules settings....................................................231 Soft limit quota notification rules settings...........................................................232 Hard limit quota notification rules settings..........................................................233 Limit notification settings....................................................................................233 Quota report settings..........................................................................................234 Custom email notification template variable descriptions...................................235
SmartQuotas
215
SmartQuotas
Quotas overview
The integrated OneFS SmartQuotas module is an optional quota-management tool that monitors and enforces administrator-defined storage limits. Through the use of accounting and enforcement quota limits, reporting capabilities, and automated notifications, you can manage storage utilization, monitor disk storage, and issue alerts when storage limits are exceeded. A storage quota defines the boundaries of storage capacity that are allowed for an entity in a OneFS cluster, such as a group, a user, or a directory. The SmartQuotas module can provision, monitor, and report disk-storage usage and can send automated notifications when storage limits are exceeded or are being approached. SmartQuotas also provides flexible reporting options that can help you analyze data usage.
Quota types
OneFS uses the concept of quota types, sometimes referred to as quota domains, as the fundamental organizational unit of storage quotas. Storage quotas comprise a set of resources and an accounting of each resource type for that set. Storage quotas creation always begins with creating one or more quota types. Every quota type is defined by a directory or an entity, which together encapsulate the files and subdirectories to be tracked. When you describe a storage quota type, three important identifiers are used: The directory that it is on u The quota entity u Whether snapshots are to be tracked against the quota limit Quota types support default user and group entities (in addition to specified users and groups) to describe quotas that have default user and group policies. You can choose a quota type from the following entities:
u u
Directory A specific directory and its subdirectories. User Either a specific user or default user (every user). Specific-user quotas that you configure take precedence over a default user quota. Group All members of a specific group or all members of a default group (every group).
Any specific-group quotas that you configure take precedence over a default group quota. Associating a group quota with a default group quota creates a linked quota.
You can create multiple quota types on the same directory, but they must be of a different type or have a different snapshot option. Quota types can be specified for any directory in OneFS and can be nested within each other, creating a hierarchy of complex storage-use policies. You should not create quotas of any type on the OneFS root (/ifs). A root-level quota may significantly degrade performance. Nested storage quotas can overlap. For example, the following quota settings ensure that the finance directory never exceeds 5 TB, while limiting the users in the finance department to 1 TB each:
u u
Set 1 TB soft quotas on each user in the finance department. A default quota type is a quota that does not account for a set of files, but instead specifies a policy for new entities that match a trigger. The default-user@/ifs/cs becomes specific-user@/ifs/cs for each specific-user that is not otherwise defined. As an example,
216
SmartQuotas
you can create a default-user quota on the /ifs/dir-1 directory, where that directory is owned by the root user. The default-user type automatically creates a new domain on that directory for root and adds the usage there:
my-OneFS-1# mkdir /ifs/dir-1 my-OneFS-1# isi quota quotas create --default-user --path=/ifs/dir-1 my-OneFS-1# isi quota quotas ls -v --path=/ifs/dir-1 Type Path Policy Snap Usage -------------------- ---------------------- ----------- ----- -------default-user /ifs/dir-1 enforcement no 0B [usage-with-no-overhead] ( 0B) [usage-with-overhead] ( 0B) [usage-inode-count] (0) * user:root /ifs/dir-1 enforcement no 0B [usage-with-no-overhead] ( 0B) [usage-with-overhead] ( 2.0K) [usage-inode-count] (1)
In this example, the default-user type created a new specific-user type automatically (user:admin) and added the new usage to it. Default-user does not have any usage because it is used only to generate new quotas automatically. Default-user enforcement is copied to a specific-user (user:admin), and the inherited quota is called a linked quota. In this way, each user account gets its own usage accounting. Defaults can overlap; for example, default-user@/ifs and default-user@/ifs/cs both may be defined. If default enforcement changes, OneFS storage quotas propagate the changes to the linked quotas asynchronously. Because the update is asynchronous, there is some lag before updates are in effect. If a default type (every user or every group) is deleted, OneFS deletes all children that are marked as inherited. As an option, you can delete the default without deleting the children, but it is important to note that this action breaks inheritance on all inherited children. Continuing with the example, add another file that is owned by the root user. Because the root type exists, the new usage is added to it.
my-OneFS-1# touch /ifs/dir-1/anotherfile my-OneFS-1# isi quota ls -v --path=/ifs/dir-1 Type Path Policy Snap Usage -------------------- ---------------------- ----------- ----- -------default-user /ifs/dir-1 enforcement no 0B [usage-with-no-overhead] ( 0B) [usage-with-overhead] ( 0B) [usage-inode-count] (0) * user:root /ifs/dir-1 enforcement no 39B [usage-with-no-overhead] ( 39B) [usage-with-overhead] ( 3.5K) [usage-inode-count] (2) * user:admin /ifs/dir-1 enforcement no 0B
Quota types
217
SmartQuotas
The enforcement on default-user is copied to the specific-user when the specific-user allocates within the type, and the new inherited quota type is also a linked quota. Configuration changes for linked quotas must be made on the parent (default) quota that the linked quota is inheriting from. Changes to the parent quota are propagated to all children. If you want to override configuration from the parent quota, you must unlink the quota first.
Accounting The accounting option tracks but does not limit disk-storage use. Using the
accounting option for a quota, you can monitor inode count and physical and logical space resources. Physical space refers to all of the space used to store files and directories, including data and metadata in the domain. Logical space refers to the sum of all files sizes, excluding file metadata and sparse regions. User data storage is tracked using logical-space calculations, which do not include protection overhead. As an example, by using the accounting option, you can do the following: u Track the amount of disk space used by various users or groups to bill each entity for only the disk space used.
u
Review and analyze reports that help you identify storage usage patterns, which you can use to define storage policies for the organization and educate users of the file system about using storage more efficiently. Plan for capacity and other storage needs.
Enforcement limits Enforcement limits include all of the functionality of the accounting option, plus the ability to limit disk storage and send notifications. Using enforcement limits, you can logically partition a cluster to control or restrict how much storage that a user, group, or directory can use. For example, you can set hard- or soft-capacity limits to ensure that adequate space is always available for key projects and critical applications and to ensure that users of the cluster do not exceed their allotted storage capacity. Optionally, you can deliver real-time email quota notifications to users, group managers, or administrators when they are approaching or have exceeded a quota limit.
If a quota type uses the accounting-only option, enforcement limits cannot be used for that quota. The actions of an administrator logged in as root may push a domain over a quota threshold. For example, changing the protection level or taking a snapshot has the potential to exceed quota parameters. System actions such as repairs also may push a quota domain over the limit. There are three types of administrator-defined enforcement thresholds.
218
SmartQuotas
Description Limits disk usage to a size that cannot be exceeded. If an operation, such as a file write, causes a quota target to exceed a hard quota, the following events occur:
l l l
the operation fails an alert is logged to the cluster a notification is issued to specified recipients.
Writes resume when the usage falls below the threshold. Soft Allows a limit with a grace period that can be exceeded until the grace period expires. When a soft quota is exceeded, an alert is logged to the cluster and a notification is issued to specified recipients; however, data writes are permitted during the grace period. If the soft threshold is still exceeded when the grace period expires, data writes fail, and a hard-limit notification is issued to the recipients you have specified. Writes resume when the usage falls below the threshold. Advisory An informational limit that can be exceeded. When an advisory quota threshold is exceeded, an alert is logged to the cluster and a notification is issued to specified recipients. Advisory thresholds do not prevent data writes.
Disk-usage calculations
For each quota that you configure, you can specify whether data-protection overhead is included in future disk-usage calculations. Overhead settings should be configured carefully, because they can significantly affect the amount of disk space that is available to users. Typically, most quota configurations do not need to include overhead calculations. If you include data-protection overhead in usage calculations for a quota, future disk-usage calculations for the quota include the total amount of space that is required to store files and directories, in addition to any space that is required to accommodate your dataprotection settings, such as parity or mirroring. For example, consider a user who is restricted by a 40 GB quota that includes data-protection overhead in its disk-usage calculations. If your cluster is configured with a 2x data-protection level (mirrored) and the user writes a 10 GB file to the cluster, that file actually consumes 20 GB of space: 10 GB for the file and 10 GB for the data-protection overhead. In this example, the user has reached 50 percent of the 40 GB quota by writing a 10 GB file to the cluster. You can configure quotas to include the space that is consumed by snapshots. A single path can have two quotas applied to it: one without snapshot usage (default) and one with snapshot usage. If snapshots are included in the quota, more files are included in
Disk-usage calculations
219
SmartQuotas
the calculation than are in the current directory. The actual disk usage is the sum of the current directory and any snapshots of that directory. You can see which snapshots are included in the calculation by examining the .snapshot directory for the quota path. Older snapshots are not added retroactively to usage when you create a new quota. Only those snapshots created after the QuotaScan job finishes are included in the calculation. If you do not include data-protection overhead in usage calculations for a quota, future disk-usage calculations for the quota include only the space that is required to store files and directories. Space that is required for the cluster's data-protection setting is not included. Consider the same example user, who is now restricted by a 40 GB quota that does not include data-protection overhead in its disk-usage calculations. If your cluster is configured with a 2x data-protection level and the user writes a 10 GB file to the cluster, that file consumes only 10 GB of space: 10 GB for the file and no space for the dataprotection overhead. In this example, the user has reached 25 percent of the 40 GB quota by writing a 10 GB file to the cluster. This method of disk-usage calculation is typically recommended for most quota configurations. Clones and cloned files are accounted for by quotas as though they consume both shared and unshared data: a clone and a copy of the same file do not consume different amounts of data. If the quota includes data protection overhead, however, the data protection overhead for shared data is not included in the usage calculation.
Quota notifications
Storage quota notifications, which are generated as a part of enforcement quotas, provide users with information about threshold violations when a violation condition occurs and while the violation condition persists. Each notification rule defines the condition that is to be enforced and the action that is to be executed when the condition is true. An enforcement quota can define multiple notification rules. Quota notifications are generated on quota domains with enforcement quota thresholds that you define. When thresholds are exceeded, automatic email notifications can be sent to specified users, or you can monitor notifications as system alerts or receive emails for these events. Notifications can be configured globally, to apply to all quota domains, or be configured for specific quota domains. Enforcement quotas support the following notification settings. A given quota can use only one of these settings. Limit notification settings Turn Off Notifications for this Quota Use Default Notification Rules Use Custom Notification Rules
Description Disables all notifications for the quota. Uses the global default notification for the specified type of quota. Enables the creation of advanced, custom notifications that apply to the specific quota. Custom notifications can be configured for any or all of the threshold types (hard, soft, or advisory) for the specified quota.
220
SmartQuotas
Quota notification rules can be written to trigger an action according to event thresholds (a notification condition). A rule can specify a schedule, such as "every day at 1:00 AM," for executing an action or immediate notification of certain state transitions. When an event occurs, a notification trigger executes one or more specified actions, such as sending an email to a user or administrator or pushing a cluster alert to the interface. Examples of notification conditions include the following:
u u u u
"Notify when a threshold is exceeded; at most, once every 5 minutes" "Notify when allocation is denied; at most, once an hour" "Notify while over threshold, daily at 2 AM" "Notify while grace period expired weekly, on Sundays at 2 AM" Instant notifications Includes the write-denied notification, triggered when a hard
threshold denies a write, and the threshold-exceeded notification, triggered at the moment a hard, soft, or advisory threshold is exceeded. These are one-time notifications because they represent a discrete event in time.
"Notify when a threshold is exceeded; at most, once every 5 minutes" "Notify when allocation is denied; at most, once an hour" "Notify while over threshold, daily at 2 AM" "Notify while grace period expired weekly, on Sundays at 2 AM"
When an event occurs, a notification is triggered according to your notification rule. For example, you can create a notification rule that sends an email to a user or administrator when a disk-space allocation threshold is exceeded by a group.
Quota reports
The OneFS SmartQuotas module provides reporting options that enable administrators to more effectively manage cluster resources and analyze usage statistics. Storage quota reports provide a summarized view of the past or present state of the quota domains. After raw reporting data is collected by OneFS, you can produce data summaries by using a set of filtering parameters and sort types. Storage-quota reports include information about violators, grouped by types of thresholds. You can general reports from a historical data sample or from current data. In either case, the reports are views of usage data at a given time. OneFS does not provide reports on data that is aggregated over time, such as trending reports, but you can use raw data to analyze trends. There is no configuration limit on the number of reports other than the space you need to store them. OneFS provides three methods of data collection and reporting:
u u
Scheduled reports are generated and saved on a regular interval. Ad hoc reports are generated and saved at the request of the user.
Quota notification rules
221
SmartQuotas
Scheduled reports are placed by default in the /ifs/.isilon/smartquotas/reports directory, but the location is configurable to any directory under /ifs. Each generated report includes quota domain definition, state, usage, and global configuration settings. By default, ten reports are kept at a time, and older reports are purged. Ad hoc reports can be created on demand to provide a current view of the storage quotas system. These live reports can be saved manually. Ad hoc reports are saved to a location that is separate from scheduled reports to avoid skewing the timed-report sets.
Creating quotas
You can create two types of storage quotas to monitor data: accounting quotas and enforcement quotas. Storage quota limits and restrictions can apply to specific users, groups, or directories. The type of quota that you create depends on your goal.
u u
Accounting quotas monitor, but do not limit, disk usage. Enforcement quotas monitor and limit disk usage. You can create enforcement quotas that use any combination of hard limits, soft limits, and advisory limits. Enforcement quotas are not recommended for snapshot-tracking quota domains.
After you create a new quota, it begins to report data almost immediately, but the data is not valid until the QuotaScan job completes. Before using quota data for analysis or other purposes, verify that QuotaScan job has finished.
To include the data-protection overhead in the accounting quota, select the Include Data-Protection Overhead check box. To include snapshot data in the accounting quota, select the Include Snapshot Data check box.
7. In the Usage Limits area, click No Usage Limit (Accounting Only) 8. Click Create Quota.
222
SmartQuotas
Results The quota appears in the Quotas & Usage list. What to do next After you create a new quota, it begins to report data almost immediately, but the data is not valid until the QuotaScan job completes. Before using quota data for analysis or other purposes, verify that the QuotaScan job has finished.
Managing quotas
The configured values of a storage quota can be modified, and you can enable or disable a quota. You can modify the default storage quotas, and you can create quota limits and restrictions that apply to specific users, groups, or directories. Quota management in OneFS is simplified by the quota search feature, which enables you to locate a quota or quotas by using filters. You can also clone quotas to speed quota
Create an enforcement quota
223
SmartQuotas
creation, and you can unlink quotas that are associated with a parent quota. Optionally, custom notifications can be configured for quotas. You can also temporarily disable a quota and then enable it when needed. Quotas can be managed through either the web administration interface or the command-line interface. Moving quota directories across quota domains is not supported.
To search for only default users, select the Only show default users checkbox. To search for only default groups, select the Only show default groups check box.
5. In the Directory Path field, type a full or partial path. You can use the wildcard character (*) in the Directory Path field.
l l
To search subdirectories, select the Include subdirectories check box. To search for only quotas that are in violations, select the Only show quotas for which usage limits are currently in violation check box.
6. Optional: Click Update Display. Quotas that match the search criteria appear in the sections where quotas are listed. Results An accounting or enforcement quota with a threshold value of zero is indicated by a dash (). You can click the column headings to sort the result set. To clear the result set and display all storage quotas, in the Quotas & Usage area, select Show all quotas and usage for this report for Report Filters, and then click Update Display.
Manage quotas
Quotas help you monitor and analyze the current or historic use of disk storage. You can search for quotas, and then you can view, modify, delete, and unlink a quota. An initial QuotaScan job must run for the default or scheduled quotas. Otherwise, the data displayed may be incomplete. Before you modify a quota, consider how the changes will affect the file system and end users.
224
SmartQuotas
The options to edit or delete a quota appear only when the quota is not linked to a default quota. The option to unlink a quota is available only when the quota is linked to a default quota.
1. Click File System Management > SmartQuotas > Quotas & Usage. 2. From the Quota Report options, select the type of quota report that you want to view or manage. l To monitor and analyze current disk storage use, click Show current quotas and usage (Live Report).
l
To monitor and analyze historic disk storage use, click Show archived quota report to select from the list of archived scheduled and manually generated quota reports.
3. For Report Filters, select the filters to be used for this quota report. l To view all information in the quota report, click Show all quotas and usage for this report.
l
To filter the quota report, click Search for specific quotas within this report, and then select the filters that you want to apply.
4. Click Update Display. The quota report displays below. 5. Optional: Select a quota to view its settings or to perform the following management actions. l To review or edit this quota, click View details.
l l
To delete this quota, click Delete. To unlink a linked quota, click Unlink. Configuration changes for linked quotas must be made on the parent (default) quota that the linked quota is inheriting from. Changes to the parent quota are propagated to all children. If you want to override configuration from the parent quota, you must first unlink the quota.
SmartQuotas
2. Navigate to the location of the exported quota configuration file. 3. At the command prompt, run the following command, where <filename> is the name of an exported configuration file: isi_classic quota import --from-file=<filename> The system parses the file and imports the quota settings from the configuration file. Quota settings that you configured before importing the quota configuration file are retained, and the imported quota settings are effective immediately.
Threshold exceeded Over-quota reminder Grace period expired Write access denied
If a directory service is used to authenticate users, you can configure notification mappings that control how email addresses are resolved when the cluster sends a quota notification. If necessary, you can remap the domain that is used for quota email notifications and you can remap Active Directory domains, local UNIX domains, or both.
To specify an archive directory that is different from the default, in the Archive Directory field, type the path or click Browse to select the path.
6. In the Manual Report Archiving area, you can configure the following size and directory options: l To configure the number of live reports that you want to archive, type the number of reports in the Limit archive size field.
l
To specify an archive directory that is different from the default, in the Archive Directory field, type the path or click Browse to select the path.
226
SmartQuotas
7. In the Email Mapping Rules area, choose each mapping rule that you want to use by selecting the check box in the Provider Type column. 8. In the Notification Rules area, define default notification rules for each rule type. l Click Default Notifications Settings to expand the list of limit notifications rules types.
l
Click Advisory Limit Notification Rules to display default settings options for this type of notification. Click Event: Advisory Limit Value Exceeded and Event: While Advisory Limit Remains Exceeded to set the options that you want. Click Soft Limit Notification Rules to display default settings options for this type of notification. Click Event: Soft Limit Value Exceeded, Event: While Soft Limit Remains Exceeded, Event: Soft Limit Grace Period Expired, and Event: Soft Limit Write Access Denied to set the options that you want. Click Hard Limit Notification Rules to display the options for this type of notification. Click Event: Hard Limit Write Access Denied and Event: While Hard Limit Remains Exceeded to set the options that you want.
l l
9. Click Save. What to do next After you create a new quota, it begins to report data almost immediately, but the data is not valid until the QuotaScan job completes. Before using quota data for analysis or other purposes, verify that the QuotaScan job has finished.
SmartQuotas
What to do next After you create a new quota, it begins to report data almost immediately, but the data is not valid until the QuotaScan job completes. Before using quota data for analysis or other purposes, verify that the QuotaScan job has finished.
The following example illustrates a custom email template to notify recipients about an exceeded quota.
Text-file contents with variables The disk quota on directory <ISI_QUOTA_PATH> owned by <ISI_QUOTA_OWNER> was exceeded. The <ISI_QUOTA_TYPE> quota limit is <ISI_QUOTA_THRESHOLD>, and <ISI_QUOTA_USAGE> is in use. Please free some disk space by deleting unnecessary files. For more information, contact Jane Anderson in IT. Email contents with resolved variables The disk quota on directory
228
SmartQuotas
Example 1 Example of a custom quota email notification text file (continued, page X of Y) /ifs/data/sales_tools/collateral owned by jsmith was exceeded. The hard quota limit is 10 GB, and 11 GB is in use. Please free some disk space by deleting unnecessary files. For more information, contact Jane Anderson in IT.
What to do next To use the custom template, click Cluster Managements > General Settings > Email Settings, and then select the custom template in the Event Notification Settings area.
SmartQuotas
2. In the Generated Quota Reports Archive area, click Generate a quota report. The Generate a Quota Report area appears. 3. Click Generate Report. Results The new report appears in the Quota Reports list.
If quota reports are not in the default directory, you can run the isi command to find the directory where they are stored. 3. At the command prompt, run one of the following commands: To view a list of all quota reports in the specified directory To view a specific quota report in the specified directory
quota settings
-a *.xml
Description The directory that the quota is on. Select to automatically create a quota for every current or future user that stores data in the specified directory. Select to automatically create a quota for every current or future group that stores data in the specified directory. Select to count all snapshot data in usage limits; cannot be changed after the quota is created. Select to count protection overhead in usage limits. Select to account for usage only.
Group Quota
230
SmartQuotas
Description Specify the type of email to use. Select to send an email notification to the owner of the entity. Select to send an email notification to another recipient and type the recipient's email address. Select from the following template types for use in formatting email notifications:
l
Notify another
Yes
Yes
Message template
Yes
Yes
Default (leave Message Template field blank to use default) Custom Yes
Select to generate an event Yes notification for the quota when exceeded. Specify the length of time (hours, days, weeks) to delay before generating a notification. Specify the notification and alert frequency: daily, weekly, monthly, yearly; depending on selection, specify intervals, day to send, time of day, multiple emails per rule. Yes
Delay
No
Frequency
No
Yes
231
SmartQuotas
Exceeded Yes
Notify owner
Select to send an Yes email notification to the owner of the entity. Select to send an Yes email notification to another recipient and type the recipient's email address. Select from the following template types for use in formatting email notifications:
l
Yes
Yes
Yes
Notify another
Yes
Yes
Yes
Message template
Yes
Yes
Yes
Yes
Default (leave
Create cluster event Select to generate an event notification for the quota. Delay Specify the length of time (hours, days, weeks) to delay before generating a notification. Specify the notification and alert frequency: daily, weekly, monthly, yearly; depending on selection, specify intervals, day to
Yes
No
No
Yes
Frequency
No
Yes
Yes
No
232
SmartQuotas
Option -
Description
send, time of day, multiple emails per rule.
Exceeded -
Remains exceeded -
Description Specify the recipient of the email notification. Select to send an email notification to the owner of the entity. Select to send an email notification to another recipient and type the recipient's email address. Select from the following template types for use in formatting email notifications:
l
Notify another
Yes
Yes
Message template
Yes
Yes
Default (leave Message Template field blank to use default) Custom Yes
Select to generate an event Yes notification for the quota when exceeded. Specify the length of time (hours, days, weeks) to delay before generating a notification. Specify the notification and alert frequency: daily, weekly, monthly, yearly; depending on selection, specify intervals, day to send, time of day, multiple emails per rule. Yes
Delay
No
Frequency
No
Yes
233
SmartQuotas
following notification settings for each threshold type. A quota can use only one of these settings. Notification setting Use Default Notification Rules Turn Off Notifications for this Quota Use Custom Notification Rules
Description Uses the default notification rules that you configured for the specified threshold type. Disables all notifications for the quota. Provides settings to create basic custom notifications that apply to only this quota.
Notes -
Off. Manually generated on-demand reports can be run at any time. On. Reports run automatically
according to the schedule that you specify.
Report frequency
Specifies the interval for this report to run: daily, weekly, monthly, or yearly; you can further refine the report schedule by using the following options.
Generate multiple reports per specified day. Set the intervals and times of day to
generate the report for that day. Scheduled report archiving Determines the maximum number of scheduled reports that are available for viewing on the SmartQuotas Reports page. When the maximum number of reports are stored, the system deletes the oldest reports to make space for new reports as they are generated.
Limit archive size for scheduled reports to a specified number of reports. Type the
234
SmartQuotas
Setting -
Description
integer to specify the maximum number of reports to keep.
Notes -
Archive Directory. Browse to the directory where you want to store quota reports for archiving.
Manual report archiving Determines the maximum number of manually generated (on-demand) reports that are available for viewing on the SmartQuotas Reports page. When the maximum number of reports are stored, the system deletes the oldest reports to make space for new reports as they are generated.
Description Path of quota domain Threshold value Disk space in use Name of quota domain owner Threshold type Grace period, in days Expiration date of grace period
Example /ifs/data 20 GB 10.5 GB jsmith Advisory 5 days Fri Feb 23 14:23:19 PST 2007
235
SmartQuotas
236
Storage pools are a logical division of nodes and files. They give you the ability to aggregate and manage large numbers of files from a single management interface. OneFS uses storage pools to efficiently manage and protect the data on a cluster.
Node pools are sets of like nodes that are grouped into a single pool of storage. Node pool membership changes automatically through the addition or removal of nodes to or from the cluster. File pools are user-defined logical groupings of files that are stored in node pools according to file pool policies. By default, the basic unlicensed technology is implemented in a cluster, and additional features are available when you license the SmartPools module. These licensed features include the ability to create multiple file pools and file pool policies that direct specific files and directories to a targeted node pool or tier and spillover management, which enables you to define how write operations are handled when a node pool or tier is full. Virtual hot spare allocation, which reserves space for data re-protection if a drive fails, is available with both the licensed and unlicensed technology.
The following table compares licensed and unlicensed storage pool features. Feature Automatic pool provisioning Spillover Policy-based data movement Virtual hot spare
u u u u u u u u u u u u u u u
Storage pool overview.........................................................................................238 Autoprovisioning.................................................................................................238 Virtual hot spare and SmartPools........................................................................239 Spillover and SmartPools....................................................................................239 Node pools.........................................................................................................240 SSD pools...........................................................................................................241 File pools with SmartPools..................................................................................241 Tiers....................................................................................................................242 File pool policies.................................................................................................243 Pool monitoring...................................................................................................243 Creating file pool policies with SmartPools..........................................................244 Managing file pool policies.................................................................................245 SmartPools settings............................................................................................248 Default file pool protection settings.....................................................................250 Default file pool I/IO optimization settings..........................................................252
Storage pools
237
Storage pools
Node pools Groups of equivalent nodes that are associated in a single pool of storage. Tiers Groups of node pools, used to optimize data storage according to OneFS platform
type.
OneFS adds the following features when the SmartPools module is licensed.
u
Custom file pools Storage pools that you define to filter files and directories into
specific node pools according to your criteria. Using file attributes such as file size, type, access time, and location that you specify in a file pool policy , custom file pools automate data movement and storage according to your unique storage needs. The licensed module also includes customizable template policies that are optimized for archiving, extra protection, performance, and VMware files.
Storage pool spillover Automated node-capacity overflow management. Spillover defines how to handle write operations when a storage pool is not writable. When spillover is enabled, data is redirected to a specified storage pool. If spillover is disabled, new data writes fail and an error message appears. If the SmartPools module is not licensed, files are stored on any available node pools across the cluster.
Autoprovisioning
Autoprovisioning is the process of automatically assigning storage by node type to improve the performance and reliability of the file storage system. When you configure a cluster, OneFS automatically assigns nodes to node pools, or autoprovisions, in your cluster to increase data-protection and cluster reliability. Autoprovisioning reduces the time required for the manual management tasks associated with configuring storage pools and resource planning.
238
Storage pools
Nodes are not provisioned, meaning they are not associated with each other and not writable, until at least three nodes of an equivalence class are assigned to the pool. If you have added only two nodes of an equivalence class to your cluster, there is no communication between nodes until one more is added. If you remove nodes from a provisioned cluster so that fewer than three equivalenceclass nodes remain, the pool is underprovisioned. In this situation, when two like nodes remain, they are still writable; if only one node remains, it is not writable but it remains readable. Node pool attributes and status are visible in the web administration interface. You can view storage pool health information through the command-line interface also.
A minimum number of virtual drives in the node pool (1-4). A minimum percentage of total disk space (0-20 percent) . A combination of minimum virtual drives and total disk space. The larger number of the two determines the space allocation, not the sum of the numbers.
If you configure both settings, the enforced minimum value satisfies both requirements. If you select the option to reduce the amount of available space, free-space calculations do not include the space reserved for the virtual hot spare. The reserved virtual hot spare free space is used for write operations unless you select the option to deny new data writes. If Reduce amount of available space is enabled while Deny new data writes is disabled, it is possible for the file system to report utilization as more than 100 percent.
Virtual hot spare reservations affect spillover. For example, if the virtual hot spare reservation is 10 percent of storage pool capacity, spillover occurs when the storage pool is 90 percent full.
Storage pools
Virtual hot spare reservations affect spillover. For example, if the virtual hot spare reservation is 10 percent of storage pool capacity, spillover occurs when the storage pool is 90 percent full.
Node pools
A node pool is a logical grouping of equivalent nodes across the cluster. OneFS nodes are grouped automatically to create a storage pool for ease of administration and application of file pool target policies. Each node in the OneFS clustered storage system is a peer, and any node can handle a data request. File pool policies can then be applied to files to target node pools that have different performance and capacity characteristics to meet different workflow requirements. Each node that is added to a cluster increases aggregate disk, cache, CPU, and network capacity. When additional nodes are added to the cluster, they are automatically added to node pools according to matching attributes, such as drive size, RAM, series, and SSD-node ratio.
240
Storage pools
SSD pools
OneFS clusters can contain both HDDs and SSDs. When OneFS autoprovisions nodes, nodes with SSDs are grouped into equivalent node pools. Your SSD strategy defines how SSDs are used within the cluster. Clusters that include both hard-disk drives (HDDs) and solid-state drives (SSDs) can be optimized by your SSD strategy options to increase performance across a wide range of workflows. SSD strategy is applied on a per file basis. When you select your options during the creation of a file pool policy, you can identify the directories and files in the OneFS cluster that require faster or slower performance. OneFS automatically moves that data to the appropriate pool and drive type. Global namespace acceleration (GNA) allows data stored on node pools without SSDs to use SSDs elsewhere in the cluster to store extra metadata mirrors, which accelerates metadata read operations. To avoid overloading node pools with SSDs, certain thresholds must be satisfied for GNA to be enabled. GNA can be enabled if 20% or more of the nodes in the cluster contain at least one SSD and 1.5% or more of the total cluster storage is SSD-based. For best results, ensure that at least 2.0% of the total cluster storage is SSD-based before enabling global namespace acceleration. If the ratio of accessible SSD-containing nodes in the cluster drops below the 20% requirement, GNA is not active despite being enabled. GNA is reactivated when the ratio is corrected. The following SSD strategy options are listed in order of slowest to fastest choices.
u
Avoid SSDs Writes all associated file data and metadata to HDDs only.
Use this option to free SSD space only after consulting with Isilon Technical Support personnel. Using this strategy may negatively affect performance.
u
Metadata read acceleration This is the default setting. Writes both file data and metadata to HDDs. An extra mirror of the file metadata is written to SSDs, if available. The SSD mirror is in addition to the number required to satisfy the protection level. Enabling GNA makes read acceleration available to files in node pools that do not contain SSDs. GNA is only for metadata and extra mirrors. Metadata read/write acceleration Writes file data to HDDs and metadata to SSDs,
when available. This strategy accelerates metadata writes in addition to reads but requires about four to five times more SSD storage than the Metadata read acceleration setting. Enabling GNA does not affect read/write acceleration.
Data on SSDs Uses SSD node pools for both data and metadata. Regardless of whether
global namespace acceleration is enabled, any SSD blocks reside on the file target pool if there is room. This SSD strategy does not result in the creation of additional mirrors beyond the normal protection level but requires significantly increased storage requirements compared with the other SSD strategy options.
Storage pools
pool policies to file pools, data can be moved automatically from one type of storage to another within a single cluster from a single point of management to meet performance, space, cost, or other requirements while retaining protection-level settings. File pool policies are based on the file attributes that you specify. For example, a file pool policy can be created for an specific file extension that requires high availability, so you can target a pool that provides the fastest reads or read/writes. Another file pool policy can be created to evaluate last accessed date, allowing you to target node pools best suited for archiving for historical or regulatory purposes.
Tiers
A tier is a user-defined collection of node pools that can be used as a target for a file pool policy. You can create tiers to assign your data to any of the node pools in the tier to meet your data-classification needs. For example, a collection of node pools can be assigned to a tier that you create for frequently accessed or mission-critical data that requires high availability and fast access. In a three-tier system, this classification may be Tier 1. You can classify data that is used less frequently or that is accessed by fewer users as Tier-2 data. Tier-3 usually comprises data that is seldom used and can be archived for historical or regulatory purposes. A node pool can belong to only one tier.
Create a tier
You can group node pools into a tier that can be used as a target for a file pool policy. 1. Click File System Management > SmartPools > Summary. The SmartPools page appears and displays two groupings: the current capacity usage and a list of tiers and node pools. 2. In the Tiers & Node Pools section, click Create a Tier. 3. In the dialog box that displays, type a name for this tier, and then click Submit. The tier appears in the list of tiers and node pools. 4. Select and drag a node pool to the tier name to add it to the tier. Continue dragging and dropping node pools until you complete the tiered group. Each node pool that you added to the tier appears under the tier name when it is in an expanded state.
Rename a tier
You can modify the name of a tier that contains node pools. A tier can contain alphanumeric characters but cannot begin with a number. 1. Click File System Management > SmartPools > Summary. The SmartPools page appears and displays two groupings: the current capacity usage and a list of tiers and node pools. 2. In the Tiers & Node Pools area, in the row of the tier you want to rename, click Edit. 3. In the dialog box that displays, type a name for this tier and click Submit. The newly named tier appears in the list of tiers and node pools.
242
Storage pools
Delete a tier
You can delete a tier, but the option is not available until you move all node pools out of that tier. 1. Click File System Management > SmartPools > Summary. The SmartPools page appears and displays two groupings: current capacity usage and a list of tiers and node pools. 2. In the Tiers & Node Pools area, in the row of the tier that you want to delete, click Delete. 3. In the confirmation dialog box that displays, click Yes to confirm the deletion. Results The tier is removed from list of tiers and node pools.
Pool monitoring
Pool health, performance, and status can be monitored through the web administration interface or the command-line interface. Information is displayed for individual nodes, including node-specific network traffic, internal and external network interfaces, and drive status. You can configure real-time and historical performance to be graphed in the web administration interface. You can assess pool health and performance by viewing the following information:
u u u
Storage pools
u u u
Specifies Name of the file Where the file is stored File-system object type Size of the file When the file was last modified When the file was created
244
Storage pools
Specifies When the file metadata was last modified When the file was last accessed Custom attributes
OneFS supports UNIX shell-style (glob) pattern matching for file name attributes and paths, using these characters: *, ?, and [a-z]. As many as four file pool policies can apply to a file (one per action) if the stop processing option is not selected. However, if the stop processing option is selected when you create a file-pool policy, only one file pool policy can be applied because OneFS applies only the first matching policy rule that it encounters. If a file type matches multiple policies, subsequent policies in the list are not evaluated. If one policy rule moves all JPG files to a nearline node pool and another policy rule moves all files smaller than 2 MB to a performance tier and the JPG rule is first in the list, then all JPG files smaller than 2 MB are moved to nearline storage instead of to the performance tier. OneFS provides customizable template policies that archive older files, increase the protection level for specified files, send files that are saved to a particular path to a higher-performance disk pool, and change the access setting for VMWare files. You also can copy any file pool policy except the default file pool policy, and then modify the settings that you need to change. After a file pool policy is created, OneFS stores and lists it with other file pool policies. When the SmartPools job runs, it traverses the stored file pool policies list from top to bottom (per file) and policies are applied in the order of that list. The file pool policy list can be reordered at any time, but the default file pool policy is always last in the list of enabled file pool policies.
Modify file pool policies Modify the default file pool policy Copy file pool policies Use a file pool policy template Reorder file pool policies Delete file pool policies
Storage pools
Results Changes to the default file pool policy are applied when the next scheduled SmartPools job runs.
If existing file pool policies direct data to a specific node pool or tier, do not add or modify a file pool policy to target anywhere for the Data storage target option. Target a specific file pool instead. 1. Click File System Management > SmartPools > Settings. The SmartPools page appears. 2. In the SmartPools Settings section, choose the settings that you want apply as the global default for Data storage target, Snapshot storage target, or Protection level. 3. Click Submit. The settings that you selected are applied to any entity that is not covered by another file pool policy.
If existing file pool policies direct data to a specific node pool or tier, do not add or modify a file pool policy to target anywhere for the Data storage target option. Target a specific file pool instead. 1. Click File System Management > SmartPools > File Pool Policies. The SmartPools page appears and displays three groupings: a list of file pool policies, a list of template policies, and latest scan job results. 2. In the File Pool Policies area, in the Actions column of the file pool policy you want to modify, click Copy. The settings options appear. 3. Make your changes in the appropriate areas and click Submit.
246
Storage pools
Results Changes to the file pool policy are applied when the next scheduled SmartPools job runs. To run the job immediately, click Start SmartPools Job.
Storage pools
SmartPools settings
SmartPools settings include directory protection, global namespace acceleration, virtual hot spare, node pool, spillover, protection management, and I/O optimization management. Setting Directory protection
Description Increases the amount of protection for directories at a higher level than the directories and files that they contain, so that data that is not lost can still be accessed.
When this setting is disabled, the directory that contains a file pool is protected according to your protectionWhen devices failures result in data loss level settings, but the devices used to (for example, three drives or two nodes in store the directory and the file may not be a +2:1 policy), enabling this setting the same. There is potential to lose nodes ensures that intact data is still accessible. with file data intact but not be able to access the data because those nodes contained the directory. As an example, consider a cluster that has a +2 default file pool protection setting and no additional file pool policies. OneFS directories are always mirrored, so they are stored at 3x, which is the mirrored equivalent of the +2 default. This configuration can sustain a failure of two nodes before data loss or inaccessibility. If this setting is enabled, all directories are protected at 4x. If the cluster experiences three node failures, although individual files may be inaccessible, the directory tree is
248
Storage pools
Setting -
Description -
Notes
available and provides access to files that are still accessible. In addition, if another file pool policy protects some files at a higher level, these too are accessible in the event of a threenode failure.
Specifies whether to allow per-file metadata to use SSDs in the node pool.
l
This setting is available only if 20 percent or more of the nodes in the cluster contain SSDs and at least 1.5 percent of the total Disabled. Restrict per-file metadata to cluster storage is SSD-based the target pool of the file, except in the case of spillover. This is the default setting. use the SSDs in any node pool.
Reserves a minimum amount of space in the node pool that can be used for data migration in the event of a drive failure. To reserve disk space for use as a virtual hot spare, select one or both of the following options:
l
Deny new data writes. Prevents write operations from using reserved disk space. VHS space to reserve. You can reserve a minimum number of virtual drives (1-4), as well as a minimum percentage of total disk space (0-20%).
Global spillover
Enabled. Redirect write operations from a node pool that is not writable to another node pool. Disabled. Return a disk space error
for write operations to a node pool that is not writable
Spillover data to
When spillover is enabled but it is important that data writes do not fail, select anywhere for the Spillover data to setting, even if file pool policies send data to specific pools. Disabling both protection management and I/O optimization management
SmartPools settings
249
Protection management
Storage pools
Setting -
Description
l
Disabling both protection management and I/O optimization management settings disables SmartPools functionality. You can modify the default settings in the
Description Specifies the node pool or tier that you want to target with this file pool policy.
Notes If GNA is not enabled and the pool that you choose to target does not contain SSDs, you cannot define a strategy. Metadata read acceleration
writes both file data and metadata to HDD pools but adds an additional SSD mirror if possible to accelerate read performance. Uses HDDs to provide reliability and an SSD, if available, to improve read performance. Recommended for most uses. When you select Metadata
If existing file pool policies direct data to a specific node pool or tier, do not add or modify a file pool policy to target anywhere for the Data storage target option. Target a specific file pool instead.
Select one of the following options to define your SSD strategy:
u
250
Storage pools
Setting -
Description data and metadata to HDDs and metadata to SSDs. Accelerates metadata reads only. Uses less SSD space than the Metadata read/write acceleration setting.
u
Notes
in the file pool policy, for performance and reliability. The extra mirror may be from a different node pool using GNA enabled or from the same node pool. The Data on SSDs strategy does not result in the creation of additional mirrors beyond the normal protection level. Both file data and metadata are stored on SSDs if available within the file pool policy. This option requires a significant amount of SSD storage.
Avoid SSDs Write all associated file data and metadata to HDDs only.
Use this to free SSD space only after consulting with Isilon Technical Support personnel; may negatively affect performance.
u
Data on SSDs Use nodes with SSDs for both data and metadata. Regardless of whether global namespace acceleration is enabled, any SSD blocks reside on the file target pool if there is room.
Notes for Data storage target apply to snapshot storage target
Specifies the node pool or tier that you want to target for snapshot storage with this file pool policy. The settings are the same as those for Data storage target, but apply to snapshot data.
Protection level
Default protection level of disk To change the protection policy to a specific level, select a new pool. Assign the default
protection policy of the disk pool to the filtered files. value from the list.
251
Storage pools
Setting -
Description Specific level. Assign a specified protection policy to the filtered files. -
Notes
Notes SmartCache can improve performance, but can also lead to data loss if a node loses power or crashes while uncommitted data is in the write cache. By default, iSCSI LUNs are configured to use a random access pattern. Other files and directories use a concurrent access pattern by default.
Defines the optimization settings for accessing data: Concurrency, Streaming, or Random.
252
CHAPTER 13 Networking
After you determine the topology of your network, you can set up and manage your internal and external networks. There are two types of networks associated with a cluster:
u
Internal Nodes use the internal network to communicate with one another.
Communication occurs through InfiniBand connections. You can optionally configure a failover network for redundancy.
External Clients connect to the cluster through the external network with Ethernet. The
Isilon cluster supports standard network communication protocols, including NFS, SMB, HTTP, and FTP. The cluster includes various external Ethernet connections, providing flexibility for a wide variety of network configurations. External speeds vary by product.
With the cluster's web administration interface, you can manage both the internal and external network settings from a centralized location.
u u u u u u
Cluster internal network overview........................................................................254 External client network overview.........................................................................254 Configuring the internal cluster network..............................................................259 Configuring an external network..........................................................................261 Managing external client connections with SmartConnect...................................275 Managing network interface provisioning rules....................................................276
Networking
253
Networking
IP address ranges on separate subnets for the int-a, int-b, and failover networks are configured. The int-b interface is enabled. Enabling an internal failover network requires that the cluster be rebooted.
Networking
How you set up your external network subnets depends on your network topology. In a basic network topology where all client-node communication occurs through a single gateway, only a single external subnet is required. If clients connect through multiple external subnets or internal connections, you must configure multiple external network subnets.
Netmask IP address range Gateway Domain name server list (optional) DNS search list (optional) SmartConnect zone name (optional) SmartConnect service address (optional) Creates a default external network subnet called subnet0, with the specified netmask, gateway, and SmartConnect service address. Creates a default IP address pool called pool0 with the specified IP address range, the SmartConnect zone name, and the external interface of the first node in the cluster as the only member. Creates a default network provisioning rule called rule0, which automatically assigns the first external interface for all newly added nodes to pool0. Adds pool0 to subnet0 and configures pool0 to use the virtual IP of subnet0 as its SmartConnect service address. Sets the global, outbound DNS settings to the domain name server list and DNS search list, if provided.
After you configure these settings, OneFS performs the following actions:
u
After you have configured the network configuration for the cluster through the commandline Configuration wizard, you can make changes to your external network settings through the web administration interface. For example, you can add external network subnets, or modify existing external network settings such as subnets, IP address pools, and network provisioning rules.
IP address pools
IP address pools are logical network partitions of the nodes and external network interfaces that belong to a cluster. IP address pools are also used to configure SmartConnect zones and IP failover support for protocols such as NFS.
u
Belong to external network subnets. Allow you to partition your cluster's network interfaces into groups. Can be to assigned to groups in your organization.
External network settings
255
u u
Networking
Multiple pools for a single subnet require a configured SmartConnect Advanced license. The IP address pool of a subnet consists of one or more ranges of IP addresses and a set of cluster interfaces. All IP address ranges in a pool must be unique. A default IP address pool is configured during the initial cluster setup using the command-line Configuration wizard. You can modify the default IP address pool at any time. Additional pools can also be added, removed, or modified. If you add external network subnets to your cluster by using the Subnet wizard, you must specify the IP address pools that belong to the subnet. IP address pools are allocated to external network interfaces either dynamically or statically. The static allocation method assigns one IP address per pool interface. The IP addresses remain assigned, regardless of that interface's status, but the method does not guarantee that all IP addresses are assigned. The dynamic allocation method distributes all pool IP addresses, and the IP address can be moved depending on the interface's status and connection policy settings.
Basic The unlicensed Basic mode balances client connections by using a round robin
policy. The Basic mode is limited to static IP address allocation and to one IP address pool per external network subnet. This mode is included with OneFS as a standard feature and does not require a license.
Advanced The licensed Advanced mode enables features such as CPU utilization,
connection counting, and client connection policies in addition to the round robin policy. The Advanced mode also allows IP address pools to be defined to support multiple DNS zones within a single subnet, and supports IP failover.
The following information describes the SmartConnect DNS client-connection balancing policies:
u
Round Robin This method selects the next available node on a rotating basis. This is the
default state (after SmartConnect is activated) if no other policy is selected.
Round robin is the only connection policy available without a SmartConnect advanced license.
u
Connection Count This method determines the number of open TCP connections on
each available node and optimizes the cluster usage.
Network Throughput This method determines the average throughput on each available node to optimize the cluster usage. CPU Usage This method determines the average CPU utilization on each available node to optimize the cluster usage.
SmartConnect requires that a new name server (NS) record is added to the existing authoritative DNS zone containing the cluster.
256
Networking
External IP failover
External IP failover redistributes IP addresses among node interfaces in an IP address pool when one or more interfaces become unavailable. To enable dynamic IP allocation and IP failover in your cluster, you must have an active SmartConnect Advanced license. The unlicensed SmartConnect Basic, provided as a standard feature in the OneFS operating system, does not support IP failover. Dynamic IP allocation ensures that all IP addresses in the IP address pool are assigned to member interfaces. Dynamic IP allocation allows clients to connect to any IP addresses in the pool and receive a response. If a node or an interface becomes unavailable, OneFS moves the IP address to other member interfaces in the IP address pool. IP failover ensures that all of the IP addresses in the pool are assigned to an available node. When an interface becomes unavailable, the dynamic IP address of the node is redistributed among the remaining available interfaces. Subsequent client connections are directed to the node that is assigned to that IP address. If your cluster has an active SmartConnect Advanced license, you may have already enabled IP failover during the process of running the Subnet wizard to configure your external network settings. You can also modify your subnet settings at any time to enable IP failover for selected IP address pools. IP failover occurs when a pool has a dynamic IP address allocation set. You can further configure IP failover for your network environment by using the following options:
u
IP allocation method This method ensures that all of the IP addresses in the pool are
assigned to an available node.
Rebalance policy This policy controls how IP addresses are redistributed when node
interface members for a given IP address pool become available after a period of unavailability.
IP failover policy This policy determines how to redistribute the IP addresses among
remaining members of an IP address pool when one or more members are unavailable.
SmartConnect requires that a new name server (NS) record be added to the existing authoritative DNS zone that contains the cluster.
NIC aggregation
Network interface card (NIC) aggregation, also known as link aggregation, is optional, and enables you to combine the bandwidth of a node's physical network interface cards into a single logical connection. NIC aggregation provides improved network throughput. Configuring link aggregation requires advanced knowledge of network switches. Consult your network switch documentation before configuring your cluster for link aggregation. NIC aggregation can be configured during the creation of a new external network subnet by using the Subnet wizard. Alternatively, NIC aggregation can be configured on the existing IP address pool of a subnet. When you configure a node through the web administration interface to enable NIC aggregation, storage administrators must be aware of the following:
u
External IP failover
257
Networking
Link Aggregation Control Protocol (LACP) Supports the IEEE 802.3ad Link
Aggregation Control Protocol (LACP). This method is recommended for switches that support LACP and is the default mode for new pools.
Legacy Fast EtherChannel (FEC) mode This method is compatible with aggregated configurations in earlier versions of OneFS. Etherchannel (FEC) This method is a newer implementation of the Legacy FEC
mode.
Active / Passive Failover This method transmits all data transmits through the
master port, which is the first port in the aggregated link. The next active port in an aggregated link takes over if the master port is unavailable.
Round-Robin Tx This method balances outbound traffic across all active ports in the
aggregated link and accepts inbound traffic on any port.
u u
Some NICs may allow aggregation of ports only on the same network card. For LACP and FEC aggregation modes, the switch must support IEEE 802.3ad link aggregation. Since the trunks on the network switch must also be set up, the node must be correctly connected with the right ports on the switch.
VLANs
Virtual LAN (VLAN) tagging is an optional setting for the external network subnet that enables a cluster to participate in multiple virtual networks. A VLAN is a group of hosts that communicate as though they are connected to the same local area network regardless of their physical location. Enabling a cluster to participate in a VLAN allows multiple cluster subnet support without multiple network switches, so that one physical switch enables multiple virtual subnets. Multiple cluster subnets can be supported without multiple network switches, so that one physical switch enables multiple virtual subnets. Configuring a VLAN requires advanced knowledge of network switches. Consult your network switch documentation before configuring your cluster for a VLAN.
IPv6 support
OneFS provides support for IPv6 through a dual-stack configuration. You can configure a cluster with IPv6 addresses. With dual-stack support in OneFS, you can use both IPv4 and IPv6 addresses. However, configuring a cluster to use IPv6 exclusively is not supported. When you set up the cluster, the initial subnet must use IPv4 addresses. The following table describes important distinctions between IPv4 and IPv6.
258
Networking
Configure settings for the int-b and failover networks Enable internal network failover Modify the int-b and failover network settings Delete IP addresses Migrate IP addresses
You can configure the int-b and failover internal networks to provide back up networks in the event of an int-a network failure. Configuration involves specifying a valid netmask and IP address range for the network.
3. In the IP Ranges area, you can add, delete or migrate your IP address ranges. Ideally, the new range is contiguous with the previous one. For example, if your current IP address range is 192.168.160.60-192.168.160.162, the new range should start with 192.168.160.163. 4. Click Submit. If you entered a contiguous range, the new range appears as one range that includes the IP addresses you added as well as the previous ones.
Networking
For the changes in netmask value to take effect, you must reboot the cluster. 1. Click Cluster Configuration > Network Configuration. 2. In the Internal Network Settings area, select the network that you want to configure the netmask for. l To select the int-a network, click int-a.
l
3. In the Netmask field, type a netmask value. You cannot modify the netmask value if the change invalidates any node addresses. 4. Click Submit. A dialog box prompts you to reboot the cluster. 5. Specify when you want to reboot the cluster. l To immediately reboot the cluster, click Yes. When the cluster finishes rebooting, the login page appears.
l
Click No to return to the Edit Internal Network page without changing the settings or rebooting the cluster.
Networking
The Confirm Cluster Reboot dialog box appears. 11. To reboot the cluster, click Yes.
Adding a subnet
OneFS provides a four-step wizard that enables you to add and configure an external subnet. This procedure explains how to start the Subnet wizard and configure the new subnet. To add a subnet, you must perform the following steps: 1. Configure the subnet's basic and advanced settings. 2. Assign an initial IP address pool to be used by the subnet. 3. Optional: Configure SmartConnect for the IP address pool. 4. Assign external network interfaces to the subnet's IP address pool.
Networking
For an IPv6 subnet, click IPv6 in the IP family list. In the Prefix length field, type an integer (ranging from 1 to 128) that represents the network prefix length.
6. In the MTU list, type or select the size of the maximum transmission units the cluster uses in network communication. Any numerical value is allowed, but might not be compatible with your network. Common settings are 1500 (standard frames) and 9000 (jumbo frames). Although OneFS supports both 1500 MTU and 9000 MTU, it is recommended that you configure switches for jumbo frames. Jumbo frames enable the cluster to more efficiently communicate with all the nodes in the cluster. To benefit from using jumbo frames, all devices in the network path must be configured for jumbo frames. 7. In the Gateway address field, type the IP address of the gateway server device through which the cluster communicates with systems outside of the subnet. 8. In the Gateway priority field, type an integer for the priority of the subnet gateway for nodes assigned to more than one subnet. You can configure only one default gateway per node, but each subnet can be assigned a gateway. When a node belongs to more than one subnet, this option enables you to define the preferred default gateway. A value of 1 represents the highest priority, and 10 represents the lowest priority. 9. If you plan to use SmartConnect for connection balancing, in the SmartConnect service IP field, type the IP address that will receive all incoming DNS requests for each IP address pool according to the client connection policy. You must have at least one subnet configured with a SmartConnect service IP in order to use connection balancing. 10. Optional: In the Advanced section, you can enable VLAN tagging if you want to enable the cluster to participate in virtual networks. Configuring a VLAN requires advanced knowledge of network switches. Consult your network switch documentation before configuring your cluster for a VLAN. 11. If you enable VLAN tagging, you must also type a VLAN ID that corresponds to the ID number for the VLAN set on the switch, with a value from 2 to 4094. 12. Optional: In the Hardware load balancing field, type the IP address for a hardware load balancing switch using Direct Server Return (DSR). This routes all client traffic to the cluster through the switch. The switch determines which node handles the traffic for the client, and passes the traffic to that node. 13. Click Next. The Step 2 of 4 -- IP Address Pool Settings dialog box appears. What to do next This is the first of four steps required to configure an external network subnet. To save the network configuration, you must complete the remaining steps. For information on the next step, see Configure an IP address pool.
262
Networking
Networking
unique among the pools served by the SmartConnect service subnet specified in Step 3 below. 2. In the Connection policy list, select the type of connection balancing policy the IP address pool for this subnet uses. The policy determines how SmartConnect distributes incoming DNS requests across the members of an IP address pool. Round Robin Connection Count Selects the next available node on a rotating basis, and is the default state if no other policy is selected. Determines the number of open TCP connections on each available node to optimize the cluster usage.
Network Throughput Uses the overall average throughput volume on each available node to optimize the cluster usage. CPU Usage Examines average CPU utilization on each available node to optimize the cluster usage.
3. In the SmartConnect service subnet list, select the name of the external network subnet whose SmartConnect service will answer DNS requests on behalf of the IP address pool. A pool can have only one SmartConnect service answering DNS requests. If this option is left blank, the IP address pool the subnet belongs to is excluded when SmartConnect answers incoming DNS requests for the cluster. If you have activated an optional SmartConnect Advanced license, complete the following steps for the options in the SmartConnect Advanced section of this wizard page. 4. In the IP allocation method list, select the method by which IP addresses are assigned to the member interfaces for this IP address pool: Static Select this IP allocation method to assign IP addresses when member interfaces are added to the IP pool. As members are added to the pool, this method allocates the next unused IP address from the pool to each new member. After an IP address is allocated, the pool member keeps the address indefinitely unless one of the following items is true:
l l l
The member interface is removed from the network pool. The member node is removed from the cluster. The member interface is moved to another IP address pool.
Dynamic Select this IP allocation method to ensure that all IP addresses in the IP address pool are assigned to member interfaces, which allows clients to connect to any IP addresses in the pool and be guaranteed a response. If a node or an interface becomes unavailable, their IP addresses are automatically moved to other available member interfaces in the pool. If you select the dynamic IP allocation method, you can specify the SmartConnect Rebalance policy and the IP failover policy in the next two steps. 5. Select the type of SmartConnect Rebalance policy to use when IP addresses are redistributed. IP addresses redistribution occurs when node interface members in an IP address pool become available. These options can only be selected if the IP allocation method is set to Dynamic.
264
Networking
Automatically redistributes IP addresses. The automatic rebalance is triggered by a change to one of the following items.
l l l
Manual Failback
Does not redistribute IP addresses until you manually issue a rebalance command through the command-line interface.
This is the default policy that automatically redistributes IP addresses. The automatic rebalance is triggered by a change to one of the following items. :
6. The IP failover policyalso known as NFS failoverdetermines how to redistribute the IP addresses among remaining members of an IP address pool when one or more members are unavailable. In order to enable IP failover, you must first set the IP allocation method to Dynamic, and then select which type of IP failover policy to use: Round Robin Connection Count Selects the next available node on a rotating basis, and is the default state if no other policy is selected. Determines the number of open TCP connections on each available node to optimize the cluster usage.
Network Throughput Uses the overall average throughput volume on each available node to optimize the cluster usage. CPU Usage Examines average CPU utilization on each available node to optimize the cluster usage.
7. Click Next to store the changes that you made to this wizard page. The Step 4 of 4 IP Address Pool members dialog box appears. What to do next This is the third of four steps required to configure an external network subnet. To save the network configuration, you must complete each of the remaining steps. For information on the next step, see Select the interface members for an IP address pool.
Networking
In the case of aggregated links, choose the aggregation mode that corresponds to the switch settings from the Aggregation mode drop-down. Configuring link aggregation requires advanced knowledge of how to configure network switches. Consult your network switch documentation before configuring your cluster for link aggregation. 2. When you have finished assigning external network interfaces to the IP address pool, click Submit. The external subnet settings you configured by using the Subnet wizard appear on the Edit Subnet page. What to do next You can change the subnet configuration settings at any time without going through the four-step wizard process. For more information, see Managing external client subnets.
266
Networking
configured on each Isilon node, but each subnet can have its own gateway. If a node belongs to more than one subnet, this option enables you to define the preferred default gateway. A value of 1 is the highest priority, with 10 being the lowest priority. SmartConnect service IP The IP address that receives incoming DNS requests from outside the cluster. SmartConnect responds to these DNS requests for each IP address pool according to the pool's client-connection policy. To use connection balance, at least one subnet must be configured with a SmartConnect service IP address.
5. Optional: Modify the Advanced settings as needed. Configuring a virtual LAN requires advanced knowledge of network switches. Consult your network switch documentation before configuring your cluster for a VLAN. If you are not using a virtual LAN, leave the VLAN options disabled. VLAN tagging You can enable VLAN tagging. VLAN tagging allows a cluster to participate in multiple virtual networks. VLAN support provides security across subnets that is otherwise available only by purchasing additional network switches. If you enabled VLAN tagging, type a VLAN ID that corresponds to the ID number for the VLAN that is set on the switch, with a value from 1 to 4094. You can enter the IP address for a hardware load balancing switch that uses Direct Server Return (DSR).
VLAN ID
Networking
2. Create a static route by running the following command: isi networks modify pool --name <subnetname>:<poolname> --add-static-routes <subnet>/<netmask><gateway>
3. To ensure that the static route was created, run the following command: isi networks ls pools -v.
3. To ensure that the static route was created, run the following command: isi networks ls pools -v.
Networking
2. Click the name of the subnet you wan too ad an IP address pool to. 3. In the IP Address Pools area, click Add pool by the I. 4. The IP Address Pool wizard appears. For more information on the three steps of the wizard, see Add an IP address pool to a new subnet, Configure SmartConnect settings for a new subnet and Select interface members for a new subnet.
269
Networking
Configure IP failover
You can configure IP failover to reassign an IP address from an unavailable node to a functional node, which enables clients to continue communicating with the cluster, even after a node becomes unavailable. 1. Click Cluster Management > Network Configuration 2. In the External Network Settings area, click the name of the subnet you want to set up IP failover for. 3. Expand the area of the pool you want to modify and click Edit in the SmartConnect Settings area. 4. Optional: In the Zone name field, you can enter a 128-character name for the zone. 5. In the Connection Policy list, select a balancing policy: Round Robin Connection Count Selects the next available node on a rotating basis, and is the default state if no other policy is selected. Determines the number of open TCP connections on each available node to optimize the cluster usage.
Network Throughput Uses the overall average throughput volume on each available node to optimize the cluster usage. CPU Usage Examines average CPU utilization on each available node to optimize the cluster usage.
6. If you purchased a license for SmartConnect Advanced, you will also have access to the following lists:
u
IP allocation method This setting determines how IP addresses are assigned to clients. Select either Dynamic or Static. Rebalance Policy This setting defines the client redirection policy for when a node becomes unavailable. The IP allocation list must be set to Dynamic in order for rebalance
policy options to be selected.
IP failover policy This setting defines the client redirection policy for when an IP
address is unavailable.
Networking
4. Click New to add a new IP address range using the Low IP and High IP fields. or click the respective value in either the Low IP or High IP columns and type a new beginning or ending IP address. 5. Click Submit. 6. In the Pool members area, click Edit. 7. In the Available Interfaces table, select one or more interfaces for the newly added node, and then click the right arrow button to move the interfaces into the Interfaces in current pool table. 8. Click Submit to assign the new node interfaces to the IP address pool.
Networking
Ethernet interfaces, both types of interfaces can be aggregated, but only with interfaces of the same type. NIC aggregation cannot be used with mixed interface types. An external interface for a node cannot be used by an IP address pool in both an aggregated configuration and an individual interface. You must remove the individual interface for a node from the Interfaces in current pool table before configuring an aggregated NIC. Otherwise, the web administration interface displays an error message when you click Submit. Configuring link aggregation requires advanced knowledge of network switches. Consult your network switch documentation before configuring your cluster for NIC aggregation. Before you begin You must enable NIC aggregation on the cluster before you can enable NIC aggregation on the switch. If the cluster is configured but the switch is not configured, then the cluster can continue to communicate. If the switch is configured, but the cluster is not configured, the cluster cannot communicate, and you are unable to configure the cluster for NIC aggregation. 1. Click Cluster Management > Network Configuration. 2. In the External Network Settings area, click the name of the subnet that contains the IP address pool that you want to add aggregated interface members to. 3. In the Pool members area, click Edit. In the case of multiple IP address pools, expand the pool that you want to add the aggregated interfaces to, and then click Edit in the Pool members area. 4. In the Available interfaces table, click the aggregated interface for the node, which is indicated by a listing of AGGREGATION in the Type column. For example, if you want to aggregate the network interface card for Node 2 of the cluster, click the interface named ext-agg, Node 2 under Available interfaces, and then click the right-arrow button to move the aggregated interface to the Interfaces in current pool table. 5. From the Aggregation mode drop-down, select the appropriate aggregation mode that corresponds to the network switch settings. Consult your network switch documentation for supported NIC aggregation modes. OneFS supports the following NIC aggregation modes: Supports the IEEE 802.3ad Link Aggregation Control Protocol Link Aggregation Control Protocol (LACP) (LACP). This method is recommended for switches that support LACP and is the default mode for new pools. Legacy Fast EtherChannel (FEC) mode Etherchannel (FEC) Active / Passive Failover This method is compatible with aggregated configurations in earlier versions of OneFS. This method is the newer implementation of the Legacy FEC mode. This method transmits all data transmits through the master port, which is the first port in the aggregated link. The next
272
Networking
active port in an aggregated link takes over if the master port is unavailable. Round-Robin Tx This method balances outbound traffic across all active ports in the aggregated link and accepts inbound traffic on any port.
6. Click Submit.
273
Networking
4. Reassign the interface members that you want to move by dragging and dropping them from one column to other, or by clicking on an interface member and using the left arrow and right arrow buttons.
Nodes support multiple network card configurations. LNI numbering corresponds to the physical positioning of the NIC ports as found on the back of the node, and LNI mappings are numbered from left to right. Aggregated LNIs are listed in the order in which they are aggregated at the time they are created. NIC names correspond to the network interface name as shown in command-line interface tools such as ifconfig and netstat. NIC Aggregated LNI Aggregated NIC Aggregated NIC (Legacy FEC mode lagg0 fec0
LNI
ext-agg = ext-1 + ext-2 ext-agg = ext-1 + ext-2 ext-agg-2 = ext-3 + ext-4 ext-agg-3 = ext-3 + ext-4 + ext-1 + ext-2 ext-agg = ext-1 + ext-2 10gige-agg-1 = 10gige-1 + 10gige-2
lagg0 lagg1
fec0 fec1
Networking
The domain name you type in the DNS search list field is used for resolving unqualified hostnames. 5. Click Submit.
Network Throughput Uses the overall average throughput volume on each available node to optimize the cluster usage. CPU Usage Examines average CPU utilization on each available node to optimize the cluster usage.
7. In the SmartConnect service subnet list, select the name of the external network subnet whose SmartConnect service answers DNS requests on behalf of the IP address pool.
Managing external client connections with SmartConnect
275
Networking
A pool can have only one SmartConnect service answering DNS requests. If this option is left blank, the IP address pool that the SmartConnect service belongs to is excluded when SmartConnect answers incoming DNS requests for the cluster. If you have purchased a license for the SmartConnect Advanced module, complete the following steps in the SmartConnect Advanced area. 8. In the IP allocation method list, select the method by which IP addresses are assigned to the member interfaces for this IP address pool.
Network Throughput Uses the overall average throughput volume on each available node to optimize the cluster usage. CPU Usage Examines average CPU utilization on each available node to optimize the cluster usage.
5. In the SmartConnect service subnet field, type the subnet used for this policy. 6. Click Submit.
Networking
rule. OneFS automatically checks for multiple provisioning rules when new rules are added to ensure there are no conflicts.
Backup-Accelerator Applies the provisioning rule only to Isilon backup-accelerator nodes that join the cluster. 6. In the then assign interface list, select one of the following an interfaces to assign to the external network subnet and IP address pool for the node specified in the rule: ext-1 ext-2 ext-3 ext-4 ext-agg ext-agg-2 ext-agg-3 The first external Gigabit Ethernet interface on the cluster. The second external Gigabit Ethernet interface on the cluster. The third external Gigabit Ethernet interface on the cluster. The fourth external Gigabit Ethernet interface on the cluster. The first and second external Gigabit Ethernet interfaces aggregated together. The third and fourth external Gigabit Ethernet interfaces aggregated together. The first four external Gigabit Ethernet interfaces aggregated together.
Create a node provisioning rule
277
Networking
All six Gigabit Ethernet interfaces aggregated together. The first external 10 Gigabit Ethernet interface on the cluster. The second external 10 Gigabit Ethernet interface on the cluster.
10gige-agg-1 The first and second external 10 Gigabit Ethernet interfaces aggregated together. 7. In the Subnet list, select the external subnet that the new node will join. 8. In the Pool list, select the IP address pool that belongs to the subnet that should be used by the new node. 9. Click Submit.
278
CHAPTER 14 Hadoop
Hadoop is a flexible, open-source framework for large-scale distributed computation. The OneFS file system can be configured for native support of the Hadoop Distributed File System (HDFS) protocol, enabling your cluster to participate in a Hadoop system. HDFS integration requires a separate license. To obtain additional information or to activate HDFS support for your EMC Isilon cluster, contact your EMC Isilon sales representative.
u u u u u u
Hadoop support overview....................................................................................280 Hadoop cluster integration..................................................................................280 Managing HDFS...................................................................................................280 Configure the HDFS protocol................................................................................280 Create a local user...............................................................................................282 Enable or disable the HDFS service......................................................................282
Hadoop
279
Hadoop
Apache Hadoop 0.20.203.0 Apache Hadoop 0.20.205 Cloudera (CDH3 Update 3) Greenplum HD 1.1
Managing HDFS
To keep the HDFS service performing efficiently on a OneFS cluster, you will need to be familiar with the user and system configuration options available as part of an HDFS implementation. There are two different methods that you can use to manage an HDFS implementation: u Hadoop client machines are configured directly through their Hadoop installation directory.
u
A secure shell (SSH) connection to a node in the Isilon cluster is used to configure the HDFS service.
280
Hadoop
2. To specify which distribution of the HDFS protocol to use, run the isi hdfs command with the --force-version option. Valid values are listed below. Please note that these values are case-sensitive.
l
AUTO: Attempts to match the distribution that is being used by the Hadoop compute node. APACHE_0_20_203: Uses the Apache Hadoop 0.20.203 release. APACHE_0_20_205: Uses the Apache Hadoop 0.20.205 release. CLOUDERA_CDH3: Uses version 3 of Cloudera's distribution, which includes Apache Hadoop. GREENPLUM_HD_1_1: Uses the Greenplum HD 1.1 distribution.
l l l
For example, the following command forces OneFS to use version 0.20.203 of the Apache Hadoop distribution:
isi hdfs --force-version=APACHE_0_20_203
3. To set the default logging level for the Hadoop daemon across the cluster, run the isi hdfs command with the --log-level option. Valid values are listed below, in descending order from the highest to the lowest logging level. The default value is NOTICE. The values are case-sensitive.
l l
EMERG: A panic condition. This is normally broadcast to all users. ALERT: A condition that should be corrected immediately, such as a corrupted system database. CRIT: Critical conditions, such as hard device errors. ERR: Errors. WARNING: Warning messages. NOTICE: Conditions that are not error conditions, but may need special handling. INFO: Informational messages. DEBUG: Messages that contain information typically of use only when debugging a program.
l l l l l l
For example, the following command sets the log level to WARNING:
isi hdfs --log-level=WARNING
4. To set the path on the cluster to present as the HDFS root directory, run the isi hdfs command with the --root-path option. Valid values include any directory path beginning at /ifs, which is the default HDFS root directory. For example, the following command sets the root path to /ifs/hadoop:
isi hdfs --root-path=/ifs/hadoop
Valid values are 4KB to 1GB. The default value is 64MB. For example, the following command sets the block size to 32 MB:
isi hdfs --block-size=32MB
6. To tune the number of worker threads that HDFS uses, run the isi hdfs command with the --num-threads option. Valid values are 1 to 256 or auto, which is calculated as twice the number of cores. The default value is auto.
Configure the HDFS protocol
281
Hadoop
7. To allocate IP addresses from an IP address pool, run isi option. Valid values are in the form <subnet>:<pool>.
For example, the following command allocates IP addresses from a pool named "pool2," which is in the "subnet0" subnet:
isi hdfs --add-ip-pool=subnet0:pool2
To disable the HDFS service, run the following command: isi services isi_hdfs_d disable
282
CHAPTER 15 Antivirus
OneFS enables you to scan the file system for computer viruses and other security threats on an Isilon cluster by integrating with third-party scanning services through the Internet Content Adaptation Protocol (ICAP). OneFS sends files through ICAP to a server running third-party antivirus scanning software. These servers are referred to as ICAP servers. ICAP servers scan files for viruses. After an ICAP server scans a file, it informs OneFS of whether the file is a threat. If a threat is detected, OneFS informs system administrators by creating an event, displaying near real-time summary information, and documenting the threat in an antivirus scan report. You can configure OneFS to request that ICAP servers attempt to repair infected files. You can also configure OneFS to protect users against potentially dangerous files by truncating or quarantining infected files. Before OneFS sends a file to be scanned, it ensures that the scan is not redundant. If a file has not been modified, OneFS will not send the file to be scanned unless the virus database on the ICAP server has been updated since the last scan.
u u u u u u u u u u u u u u
On-access scanning............................................................................................284 Antivirus policy scanning.....................................................................................284 Individual file scanning.......................................................................................284 Antivirus scan reports.........................................................................................285 ICAP servers........................................................................................................285 Supported ICAP servers.......................................................................................285 Anitvirus threat responses...................................................................................286 Configuring global antivirus settings...................................................................287 Managing ICAP servers........................................................................................289 Create an antivirus policy....................................................................................290 Managing antivirus policies.................................................................................291 Managing antivirus scans....................................................................................291 Managing antivirus threats..................................................................................292 Managing antivirus reports..................................................................................293
Antivirus
283
Antivirus
On-access scanning
You can configure OneFS to send files to be scanned before they are opened, after they are closed, or both. Sending files to be scanned after they are closed is faster but less secure. Sending files to be scanned before they are opened is slower but more secure. If OneFS is configured to ensure that files are scanned after they are closed, when a user creates or modifies a file on the cluster, OneFS queues the file to be scanned. OneFS then sends the file to an ICAP server to be scanned when convenient. In this configuration, users can always access their files without any delay. However, it is possible that after a user modifies or creates a file, a second user might request the file before the file is scanned. If a virus was introduced to the file from the first user, the second user will be able to access the infected file. Also, if an ICAP server is unable to scan a file, the file will still be accessible to users. If OneFS ensures that files are scanned before they are opened, when a user attempts to download a file from the cluster, OneFS first sends the file to an ICAP server to be scanned. The file is not sent to the user until the scan is complete. Scanning files before they are opened is more secure than scanning files after they are closed, because users can access only scanned files. However, scanning files before they are opened requires users to wait for files to be scanned. You can also configure OneFS to deny access to files that cannot be scanned by an ICAP server, which can increase the delay. For example, if no ICAP servers are available, users will not be able to access any files until the ICAP servers become available again. If you configure OneFS to ensure that files are scanned before they are opened, it is recommended that you also configure OneFS to ensure that files are scanned after they are closed. Scanning files as they are both opened and closed will not necessarily improve security, but it will usually improve data availability when compared to scanning files only when they are opened. If a user wants to access a file, the file may have already been scanned after the file was last modified, and will not need to be scanned again provided that the ICAP server database has not been updated since the last scan.
284
Antivirus
The time that the scan started. The time that the scan ended. The total number of files scanned. The total size of the files scanned. The total network traffic sent. The network throughput that was consumed by virus scanning. Whether the scan succeeded. The total number of infected files detected. The names of infected files. The threats associated with infected files. How OneFS responded to detected threats. The name and IP address of the user that triggered the scan. This information is not included in reports triggered by antivirus scan policies.
ICAP servers
The number of ICAP servers that are required to support an Isilon cluster depends on how virus scanning is configured, the amount of data a cluster processes, and the processing power of the ICAP servers. If you intend on scanning files only according to antivirus scan policies, it is recommended that you have a minimum of two ICAP servers for a cluster. If you intend on scanning files on-access, it is recommended that you have at least one ICAP server for each node in the cluster. If you configure more than one ICAP server for a cluster, OneFS distributes files to the ICAP servers on a rotating basis, and does not modify the distribution based on the processing power of the ICAP servers. Because of this, it is important to ensure that the processing power of each ICAP server is relatively equal. If one server is significantly more powerful than another, OneFS does not send more files to the more powerful server.
Symantex Scan Engine 5.2 and later. Trend Micro Interscan Web Security Suite 3.1 and later. Kaspersky Anti-Virus for Proxy Server 5.5 and later. McAfee VirusScan Enterprise 8.7 and later with VirusScan Enterprise for Storage 1.0 and later.
285
Antivirus
Alert All threats that are detected cause an event to be generated in OneFS at the warning
level, regardless of the threat response configuration.
Repair The ICAP server attempts to repair the infected file before returning the file to
OneFS.
Quarantine OneFS quarantines the infected file. A quarantined file cannot be accessed
by any user. However, a quarantined file can be removed from quarantine by the root user if the root user is connected to the cluster through secure shell (SSH). If you backup your cluster through NDMP backup, quarantined files will remain quarantined when the files are restored. If you replicate quarantined files to another Isilon cluster, the quarantined files will continue to be quarantined on the target cluster. Quarantines operate independently of access control lists (ACLs).
Truncate OneFS truncates the infected file. When a file is truncated, OneFS reduces the
size of the file to zero bytes to render the file harmless.
You can configure OneFS and ICAP servers to react in one of the following ways when threats are detected:
u
Repair or quarantine Attempts to repair infected files. If an ICAP server fails to repair a
file, OneFS quarantines the file. If the ICAP server repairs the file successfully, OneFS sends the file to the user. Repair or quarantine can be useful if you want to protect users from accessing infected files while retaining all data on a cluster.
Repair or truncate Attempts to repair infected files. If an ICAP server fails to repair a file, OneFS truncates the file. If the ICAP server repairs the file successfully, OneFS sends the file to the user. Repair or truncate can be useful if you are not concerned with maintaining all data on your cluster, and you want to free storage space. However, data in infected files will be lost. Alert only Only generates an event for each infected file. It is recommended that you do
not apply this setting.
Repair only Attempts to repair infected files. Afterwards, OneFS sends the files to the
user, whether or not the ICAP server repaired the files successfully. It is recommended that you do not apply this setting. If you only attempt to repair files, users will still be able to access infected files if the ICAP server fails to repair it.
Quarantine Quarantines all infected files. It is recommended that you do not apply this setting. If you quarantine files without attempting to repair them, you might deny access to infected files that could have been repaired. Truncate Truncates all infected files. It is recommended that you do not apply this
setting. If you truncate files without attempting to repair them, you might delete data unnecessarily.
286
Antivirus
Click Only scan files smaller than the maximum file size and specify a maximum file size.
3. In the Filename restrictions area, specify whether to exclude files from being scanned based on file names and extensions. l Click Scan all files.
l l
Click Only scan files with the following extensions or filenames. Click Scan all files except those with the following extensions or filenames.
4. Optional: If you chose to exclude files based on file names and extensions, specify the criteria by which files will be selected. a. In the Extensions area, click Edit list, and specify extensions. b. In the Filenames area, click Edit list, and specify filenames. You can specify the following wild cards: Wildcard *
Description Matches any string in place of the asterisk. For example, specifying "m*" would match "movies" and "m123" .
[]
Matches any characters contained in the brackets, or a range of characters separated by a dash. For example, specifying "b[aei]t" would match "bat", "bet", and "bit". For example, specifying "1[4-7]2" would match "142", "152", "162", and "172". You can exclude characters within brackets by following the first bracket with an exclamation mark. For example, specifying "b[!ie]" would match "bat" but not "bit" or "bet". You can match a bracket within a bracket if it is either the first or last character. For example, specifying "[[c]at" would match "cat", and "[at".
287
Antivirus
Wildcard -
Description
You can match a dash within a bracket if it is either the first or last character. For example, specifying "car[-s]" would match "cars", and "car-".
Matches any character in place of the question mark. For example, specifying "t?p" would match "tap", "tip", and "top".
5. Click Submit.
To scan files after they are closed, select Scan files when they are closed.
3. In the Directories to be scanned area, specify the directories that you want to apply onaccess settings to. If no directories are specified, on-access scanning settings are applied to all files. If you specify a directory, only files from the specified directories will be scanned as they are accessed. 4. Click Submit.
288
Antivirus
Click Disable.
Antivirus
2. In the ICAP Servers table, in the row of the ICAP server that you want to temporarily disconnect from, click Disable.
6. In the Run policy area, specify whether you want to run the policy according to a schedule or manually. Scheduled policies can also be run manually at any time. Run the policy only manually. Click Manually
Run the policy according a. Click Scheduled. to a schedule. b. In the Interval area, specify on what days you want the policy to run. c. In the Frequency area, specify how often you want the policy to run on the specified days. 7. Click Submit.
290
Antivirus
Click Disable.
Scan a file
You can manually scan an individual file for viruses. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Run the isi
avscan manual command.
291
Antivirus
Rescan a file
You can rescan the file for viruses if, for example, you believe that a file is no longer a threat. 1. Click Data Protection > Antivirus > Detected Threats. 2. In the Detected Threats table, in the row of a file, click Rescan.
Antivirus
View threats
You can view files that are identified as threats by an ICAP server. 1. Click Data Protection > Antivirus > Detected Threats. 2. In the Detected Threats table, view potentially infected files.
Status Displays an icon that indicates the status of the detected threat. The icon appears in one of the following colors:
u u u
Red OneFS did not take any action on the file. Orange The file was truncated. Yellow The file was quarantined.
u u u u
Threat Displays the name of the detected threat as it is recognized by the ICAP server. Filename Displays the name of the file. Directory Displays the directory in which the file is located. Remediation Indicates how OneFS responded to the file when the threat was detected. If OneFS did not quarantine or truncate the file, Infected appears. Detected Displays the time that the file was detected. Policy Displays the name of the antivirus policy that detected the threat. If the threat was detected as a result of a manual antivirus scan of an individual file, Manual scan appears. Currently Displays the current state of the file. File size Displays the size of the file, in bytes. Truncated files display a size of zero bytes.
u u
u u
View threats
293
Antivirus
Anti-Virus scan found threats A threat was detected by an antivirus scan. These events do not provide threat details, but refer to specific reports on the Antivirus Reports page. No ICAP Servers available OneFS is unable to communicate with any
ICAP servers.
294
CHAPTER 16 iSCSI
As an alternative to file-based storage, block-based storage is a flexible way to store and access nearly any type of data. The Isilon iSCSI module enables you to provide block storage for Microsoft Windows, Linux, and VMware systems over an IP network. The Isilon iSCSI module requires a separate license. To obtain additional information about the iSCSI module or to activate the module for your cluster, contact your EMC Isilon sales representative. The Isilon iSCSI module enables you to create and manage iSCSI targets on an Isilon cluster. The targets become available as SCSI block devices on which clients can store structured and unstructured data. iSCSI targets contain one or more logical units, each uniquely identified by a logical unit number (LUN), which the client can format on the local file system and connect to (such as a physical disk device). You can configure separate data protection levels for each logical unit with Isilon FlexProtect or data mirroring. For basic access control, you can configure each target to limit access to a list of initiators. You can also require initiators to authenticate with a target by using the Challenge-Handshake Authentication Protocol (CHAP). The iSCSI module includes the following features:
u u u u u u u u u u u u u u u u u u u
Support for using a Microsoft Internet Storage Name Service (iSNS) server Isilon SmartConnect Advanced dynamic IP allocation Isilon FlexProtect Data mirroring from 2x to 8x LUN cloning One-way CHAP authentication Initiator access control iSCSI targets and LUNs........................................................................................296 iSNS client service...............................................................................................296 Access control for iSCSI targets...........................................................................297 iSCSI considerations and limitations...................................................................297 Supported SCSI mode pages...............................................................................298 Supported iSCSI initiators...................................................................................298 Configuring the iSCSI and iSNS services..............................................................298 Create an iSCSI target..........................................................................................299 Managing iSCSI targets.......................................................................................301 Configuring iSCSI initiator access control............................................................302 Creating iSCSI LUNs.............................................................................................305 Managing iSCSI LUNs..........................................................................................309
iSCSI
295
iSCSI
Description This is the default LUN type for clone LUNs and imported LUNs, and the only type available for newly created LUNs. Normal LUNs can be either writeable or read-only. A snapshot LUN is a copy of a normal LUN or another snapshot LUN. Although snapshot LUNs require little time and disk space to create, they are read-only. You can create snapshot LUNs by cloning existing normal or snapshot LUNs, but you cannot create snapshot clones of clone LUNs. A clone LUN is a copy of a normal, snapshot, or clone LUN. A clone LUN, which is a compromise between a normal LUN and a snapshot LUN, is implemented using overlay and mask files in conjunction with a snapshot. Clone LUNs require little time and disk space to create, and the LUN is fully writeable. You can create clone LUNs by cloning or importing existing LUNs.
Snapshot
Clone
iSCSI
CHAP authentication
The iSCSI module supports the Challenge-Handshake Authentication Protocol (CHAP) to authenticate initiator connections to iSCSI targets. You can restrict initiator access to a target by enabling CHAP authentication and then adding user:secret pairs to the target's CHAP secrets list. Enabling CHAP authentication requires initiators to provide a valid user:secret pair to authenticate their connections to the target. CHAP authentication is disabled by default. The Isilon iSCSI module does not support mutual CHAP authentication.
Multipath I/O (MPIO) is recommended only for iSCSI workflows with primarily readonly operations, because the node must invalidate the data cache on all other nodes during file-write operations and because performance decreases in proportion to the number of write operations. If all MPIO sessions are connected to the same node, performance should not decrease. The Isilon iSCSI module supports one-way Challenge-Handshake Authentication Protocol (CHAP), with the target authenticating the initiator. The authentication configuration is shared by all of the nodes, so a target authenticates its initiator regardless of the node the initiator is connecting through. Mutual CHAP authentication between an initiator and a target is not supported. The Isilon iSCSI module supports the importing of normal LUNs only; importing snapshot LUNs and clone LUNs is not supported. You cannot back up and then restore a snapshot or clone LUN, or replicate snapshot or clone LUNs to another cluster. It is recommended that you deploy a backup application to back up iSCSI LUNs on the iSCSI client, as the backup application ensures that the LUN is in a consistent state at the time of backup.
l
The Isilon iSCSI module does not support the following: Internet Protocol Security (IPsec) Multiple connections per session (MCS) iSCSI host bus adaptors (HBAs)
Access control for iSCSI targets
297
l l
iSCSI
* For the caching mode page, OneFS supports the write cache enable (WCE) parameter only. ** OneFS supports querying this mode page through the Mode Sense command, but does not support changing the fields of this page through the Mode Select command.
iSCSI Initiator Microsoft iSCSI Initiator 2.08 or later (Certified) Microsoft iSCSI Initiator (Certified) Microsoft iSCSI Initiator (Certified) Linux Open-iSCSI Initiator (Supported) iSCSI Initiator (Certified) iSCSI Initiator (Certified) iSCSI Initiator (Certified)
All of the current iSCSI sessions will be terminated for all the nodes in the cluster. Initiators cannot establish new sessions until the iSCSI service is re-enabled.
298
iSCSI
1. Click File System Management > iSCSI > Settings. 2. In the iSCSI Service area, set the service state that you want: l If the service is disabled, you can enable it by clicking Enable.
l
iSNS server port: Type the iSNS server port number. The default port number is 3205.
3. Click Test connection to validate the iSNS configuration settings. If the connection to the iSNS server fails, check the iSNS server address and the iSNS server port number. 4. Click Submit. 5. Change the service to the state that you want: l If the service is disabled, you can enable it by clicking Enable. Enabling the service allows OneFS to register information about iSCSI targets.
l
If the service is enabled, you can disable it by clicking Disable. Disabling the service prevents OneFS from registering information about iSCSI targets.
The Current Sessions area displays information about each current connection between an initiator and a target, including the client and target IP addresses; node, target, and LUN; operations per second; and the inbound, outbound, and total throughput in bits per second. You can view details about a target by clicking the target name.
iSCSI
2. In the Targets area, click Add target. 3. In the Name field, type a name for the target. The name must begin with a letter and can contain only lowercase letters, numbers, and hyphens (-). 4. In the Description field, type a descriptive comment for the target. 5. In the Default path field, type the full path of the directory, beginning with /ifs, where the logical unit number (LUN) directory is created, or click Browse to select a directory. This directory is used only if no other directory is specified during LUN creation or if a LUN is not created. The directory must be in the /ifs directory tree. The full path to the directory is required, and wildcard characters are not supported. 6. Add one or more SmartConnect pools for the target to connect with. This setting overrides any global default SmartConnect pools that are configured for iSCSI targets. a. For the SmartConnect pool(s) setting, click Edit list. b. Move pools between the Available Pools and Selected Pools lists by clicking a pool and then clicking the right or left arrow. To remove all selected pools at once, click clear. c. Click OK. 7. Click Submit. 8. Optional: In the Initiator Access Control area, enable and configure the settings for initiator access control. a. Click Enable to restrict target access to initiators that are added to the initiator access control settings. b. Click Add initiator. c. In the Initiator name field, type the name of the initiator that you want to allow to access this target, or click Browse to select from a list of initiators. An initiator name must begin with an iqn. prefix. d. Click OK. To continue adding initiators, click OK and add another. When you are finished adding initiators, click OK. 9. Optional: In the CHAP Authentication area, enable and configure Challenge-Handshake Authentication Protocol (CHAP) settings. If CHAP authentication is enabled and the CHAP secrets list is empty, no initiators can access the target. a. Click Enable to require initiators to authenticate with the target. b. Click Add username. c. In the Username field, type the name that the initiator will use to authenticate with the target. You can specify an initiator's iSCSI qualified name (IQN) as the username. Depending on whether you specify an IQN, valid usernames differ in the following ways: If you specify an IQN as the username, the Username value must begin with an iqn. prefix. The characters that are allowed after the iqn. prefix are alphanumeric characters, periods (.), hyphens (-), and colons (:). All other usernames can use alphanumeric characters, periods (.), hyphens (-), and underscores (_).
300
iSCSI
CHAP usernames and passwords are case-sensitive. d. In the Secret and Confirm secret fields, type the secret that the initiator will use to authenticate with the target. A CHAP secret must be 12 to 16 characters long and can contain any combination of letters, numbers, and symbols. e. Click OK. 10. Click Submit.
l l
Changing the default path does not affect existing logical units. Changing the security settings does not affect existing connections.
4. Click Submit.
iSCSI
logical units that are associated with the target as well as the settings for initiator access control and authentication. 1. Click File System Management > iSCSI > Targets & Logical Units. 2. In the Targets area, click the name of a target. 3. Review the following sections for information on the target. To modify these settings, click Edit target. l Target Details: Displays the target name, IQN, description, default path, capacity, and SmartConnect pool settings. The name and IQN cannot be modified.
l
Logical Units: Displays any logical units that are contained in the target. You can add or import a logical unit, or manage existing logical units. You can also select the columns to display or hide. Allowed Initiators: Displays the target's initiator access control status, and lists the names of any initiators that are allowed to access the target when access control is enabled. CHAP Authentication: Displays the target's CHAP authentication status, and lists all user:secret pairs for the target.
If access control is enabled, click Disable to allow all initiators access to the target. If you disable access control, the list of allowed initiators is ignored.
302
iSCSI
If you remove all of the allowed initiators for a target and access control is enabled, the target will deny new connections until you disable access control. Removing an allowed initiator for a target does not affect the initiator's access to other targets.
1. Click File System Management > iSCSI > Targets & Logical Units. 2. In the Targets area, under Actions, click Edit for the target that you want to modify. 3. In the Initiator Access Control area, under Actions, click Delete for the initiator that you want to remove from the access list. 4. In the confirmation dialog box, click Yes.
303
iSCSI
If you specify an IQN as the username, the Username value must begin with an iqn. prefix. The characters that are allowed after the iqn. prefix are alphanumeric characters, periods (.), hyphens (-), and colons (:). All other usernames can use alphanumeric characters, periods (.), hyphens (-), and underscores (_). CHAP usernames and passwords are case-sensitive.
5. In the Secret and Confirm secret fields, type the secret that the initiator will use to authenticate with the target. A CHAP secret must be 12 to 16 characters long and can contain any combination of letters, numbers, and symbols. 6. Click OK.
304
iSCSI
If CHAP authentication is enabled, click Disable to stop authenticating initiators with the target. If CHAP authentication is disabled, the CHAP secrets list is ignored.
iSCSI
4. From the Target list, select the target that will contain the logical unit. 5. Select one of the LUN number options. l To assign the next available number to the logical unit, click Automatic. This is the default setting.
l
To manually assign a number to the logical unit, click Manual and then, in the Number field, type an integer value. The value must be within the range 0-255 and must not be assigned to another logical unit within the target.
By default, the LUN number forms part of the directory name that is created for storing the LUN data. 6. To manually specify the path where the LUN directory is created, in the Path field, type the full path of the directory, beginning with /ifs, or click Browse to select the directory. The directory must be in the /ifs directory tree. You must specify the full path to the directory, and wildcard characters are not allowed. The default path is /ifs/iscsi/ ISCSI.LUN.<TargetName>.<LUNnumber>, where <TargetName> is the Target value and <LUNnumber> is the LUN number. 7. In the Size field, specify the LUN capacity by typing an integer value and then selecting a unit of measure from the list (MB, GB, or TB). The minimum LUN size is 1 MB. The maximum LUN size is determined by the OneFS file system. After you create a LUN, you can increase its size, but you cannot decrease it. 8. Select one of the Provisioning options. l To specify that blocks are unallocated until they are written, click Thin provision.
l
To immediately allocate all the blocks, click Pre-allocate space. This is the default setting. Allocation of all the blocks for a large LUN can take hours or even days.
9. Select one of the LUN access options. l To make the LUN accessible, click Online. This is the default setting.
l
10. Select one of the Write access options. l To allow iSCSI initiators to write to the LUN, click Read-Write. This is the default setting.
l
11. Under Protection Settings, from the Disk pool list, select the disk pool to contain the logical unit. 12. From the SSD strategy list, select to specify a strategy to use if solid-state drives (SSDs) are available. l Metadata read acceleration (Recommended): Writes metadata and all user data on hard disk drives (HDDs) and additionally creates a mirror backup of the data on an SSD. Depending on the global namespace acceleration setting, the SSD mirror may be an extra mirror in addition to the number required to satisfy the protection level.
l
Metadata read/write acceleration with performance redundancy (Requires more SSD space): Writes all metadata on an SSD and writes all user data on HDDs. Data on SSDs (Requires most SSD space): Similar to metadata acceleration, but also writes one copy of the file's user data (if mirrored) or all of the data (if not mirrored) on SSDs. Regardless of whether global namespace acceleration is
306
iSCSI
enabled, any SSD blocks reside on the file's target pool if there is room. This SSD strategy does not create additional mirrors beyond the normal protection level.
l
Avoid SSDs (Reduces performance): Never uses SSDs; writes all associate file data and metadata to HDDs only.
13. From the Protection level list, select a protection policy for the logical unit. Select Use iSCSI default (2x), which is the recommended setting for best performance, or one of the mirrored options, such as 2x to 8x. 14. Select one of the Write Cache options. l To prevent write caching for files that contain LUN data, click Disabled. This is the recommended setting for LUNs.
l
To allow write caching for files that store LUN data, click Enable.
The Write Cache option controls whether file writes are sent to the coalescer or the endurant cache. With Write Cache disabled, which is the default and recommended setting, all file writes are sent to the endurant cache. The endurant cache is a committed data guarantee. If Write Cache is enabled, all file writes are sent to the coalescer. Write caching can improve performance, but can lead to data loss if a node loses power or crashes while uncommitted data is in the write cache. 15. Select one of the Data access pattern options. l To select a random access pattern, click Random. This is the recommended setting for LUNs.
l l
To select a concurrent access pattern, click Concurrency. To select a streaming access pattern, click Streaming. Streaming access patterns can improve performance in some workflows.
307
iSCSI
Result A snapshot of the source LUN is created. The clone LUN is then created by copying the LUN data from the snapshot. After completing the copy, the snapshot is deleted. The copy process may take several hours to complete for large LUNs if the source LUN has a pre-allocated provisioning policy. The copy process may also take several minutes for thinly provisioned LUNs that are significantly used. A snapshot of the source LUN is created. The clone LUN is configured to reference the data from the snapshot. The snapshot is deleted when the clone is deleted. A snapshot of the source LUN is created. The system then creates a clone LUN that references data from the snapshot. The clone LUN is created by copying the LUN data from the snapshot. The copy process may take several minutes to complete for large LUNs if the source LUN has a pre-allocated provisioning policy. The copy process may also take several minutes for thinly provisioned LUNs that are heavily used. The clone LUN is configured to reference the data from the same snapshot that the source LUN references. The underlying snapshot is not deleted when a LUN is deleted unless the LUN being deleted is the last LUN referencing the snapshot. The clone LUN is configured to reference the data from the same snapshot that the source LUN references. The underlying snapshot is not deleted when a LUN is deleted unless the LUN being deleted is the only LUN referencing the snapshot. A snapshot of the source LUN is created. The clone LUN is then created by copying the LUN data from the snapshot. After completing the copy, the snapshot is deleted. The copy process may take several minutes to complete for large LUNs if the source LUN has a pre-allocated provisioning policy. The copy process may also take several minutes for thinly provisioned LUNs that are heavily used
Normal
Snapshot
Normal
Clone
Snapshot
Normal
Snapshot
Snapshot
Snapshot
Clone
Clone
Normal
308
iSCSI
Result Not allowed. A clone of the clone LUN is created. The clone LUN is configured to reference data from the snapshot.
Modify a LUN Delete a LUN Migrate a LUN to another target Import a LUN View LUN settings
iSCSI
4. Click one of the To LUN number options. l To assign the next available number to the logical unit, click Automatic. This is the default setting.
l
To manually assign a number to the logical unit, click Manual and then, in the Number box, type an integer value. The value must be within the range 0-255 and must not be assigned to another logical unit.
5. To configure the path where the LUN directory is created, in the To path box, type the full path of the directory, or click Browse to select the directory. If a path is not specified, the LUN directory is unchanged from the original directory where that LUN was created. 6. Click Submit.
To manually assign a number to the logical unit, click Manual, and then in the Number field, type an integer value. The value must be within the range 0-255 and must not be assigned to another logical unit.
7. Select one of the LUN access options. l To make the LUN accessible, click Online. This is the default setting.
l
8. Select one of the Write access options. l To allow iSCSI initiators to write to the LUN, click Read-Write. This is the default setting.
l
9. Select one of the caching options. l To allow write caching for files storing LUN data, click Enabled.
l
To prevent write caching for files storing LUN data, click Disabled.
310
iSCSI
Target: Displays the name of the iSCSI target that contains the logical unit. You can modify the target by using the move operation. Description: Displays an optional description for the logical unit. You can modify the description by clicking Edit LUN. Type: Displays the LUN type (normal, clone, or snapshot). You cannot modify this setting. Size: Displays the LUN capacity. You can increase the size of normal or snapshot LUNs by clicking Edit LUN, but you cannot decrease the size. You cannot modify the size of snapshot LUNs. Status: Displays the connection status (online or offline) and write access permissions (read-only or read-write) of the LUN. You can modify write-access settings for normal or clone LUNs by clicking Edit LUN. You cannot modify writeaccess settings for snapshot LUNs. Path: Displays the path to the directory where the LUN files are stored. You can change the path for normal or snapshot LUNs by using the move operation. You cannot modify the path for snapshot LUNs. Disk pool: Displays the disk pool of the LUN. You can modify the disk pool by clicking Edit LUN. Protection level: Displays the mirroring level (such as 2x, 3x, 4x, and so on) or FlexProtect protection policy for the LUN. You can modify the protection policy for normal or clone LUNs by clicking Edit LUN. You cannot modify these settings for snapshot LUNs. Write Cache: Displays whether SmartCache is enabled or disabled. You can change this setting for normal or clone LUNs by clicking Edit LUN. You cannot modify these settings for snapshot LUNs. Data access pattern: Displays the access pattern setting (Random, Concurrency, or Streaming) for the LUN. You can change the access pattern for normal or clone LUNs by clicking Edit LUN. You cannot modify these settings for snapshot LUNs. SCSI name: Displays the iSCSI qualified name (IQN) of the LUN. You cannot modify this setting. EUI: Displays the extended unique identifier (EUI), which uniquely identifies the LUN. You cannot modify this setting. NAA: Displays the LUN's T11 Network Address Authority (NAA) namespace. You cannot modify this setting. Serial number: Displays the serial number of the LUN. You cannot modify this setting.
311
iSCSI
312
OneFS integrates with VMware infrastructures, including vSphere, vCenter, and ESXi. VMware integration enables you to view information about and interact with Isilon clusters through VMware applications. OneFS interacts with VMware infrastructures through VMware vSphere API for Storage Awareness (VASA) and VMware vSphere API for Array Integration (VAAI). OneFS integrates with VMware vCenter through the Isilon for vCenter plug-in. The Isilon for vCenter plug-in enables you to locally backup and restore virtual machines on an Isilon cluster. For more information about Isilon for vCenter, see the following documents:
u u u u u u u
Isilon for vCenter Release Notes Isilon for vCenter Installation Guide Isilon for vCenter User Guide
VASA...................................................................................................................314 VAAI....................................................................................................................315 Configuring VASA support...................................................................................315 Disable or re-enable VASA...................................................................................316
VMware integration
313
VMware integration
VASA
OneFS communicates with VMware vSphere through VMware vSphere API for Storage Awareness (VASA). VASA support enables you to view information about Isilon clusters through vSphere, including Isilon-specific alarms in vCenter. VASA support also enables you to integrate with VMware profile driven storage by providing storage capabilities for Isilon clusters in vCenter. For OneFS to communicate with vSphere through VASA, your VMware environment must include ESXi 5.0 or later hypervisors. To configure VASA support, you must access the cluster through the root account. Because SmartLock compliance mode disables root access to the cluster, the cluster must not be running in compliance mode.
Description There is not enough available space on the cluster to allocate space for writing data to thinly provisioned LUNs. If this condition persists, you will not be able to write to the virtual machine on this cluster. To resolve this issue, you must free storage space on the cluster.
Archive The Isilon cluster is composed of Isilon NL-Series nodes. The cluster is configured
for maximum capacity.
Performance The Isilon cluster is composed of Isilon i-Series, Isilon X-Series, or Isilon SSeries nodes. The cluster is configured for maximum performance. The Isilon I-Series and X-Series nodes contain Solid State Drives (SSDs). If a cluster is composed of i-Series, X-Series , or S-Series nodes, but does not contain SSDs, the cluster is recognized as a capacity cluster.
Capacity The Isilon cluster is composed of Isilon X-Series nodes that do not contain SSDs. The cluster is configured for a balance between performance and capacity. Hybrid The Isilon cluster is composed of nodes associated with two or more storage
capabilities. For example, if the cluster contained both Isilon S-Series and NL-Series nodes, the storage capability of the cluster is displayed as Hybrid.
314
VMware integration
VAAI
OneFS uses VMware vSphere API for Array Integration (VAAI) to support offloading specific virtual machine storage and management operations from VMware ESXi hypervisors to an Isilon cluster. VAAI support enables you to accelerate the process of creating virtual machines and virtual disks. For OneFS to interact with your vSphere environment through VAAI, your VMware environment must include ESXi 5.0 or later hypervisors. If you enable VAAI capabilities for an Isilon cluster, when you clone a virtual machine residing on the cluster through VMware, OneFS clones the files related to that virtual machine. For more information on file clones, see Clones.
Hardware Assisted Locking Full Copy Block Zeroing OneFS does not support the thin provisioning block reclaim mechanism.
Enable VASA
You must enable an Isilon cluster to communicate with VMware vSphere API for Storage Awareness (VASA) by enabling the VASA daemon. 1. Open a secure shell (SSH) connection to any node in the cluster and log in. 2. Enable VASA by running the following command: isi services isi_vasa_d enable
VMware integration
For more information about exporting a security certificate, see the documentation of your browser. Record the location of where you saved the certificate. You will need this file path when adding the vendor provider in vCenter.
Name Type a name for this VASA provider. Specify as any string. For example, type EMC Isilon Systems. URL Type http://<IPAddress>:8081/vasaprovider, where <IPAddress>
is the IP address of a node in the Isilon cluster.
u u u
Login Type root. Password Type the password of the root user. Certificate location Type the file path of the vendor provider certificate for this
cluster.
3. Disable or enable the VASA daemon by running one of the following commands: l isi services isi_vasa_d disable
l
316
The File System Explorer is a web-based interface that enables you to manage the content stored on the cluster. You can use the File System Explorer to navigate the Isilon file system (/ifs), add directories, and manage file and directory properties including data protection, I/O optimization, and UNIX permissions. Isilon file system directory permissions are initially set to allow full access for all users. Any user can delete any file, regardless of the permissions on the individual file. Depending on your environment, you want to establish permission restrictions through the File System Explorer. You can view and configure file and directory properties from within Windows clients that are connected to the cluster. However, because Windows and UNIX permissions differ from one another, you must be careful not to make any unwanted changes that affect file and directory access. The File System Explorer displays up to 1000 files in a directory. If more than 1000 files exist within a directory, the files are displayed without additional information, such as file size and last modified date.
u u u u u
Browse the file system........................................................................................318 Create a directory................................................................................................318 Modify file and directory properties.....................................................................318 View file and directory properties........................................................................318 File and directory properties................................................................................319
317
You can expand and collapse directories in the Directories pane. The contents of the selected directory are displayed in the right pane. You can view the contents of another directory by clicking the directory in the Directories pane.
Create a directory
You can create a directory under /ifs through the File System Explorer. 1. Navigate to File System Management > File System Explorer. 2. In the Directories pane, specify where you want to create the directory. 3. Click Add Directory. 4. In the New Directory Properties dialog box, in the Directory name field, type a name for the directory. 5. From the User list, select the owner of the directory. 6. From the Group list, select the group for the directory. 7. From the Permissions table, specify the basic permissions for the directory. 8. Click Submit.
318
Settings management Specifies whether protection settings are managed manually or by SmartPools. If you modify either or both protection settings, this property automatically refreshes to Manually managed. If you specify Managed by SmartPools, the protection settings will automatically refresh to match the SmartPools specifications the next time the SmartPools job is run. Disk pool The disk pool whose protection policy is applied if SmartPools is configured to manage protection settings. This property is available only if SmartPools is licensed and enabled on the cluster. SSD The SSD strategy that will be used for user data and metadata if solid-state drives
(SSDs) are available. The following SSD strategies are available:
u
Metadata acceleration OneFS creates a mirror backup of file metadata on an SSD and writes the rest of the metadata plus all user data to hard disk drives (HDDs). Depending on the global namespace acceleration setting, the SSD mirror might be an extra mirror in addition to the number required to satisfy the protection level. Avoid SSDs OneFS does not write data or metadata to SSDs. OneFS writes all data and metadata to HDDs only. Data on SSDs Similar to metadata acceleration, OneFS creates a mirror backup of file metadata on an SSD and writes the rest of the metadata plus all user data to hard disk drives. However, OneFS also writes one copy of the file user data (if mirrored) or all of the data (if not mirrored) to SSDs. All SSD blocks reside on the file target pool if there is adequate space available, regardless of whether global namespace acceleration is enabled. OneFS does not create additional mirrors beyond the normal protection level.
Protection level The FlexProtect or data mirroring protection policy for this file or
directory. If SmartPools is licensed and enabled on the cluster, the default protection policy for files and directories is inherited from the specified disk pool.
SmartCache Specifies whether write caching with SmartCache is enabled for this file or
directory.
Data access pattern The optimization settings for accessing data. The following data
access patterns are available:
u u
Concurrency File or directory is optimized to support many clients simultaneously. Streaming File or directory is optimized for high-speed streaming of a single file. For
example, this pattern can be useful if a single client needs to read very quickly from a single file.
File and directory properties
319
The default data access pattern of iSCSI LUNs is the random access pattern. The default data access pattern of other files and directories is the concurrent access pattern.
UNIX Permissions
u u u
User The owner of the file or directory. Group The group of the file or directory. Permissions The basic permissions for the file or directory.
320