STRSW Ilt Ontapadm Studentguide
STRSW Ilt Ontapadm Studentguide
Student Guide
Content Version 1.0
NETAPP UNIVERSITY
Student Guide
Course ID: STRSW-ILT-ONTAPADM
Catalog Number: STRSW-ILT-ONTAPADM-SG
COPYRIGHT
© 2016 NetApp, Inc. All rights reserved. Printed in the U.S.A. Specifications subject to change without notice.
No part of this document covered by copyright may be reproduced in any form or by any means—graphic, electronic, or mechanical,
including photocopying, recording, taping, or storage in an electronic retrieval system—without prior written permission of NetApp, Inc.
TRADEMARK INFORMATION
NetApp, the NetApp logo, Go Further, Faster, ASUP, AutoSupport, Campaign Express, Clustered Data ONTAP, Customer Fitness,
CyberSnap, Data ONTAP, DataFort, FilerView, Fitness, Flash Accel, Flash Cache, Flash Pool, FlashRay, FlexArray, FlexCache,
FlexClone, FlexPod, FlexScale, FlexShare, FlexVol, GetSuccessful, LockVault, Manage ONTAP, Mars, MetroCluster, MultiStore,
OnCommand, ONTAP, ONTAPI, RAID DP, SANtricity, SecureShare, Simplicity, Simulate ONTAP, SnapCenter, Snap Creator,
SnapCopy, SnapDrive, SnapIntegrator, SnapLock, SnapManager, SnapMirror, SnapMover, SnapProtect, SnapRestore, Snapshot,
SnapValidator, SnapVault, StorageGRID, Tech OnTap, Unbound Cloud, and WAFL are trademarks or registered trademarks of
NetApp, Inc. in the United States and/or other countries.
Other product and service names might be trademarks of NetApp or other companies. A current list of NetApp trademarks is available
on the Web at http://www.netapp.com/us/legal/netapptmlist.aspx.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Introductions
I am Marc. I am a NetApp partner selling to
Enterprise customers in the medical field…
Take time to get to know one another. If you are participating in a NetApp Virtual Live class, your instructor will ask you
to use the chat window or a conference connection to speak. If you are using a conference connection, unmute your line to
speak and be sure to mute again after you speak.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP SMB
Administration
Welcome ONTAP NAS
Fundamentals
ONTAP NFS
Administration
Foundational Intermediate
© 2016 NetApp, Inc. All rights reserved. 7
The ONTAP 9 Data Management Software learning path consists of multiple courses that focus on particular topics.
Fundamental courses build knowledge as you progress up the foundational column and should therefore be taken in the
order shown. Likewise, administration courses also build knowledge as you progress up the intermediate column, but they
require the prerequisite foundational knowledge.
You can navigate the learning path in one of three ways:
Complete all of the fundamental courses and then progress through the administration courses. This navigation is the
recommended progression.
Take a fundamental course and then take its complementary administration course. The courses are color-coded to
make complementary courses easier to identify (green=cluster topics, blue=protocol topics, and orange=data
protection topics).
Take the course or courses that best fit your particular needs. For example, if you manage only SMB file shares, you
can take ONTAP NAS Fundamentals and then take ONTAP SMB Administration. Most courses require some
prerequisite knowledge. For this example, the prerequisites are ONTAP Cluster Fundamentals and ONTAP Cluster
Administration.
The “you are here” indicator shows where this course appears in the ONTAP learning path. You should take ONTAP
Cluster Fundamentals in preparation for this course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Course
Agenda Afternoon
Module 3: Management
Module 4: Network Management
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Course
Agenda Afternoon
Module 7: Storage Efficiency
Module 8: NAS Protocols
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Course
Agenda Afternoon
Module 11: Data Protection Features
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
NetappU NetappU
Sunnyvale Research
Triangle
Park
Launch your exercise equipment kit from your laptop or from the classroom desktop. To connect to your exercise
equipment, use Remote Desktop Connection or the NetApp University portal.
The Windows 2012 Server is your Windows domain controller for the LEARN windows domain. The Windows Server
hosts the domain DNS server.
Your exercise equipment consists of several servers:
A two-node NetApp ONTAP cluster
A one-node ONTAP cluster
A CentOS Linux server
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 15 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
If you encounter an issue, notify your instructor immediately so that it can be resolved promptly.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The NetApp University Overview page is your front door to learning. Find training that fits your learning map and your
learning style, learn how to become certified, link to blogs and discussions, and subscribe to the NetApp newsletter Tech
OnTap.
http://www.netapp.com/us/services-support/university/index.aspx
The NetApp University Community page is a public forum for NetApp employees, partners, and customers. NetApp
University welcomes your questions and comments.
https://communities.netapp.com/community/netapp_university
The NetApp University Support page is a self-help tool that enables you to search for answers to your questions and to
contact the NetApp University support team. http://netappusupport.custhelp.com
Are you new to NetApp? If so, register for the New to NetApp Support Webcast to acquaint yourself with facts and tips
that can help you have a successful support experience.
http://www.netapp.com/us/forms/supportwebcastseries.aspx?REF_SOURCE=new2ntapwl-netappu
The NetApp Support page is your introduction to all products and solutions support: http://mysupport.netapp.com. Use the
Getting Started link (http://mysupport.netapp.com/info/web/ECMP1150550.html) to establish your support account and
hear from the NetApp CEO. Search for products, downloads, tools, and documentation or link to the NetApp Support
Community (http://community.netapp.com/t5/Products-and-Solutions/ct-p/products-and-solutions).
Join the Customer Success Community to ask support-related questions, share tips, and engage with other users and
experts.
https://forums.netapp.com/
Search the NetApp Knowledgebase to harness the accumulated knowledge of NetApp users and product experts.
https://kb.netapp.com/support/index?page=home
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Colocation Public
Private
Managing data in hybrid cloud architectures can be a challenge. Sometimes the hybrid architecture becomes a collection
of separate, incompatible repositories for data, known as “data silos.” As a solution, the Data Fabric powered by NetApp
connects cloud resources. When clouds are connected by the Data Fabric strategy, IT can draw from the resources of each
cloud. IT can also move data and applications to new cloud services and place every workload on the most appropriate
platform.
Data Fabric is the NetApp vision for the future of data management. Data Fabric is the NetApp architecture for the hybrid
cloud.
Data Fabric seamlessly connects multiple data-management environments across disparate clouds into a cohesive,
integrated whole. Organizations control the management, security, protection, and access of data across the hybrid cloud,
no matter where the data is located. IT has the flexibility to select the right set of resources and the freedom to change the
resources whenever necessary.
For more information about Data Fabric, see the Welcome to Data Fabric video:
http://www.netapp.com/us/campaigns/data-fabric/index.aspx.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
NetApp
ONTAP ONTAP
Colocation Public Cloud Software
Private
ONTAP NetApp ONTAP Select Software
NetApp ONTAP software is known primarily as the storage software that runs on FAS and All Flash FAS controllers.
ONTAP software is the foundation of Data Fabric, the NetApp vision for the future of data management.
As part of the Data Fabric vision, ONTAP software can also run in the cloud as a software storage system; for example,
NetApp ONTAP Cloud for Amazon Web Services (ONTAP Cloud for AWS).
For remote offices that lack space and resources, NetApp ONTAP Select software can be run as a virtual machine on the
VMware vSphere platform. This virtualization enables the distribution of data center management and end-to-end data
protection for the remote office environment.
Although this course focuses on ONTAP clusters, the knowledge is also applicable to ONTAP Cloud and ONTAP Select
software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Services Unify management across all environments and through all layers
Ecosystem Integrate storage systems and data management with application software
Integration frameworks
Data
Deliver a set of capabilities that enable management of and access to data
Management
Storage Increase the performance, availability, durability, scalability, and supportability of
Management system hardware components
Creators of a data fabric system must take a data-centric view of IT infrastructure in every layer of the environment:
platform, transport, storage management, data management, ecosystem integration, and services. The architecture contains
the products and solutions that unbind data from underlying systems so that the data can be accessed across the fabric.
With Data Fabric powered by NetApp, IT architects have many building blocks to choose from at each layer. These
building blocks adhere to the principles of a true Data Fabric strategy.
Data Fabric starts at the platform layer. The platform layer includes the storage systems that are the building blocks for the
endpoints of the Data Fabric strategy. The course focuses on ONTAP software, running on FAS or All Flash FAS
systems, as the endpoint. The course also discusses the features that the other Data Fabric layers contain.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You might wonder, “What is a cluster?” The course examines cluster components individually, but first, consider a high-
level view.
A cluster is one or more FAS or All Flash FAS controllers that run the ONTAP software. In ONTAP terminology, a
controller is called a node. In clusters with more than one node, a cluster interconnect is required so that the nodes appear
as one cluster.
A cluster can be a mix of various FAS and All Flash FAS models, depending on the workload requirements. Nodes can be
added to or removed from a cluster as workload requirements change. For more information about the number and types
of nodes, see the ONTAP 9 Storage Platform Mixing Rules in the NetApp Library.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
For information about specific controller models, see the product documentation on the NetApp Support site, or see the
Hardware Universe at http://hwu.netapp.com/.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Cluster interconnect:
Connection of nodes
Private network
Management network:
For cluster administration
Potential for management and data to be
on a shared Ethernet network
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Nodes have various physical ports that are available for cluster, management, and data traffic. The ports need to be
configured appropriately for the environment.
Ethernet ports can be used directly or can be aggregated by using interface groups. Also, physical Ethernet ports and
interface groups can be segmented by using virtual LANs (VLANs). Interface groups and VLANs are called virtual ports,
which are treated like physical ports.
A logical interface (LIF) represents a network access point to a node in the cluster. A LIF can be associated with a
physical port, an interface group, or a VLAN to interface with the management or data network.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggregate
Physical Layer
RAID Groups of Disks
The ONTAP storage architecture dynamically maps physical storage resources to logical containers.
In ONTAP software, disks are grouped into RAID groups. An aggregate is a collection of physical disk space that
contains one or more RAID groups. Each aggregate has a RAID configuration and a set of assigned disks. The disks,
RAID groups, and aggregates make up the physical storage layer.
Within each aggregate, you can create one or more FlexVol volumes. A FlexVol volume is an allocation of disk space that
is a portion of the available space in the aggregate. A FlexVol volume can contain files or LUNs. The FlexVol volumes,
files, and LUNs make up the logical storage layer.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Disk:
Disk ownership can be assigned to one
controller.
A disk can be used as a spare or added to a
RAID group.
RAID group:
A RAID group is a collection of disks.
RAID groups protect data in the aggregate.
Aggregate:
One or more RAID groups can be used to
form an aggregate.
Data is written across all groups.
One controller owns an aggregate.
© 2016 NetApp, Inc. All rights reserved. 13
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SVM:
Data
Container for data volumes
LIF Access of client data through a LIF
Client Access
Volume:
SVM with FlexVol Logical data container for files or LUNs
Volumes
One or more FlexVol volumes on an
SVM or one scalable infinite volume
on an SVM
LIF:
Representation of the network address
Cluster that is associated with a port
Access to client data
© 2016 NetApp, Inc. All rights reserved. 14
A storage virtual machine (SVM) contains data volumes and LIFs. The data volumes store client data, which is accessed
through a LIF.
A volume is a logical data container that might contain files or LUNs. SVMs might have one or more FlexVol volumes or
a single, scalable infinite volume.
A LIF represents the IP address or worldwide port name (WWPN) that is associated with a port. Data LIFs are used to
access client data.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click Submit. answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A data SVM contains data volumes and LIFs that serve data to clients. Unless otherwise specified, the term SVM refers to
data SVM. In the CLI, SVMs are displayed as “Vservers.” SVMs might have one or more FlexVol volumes or one
scalable infinite volume that can be used as a content repository.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Cluster
When the SVM is created, a root volume is also created, which serves as the NAS client entry point to the namespace that
an SVM provides. NAS client data access depends on the health of the root volume in the namespace. SAN client data
access is independent of the root volume health in the namespace.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
An SVM can contain one or more FlexVol volumes. In a NAS environment, volumes represent the file system where
clients store data. In a SAN environment, a LUN is created in the volumes for a host to access.
In a SAN environment, the host operating system controls the reads and writes for the file system.
Qtrees can be created to partition a FlexVol volume into smaller segments, much like directories. Qtrees can also be used
to manage quotas, security styles, and CIFS opportunistic lock (oplock) settings.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Data LIFs that are assigned a NAS protocol follow slightly different rules than LIFs that are assigned a SAN protocol.
NAS LIFs are created so that clients can access data from a specific SVM. NAS LIFs are multiprotocol and can be
assigned an NFS, CIFS, or both. When the LIF is created, you can manually assign an IP address or specify a subnet so
that the address is assigned automatically. NAS LIFs can fail over or migrate to any node in the cluster.
SAN LIFs are created so that a host can access LUNs from a specific SVM. SAN LIFs are single-protocol and can be
assigned either the FC or iSCSI protocol. When a LIF is assigned the FC protocol, a WWPN is automatically assigned.
When a LIF is assigned the iSCSI protocol, you can either manually assign an IP address or specify a subnet so that the
address is assigned automatically. Although SAN Data LIFs do not fail over, they can be migrated. However, restrictions
exist on migration.
For more information about migrating SAN LIFs, see the ONTAP 9 SAN Administration Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click Submit. answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP
Common Data Services
ONTAP 9 has many deployment options and can be used in different environments. Simply put, “It’s just ONTAP!” After
being deployed—whether on an engineered system, commodity hardware, or the cloud—all ONTAP software is managed
the same way.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
For more details about the supported engineered platform for ONTAP 9, see the Hardware Universe at
http://hwu.netapp.com/.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP
Common Data Services
The current section focuses on software-defined storage or third-party servers that run ONTAP Select software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP Select
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Virtual Machines/
A software-only platform for the private
… cloud:
Hosts
ONTAP on commodity server hardware
File Services Storage provided by local direct-attached
storage (DAS)
ONTAP Data center and remote office solution
Select
Customer problems that are addressed:
Hypervisor Business agility
Commodity Enterprise availability
Server
Efficient and effective protection
DAS …
Support for VMware vSphere and KVM
© 2016 NetApp, Inc. All rights reserved. 31
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click Submit. answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Which two configurations does ONTAP Select software support? (Choose two.)
a. single-node
b. two-node switchless
c. four-node
d. MetroCluster software
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP
Common Data Services
The current section focuses on cloud storage as provided by ONTAP Cloud for AWS.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Cloud Manager
Manage and monitor cloud storage on site or directly in the cloud.
OnCommand System Manager
NetApp OnCommand OnCommand Unified Manager
Management Suite
OnCommand Insight *Amazon Elastic Block
Store (Amazon EBS)
ONTAP Cloud software deploys ONTAP 9 as software in the cloud. ONTAP Cloud software further enables a common
set of data services in the cloud. You can choose to own, lease, or rent on demand. You can explore and test the full power
of ONTAP 9 in the cloud with little risk. NetApp OnCommand Cloud Manager and OnCommand Insight simplify
monitoring, provisioning, and data movement of all ONTAP 9 instances across clouds.
NOTE: Amazon Elastic Block Store (Amazon EBS) provides persistent block-level storage volumes for use with
Amazon Elastic Compute Cloud (Amazon EC2) instances in the AWS cloud. You can find more information about
Amazon EC2 and Amazon EBS at http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/AmazonEBS.html.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP Cloud
ONTAP Cloud Explore ONTAP Cloud Standard
Standard
Hourly Hourly
BYOL
Multiprotocol
Functionality
ONTAP
FlexClone
Premium Data Protection
Bundle
M3.XL, M3.2XL M3.XL, M3.2XL
EC2* Instances M3.XL
R3.XL, R3.2XL R3.XL, R3.2XL
Functionality
AWS
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click Submit. answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Which two options are licensing choices for ONTAP Cloud software?
(Choose two.)
a. node-locked
b. clusterwide
c. pay as you go
d. BYOL
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Storage virtual machine (SVM) Owns a set of logical storage and network resources
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SVM root volume Serves as the NAS-client entry point to the namespace
Node root volume Contains cluster configuration data and network resources
Data logical interface (LIF) Provides a network access point for clients or
hosts to access data in an SVM
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Single-Node
MetroCluster
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Some features and operations are not supported for single-node clusters. Because single-node clusters operate in a
standalone mode, storage failover (SFO) and cluster high availability are unavailable. If the node goes offline, clients
cannot access data that is stored in the cluster. Also, any operation that requires more than one node cannot be performed.
For example, you cannot move volumes, perform most copy operations, or back up cluster configurations to other nodes.
Lastly, an infinite volume must contain aggregates from at least two nodes. Therefore, Infinite Volume is not supported on
single-node clusters.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
HA pairs provide hardware redundancy that is required for nondisruptive operations (NDO) and fault tolerance. The
hardware redundancy gives each node in the pair the software functionality to take over and return partner storage. The
features also provide the fault tolerance required to perform NDO during hardware and software upgrades or maintenance.
A storage system has various single points of failure, such as certain cables or hardware components. A high-availability
(HA) pair greatly reduces the number of single points of failure. If a failure occurs, the partner can take over and continue
serving data until the failure is fixed. The controller failover function provides continuous data availability and preserves
data integrity for client applications and users.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Node 1 Node 2
Node 2 Storage
Node 1 Storage
Primary Connection
Standby Connection
NOTE: Multipath high-availability (MPHA)
redundant storage connections are not shown.
© 2016 NetApp, Inc. All rights reserved. 10
Each node in an HA pair requires an HA interconnect between the controllers and connections to both the node’s disk
shelves and the partner node’s shelves.
The example here uses a standard FAS8080 EX HA pair with native DS4246 disk shelves. The controllers in the HA pair
are connected through an HA interconnect that consists of adapters and cables. When the two controllers are in the same
chassis, adapters and cabling are not required because connections are made through an internal interconnection. To
validate an HA configuration, use the Hardware Universe.
For multipath high-availability (MPHA) support, redundant primary and secondary connections are also required. For
simplicity, the connections are not shown on the slide. MPHA is required on all HA pairs except some FAS2500 series
system configurations, which use a single-path high HA configuration and lack redundant standby connections.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In a two-node switchless
cluster, ports are Onboard
Cluster Interconnect Ports
connected between nodes. *GbE=Gigabit Ethernet
10-GbE*
4 x Ports on a FAS8060
In clusters that have more than one node, a cluster interconnect is required. The example here shows a FAS8060 system
that has two controllers installed in the chassis. Each controller has a set of four onboard 10-Gigabit Ethernet (10-GbE)
ports that can be used to connect to the cluster interconnect.
In a two-node switchless cluster, a redundant pair of ports is cabled together as shown on the slide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Inter-Switch
Links (ISLs)
Cluster Interconnect
Cluster Switch Cluster Switch
If your workload requires more than two nodes, the cluster interconnect requires switches. The cluster interconnect
requires two dedicated switches for redundancy and load balancing. Inter-Switch Links (ISLs) are required between the
two switches. There should always be at least two cluster connections, one to each switch, from each node. The required
connections vary, depending on the controller model.
After the cluster interconnect is established, you can add more nodes as your workload requires.
For more information about the maximum number and models of controllers supported, see the ONTAP Storage Platform
Mixing Rules in the NetApp Library.
For more information about the cluster interconnect and connections, see the ONTAP Network Management Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The MetroCluster continuous-availability and disaster recovery software uses mirroring to protect the data in a cluster.
The MetroCluster software provides disaster recovery through one MetroCluster command. The command activates the
mirrored data on the survivor site.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You must connect the controller, disks, and cables first. Powering on should start with the networking, then disk shelves,
and finally the controllers.
If the system is new and does not require a software upgrade (or downgrade), simply start the setup process.
If the system requires an upgrade or downgrade, install the software first. After the software installation is complete,
initialize the disks. Initialization takes some time.
When the system boots completely, run a setup procedure to set up and configure the system or cluster. After the
configuration is complete, you can create storage resources.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Connect controllers to disk shelves. Verify that shelf IDs are set properly.
If required for your controller type, connect NVRAM HA cable between partners. The connections can be through the
chassis, 10-GbE, or InfiniBand, depending on your storage controllers.
Connect controllers to networks. If present, connect any tape devices. (You can connect tape devices later.)
Connect controllers and disk shelves to power.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
HA interconnects connect the two nodes of each HA pair for all controllers. The connections are internally provided over
the backplane in the chassis of a dual-controller configuration. For chassis with single controllers, a dedicated InfiniBand
or 10-GbE link is required, depending on the model and enclosure. Visit the NetApp Support site to see the appropriate
hardware configuration guide for your model storage controller.
The following types of traffic flow over the HA interconnect links:
Failover: The directives are related to performing SFO between the two nodes, regardless of whether the failover is:
– Negotiated (planned and in response to an administrator request)
– Not negotiated (unplanned and in response to an improper system shutdown or booting)
Disk firmware: Nodes in an HA pair coordinate the update of disk firmware. While one node updates the firmware,
the other node must not perform any I/O to that disk.
Heartbeats: Regular messages demonstrate availability.
Version information: The two nodes in an HA pair must be kept at the same major and minor revision levels for all
software components.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What are the advantages of using MPHA cabling instead of single-path cabling?
Why do I need an alternate control path (ACP)? The FC-connected shelves did
not need an ACP.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The order that is shown is recommended for powering on the hardware devices in a cluster.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Each controller should have a console connection, which is required to get to the firmware and the boot menu; for
example, for the setup, installation, and initialization options. A remote management device connection, although not
required, is helpful if you cannot get to the UI or console. Remote management enables remote booting, the forcing of
core dumps, and other actions.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Remains operational
Device
Storage Controller
Uses the following setup command:
system node service-processor
© 2016 NetApp, Inc. All rights reserved. 29
Some storage system models include an e0M interface. The interface is dedicated to ONTAP management activities. An
e0M interface enables you to separate management traffic from data traffic on your storage system, for better security and
throughput.
To set up a storage system that has the e0M interface, remember the following information:
The Ethernet port that is indicated by a wrench icon on the rear of the chassis connects to an internal Ethernet switch.
Follow the ONTAP setup script.
To manage LAN in environments where dedicated LANs isolate management traffic from data traffic, e0M is the
preferred interface.
Configure e0M separately from the Remote LAN Module (RLM) or SP configuration.
Both configurations require unique IP and MAC addresses to enable the Ethernet switch to direct traffic to either the
management interfaces or the RLM or SP.
For more information on configuring remote support, refer to the ONTAP System Administration Guide and ONTAP
Remote Support Agent Configuration Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Find more information about the service processor in the ONTAP System Administrator Reference.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
After you boot the system, if the node stops at the firmware prompt (which happens if the firmware environment variable
AUTOBOOT is set to false), type boot_primary to enable the node to continue to the boot menu. If AUTOBOOT is set
to true, the node goes straight to the boot menu.
If you use TFTP, beware of older TFTP servers that have limited capabilities and might cause installation failures.
Because all disks are initialized parallel to one another, the time that is required to initialize the disks is based on the size
of the largest disk that is attached to the node, not on the sum capacity of the disks. After the disks are initialized, the
node’s first aggregate and its vol0 volume are created automatically.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
After installing the hardware, you should use System Setup whenever possible to set up the cluster. However, if System
Setup does not support your configuration, you can use the CLI to run the cluster setup wizard.
Before setting up a cluster, you should use a cluster setup worksheet and record the values that you need during the setup
process. Worksheets are available on the NetApp Support website. If you use the System Setup software, enter the
information that you collected on the worksheet as the software prompts you.
If you use the CLI and the cluster setup wizard to create the cluster, you create the cluster on a single node in the cluster.
The wizard helps you to configure the cluster interconnect, create the cluster admin, add feature license keys, and create
the node management interface for the first node. After creating the cluster, you use the node setup wizard to join nodes to
the cluster, one at a time. The node setup wizard helps you to configure each node's node management interface.
After using the CLI to add all nodes, you also need to manually configure a few items. Synchronizing the time ensures
that every node in the cluster has the same time and prevents CIFS and Kerberos failures. You need to decide where to
send event notifications: to an email address, a syslog server, or an SNMP traphost. NetApp also recommends that you
configure the AutoSupport support tool.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You should use System Setup whenever possible to set up the cluster, because System Setup automates much of the
cluster creation.
The software first automatically discovers nodes by using DHCP or a specific range of IP addresses. During the
discovery, the software also identifies the HA pairs for proper failover configuration. The software automatically
configures multiple nodes in a cluster. Unlike with the CLI, with the System Setup software you do not need to create the
cluster on the first node and join the other nodes. No joining is required.
Optionally, you can also configure networking and aggregates, create a data SVM, and configure NFS, CIFS, and iSCSI
protocols. Lastly, you can enable AutoSupport and critical alerts.
For more information about setting up a cluster, see the ONTAP 9 Software Setup Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
After you use System Setup to create the cluster, a link is provided to launch OnCommand System Manager. Log in as
cluster administrator to manage the entire cluster. You manage all cluster resources, the creation and management of
SVMs, access control and roles, and resource delegation.
To log in to the cluster, use the default user name admin and the password that you configured during cluster creation.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can use many tools to create and manage cluster resources. Each tool has advantages and disadvantages.
OnCommand System Manager is a web-based UI that provides a visual representation of the available resources.
Resource creation is wizard-based and adheres to best practices. However, not all operations are available. Some
advanced operations might need to be performed by using commands in the CLI.
You can use the CLI to create and configure resources. Enter commands manually or through scripts. Instead of the
wizards that System Manager uses, the CLI might require many manual commands to create and configure a resource.
Although manual commands give the administrator more control, manual commands are also more prone to mistakes that
can cause issues. One advantage of using the CLI is that the administrator can quickly switch focus without needing to
move through System Manager pages to find different objects.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The cluster has different CLIs or shells for different purposes. This course focuses on the clustershell, which starts
automatically when you log in to the cluster.
Clustershell features include inline help, an online manual, history and redo commands, and keyboard shortcuts. The
clustershell also supports queries and UNIX-style patterns. Wildcards enable you to match multiple values in command-
parameter arguments.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Typing the first two levels of the command directory puts you in the command directory. You can then type a command
from that level or type a fully qualified command from a different command directory.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
At the command line, press the question mark (?) key to show the command directories and commands that are available
at that command level.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Answers:
2. There is not a show command at this level.
3a. The cluster has two nodes.
3b. Both nodes should be healthy and eligible.
4. You are in the cluster command scope
5a. A show command is available.
5b. top or .. will return you to the root of the command directory.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Press the Tab key to show available directories, commands, and parameters or to automatically complete a command (or a
portion of a command). You can also use the Tab key to complete nonambiguous substrings of commands, parameters,
and values.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The clustershell features privilege levels that force administrators to be mindful of commands that can harm the health of
the storage system. The admin privilege level is used for most tasks. Advanced and diagnostic levels are reserved for more
risky functions.
ONTAP provides multiple sets of commands that are based on privilege level. ONTAP offers administrative, advanced,
and diagnostic levels. Use the priv command to set the privilege level.
The administrative level provides access to commands that are sufficient for managing your storage system. The advanced
and diag levels provide access to the same administrative commands, plus additional troubleshooting and diagnostic
commands.
Advanced and diag-level commands should be used only with the guidance of NetApp technical support.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Use the .. command to move up one level in the command hierarchy. Use the top command to move to the top level
of the command hierarchy.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can abbreviate commands and parameters in the clustershell if the abbreviation is unambiguous in the current
context. You can also run commands out of context if the command is not available in any other context.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
System Manager is a graphical management interface that enables you to manage storage systems and storage objects
(such as disks, volumes, and aggregates) and perform management tasks that relate to storage systems. As a cluster
administrator, you can use System Manager to administer the entire cluster and its resources.
System Manager is no longer available as an executable file and is now included with ONTAP as a web service, enabled
by default. System Manager is accessible through a web browser. System Manager for ONTAP 9 has a slightly different
layout than older versions.
System Manager enables you to perform many tasks:
Configure and manage storage objects, such as disks, aggregates, volumes, qtrees, and quotas.
Configure protocols, such as CIFS and NFS, and provision file sharing.
Configure protocols such as FC, FCoE, and iSCSI, for block access.
Create and configure network components such as subnets, broadcast domains, data and management interfaces, and
interface groups.
Set up and manage mirroring and vaulting relationships.
Manage clusters, storage nodes, and SVMs.
Create and configure SVMs, manage storage objects that are associated with SVMs, and manage SVM services.
Monitor and manage HA configurations in a cluster.
Configure SPs to remotely log in, manage, monitor, and administer the node, regardless of the state of the node.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
System Manager has most of the same features as previous versions, with a new layout:
The tabs are relocated from the left side to a row on the top. You do not need to navigate the Cluster, SVM, and
Nodes tabs.
Tabs are ordered by frequency of use.
Tabs are dynamic, depending on licensed features; for example the LUN tab has been added because iSCSI was
licensed.
Quickly allocate resources by clicking the plus sign in the upper-right corner.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Have a roundtable discussion with the class to answer these questions. You should also add any comments about
experiences or “lessons learned” during the exercises that others may find helpful.
If you encounter an issue, notify your instructor immediately so that it can be resolved promptly.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In this module, you learn how to configure key features of NetApp ONTAP software, such as role-based access control
(RBAC), feature licensing, Network Time Protocol (NTP), and the AutoSupport tool. You also learn about policies and
job schedules, which are used throughout this course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The cluster might require initial configuration, depending on the environment. This module discusses access control, date
and time, licenses, jobs, and schedules. Some of the items might already be configured if the cluster was created by using
the System Setup software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
This module focuses on cluster administration. Two types of administrators can manage a cluster.
What an SVM administrator can configure is based on how the cluster administrator has configured the SVM
administrator’s user account.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The admin storage virtual machine (SVM) is used to manage the cluster.
There is only one admin SVM, which represents the cluster. Through the cluster management logical interface (LIF), you
can manage any node, resource, or data SVM.
Unless otherwise specified, as with the admin SVM, the term SVM typically refers to a data-serving SVM, which applies
to both SVMs with FlexVol volumes and SVMs with Infinite Volume. Also, in the CLI, SVMs are displayed as Vservers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can use the default system administration account to manage a storage system, or you can create additional
administrator user accounts to manage administrative access to the storage system.
You might want to create an administrator account for the following reasons:
You can specify administrators and groups of administrators with differing degrees of administrative access to your
storage systems.
You can limit an administrator’s access to specific storage systems by providing an administrative account on only
those systems.
Creating different administrative users enables you to display information about who is performing which commands
on the storage system.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Role 1 Capability:
Admin 1
Capability 1 Includes a command
Capability 2 Role 1 Includes an access level
Role 2 all
readonly
Role 2 Role 3 none
Capability 3
Role:
A named set of capabilities and commands
Role 3 Admin 2 Defined for cluster or SVM administration
Capability 1 Role 1 User:
Authenticated by the cluster
Capability 2 Role 2 Authenticated for administration, not for data
Capability 3
access
Created as cluster or SVM administrators
© 2016 NetApp, Inc. All rights reserved. 8
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software includes administrative access-control roles that can be used to subdivide administration duties for SVM
administration tasks.
The vsadmin role is the superuser role for an SVM. The admin role is the superuser for a cluster.
The vsadmin role grants the data SVM administrator full administrative privileges for the SVM. Additional roles include
the vsadmin-protocol role, the vsadmin-readonly role, and the vsadmin-volume role. Each role provides a unique SVM
administration privilege.
A cluster administrator with the readonly role can grant read-only capabilities. A cluster administrator with the none role
cannot grant capabilities.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Role name
Command directory
Query
Cluster administrators can create access-control roles to apply to cluster or SVM administrators. The roles can grant or
limit authority to perform certain administrative tasks. An access-control role consists of a role name and a command or a
command directory to which the role has access. The role can include an access level (none, readonly, or all) and a query
that applies to the specified command or command directory. The example on the slide creates a role that is named
svm1vols and that grants access to the volume commands but limits access to aggregates that start with the string “aggr7”.
The role is assigned to a user who is named Ken.
After the role is created, you can apply the role to individual administrators:
c1::> security login modify –vserver svm1 –user ken -role svm1vols
c1::> security login role create –vserver svm1 -role svm1vols -cmddirname volume
-query "-aggr aggr7*" -access all
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Active Directory authentication for cluster and SVM administrators provides a dedicated, CIFS-licensed SVM that serves
as a communication tunnel to the administration server. The enhancement satisfies customers who want to use Active
Directory to authenticate their storage and SVM administrators but do not need CIFS data access.
You must also create cluster user accounts for the domain users.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
cliset: Permit the user to create or modify settings by using the clustershell.
cliget: Permit the user to view settings by using the clustershell.
httpset: Permit the user to create or modify settings with management tools that use HTTP.
httpget: Permit the user to view settings with management tools that use HTTP.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The admin SVM is created to manage the cluster and serve data to the cluster
administrators.
a. True
b. False
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
How might multitenancy affect the way that you use RBAC?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Problems can occur when the cluster time is inaccurate. ONTAP software enables you to manually set the time zone, date,
and time on the cluster. However, you should configure the NTP servers to synchronize the cluster time.
To configure the date and time, in NetApp OnCommand System Manager, on the cluster’s system tools Configurations
tab, click the Date and Time link. Click Edit, select the time zone from the Time Zone list, enter the NTP address in the
Time Servers field, and then click Add.
Adding the NTP server automatically configures all the nodes in the cluster, but each node needs to synchronize
individually. The synchronization for all the nodes in the cluster might require a few minutes.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A license is a record of one or more software entitlements. License keys, also known as license codes, enable you to use
certain features or services on your cluster. Each cluster requires a cluster base license key, which you can install either
during or after the cluster setup. Some features require additional licenses. ONTAP feature licenses are issued as
packages, each of which contains one or more features. A package requires a license key, and installing the key enables
you to access all features in the package. ONTAP prevents you from installing a feature license before a cluster base
license key is installed.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Site license:
Single license that enables the feature on the entire cluster
Not carried with nodes that are removed from the cluster
Evaluation license:
Also known as a demo license
Temporary license with an expiration date
Clusterwide and not locked to a node
Standard license: A standard license is issued for a node with a specific system serial number and is valid only for
the node that has the matching serial number. Installing a standard, node-locked license entitles a node, but not the
entire cluster, to the licensed functionality. For the cluster to be enabled, though not entitled, to use the licensed
functionality, at least one node must be licensed for the functionality. However, if only one node in a cluster is
licensed for a feature, and that node fails, then the feature no longer functions on the rest of the cluster until the
licensed node is restarted.
Site license: A site license is not tied to a specific system serial number. When you install a site license, all nodes in
the cluster are entitled to the licensed functionality. The system license show command displays site licenses
under the cluster serial number. If your cluster has a site license and you remove a node from the cluster, the node
does not carry the site license with it, and that node is no longer entitled to the licensed functionality. If you add a
node to a cluster that has a site license, the node is automatically entitled to the functionality that the site license
grants.
Evaluation license: An evaluation license enables you to try certain software functionality without purchasing an
entitlement. If your cluster has an evaluation license for a package and you remove a node from the cluster, the node
does not carry the evaluation license.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
rtp-nau::> license ?
(system license)
add Add one or more licenses
capacity> The capacity directory
clean-up Remove unnecessary licenses
delete Delete a license
entitlement-risk> The entitlement-risk directory
show Display licenses
status> Display license status
© 2016 NetApp, Inc. All rights reserved. 23
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Which two statements about standard license keys are true? (Choose two.)
a. They are node-locked.
b. They are 28 characters long.
c. They require only one license code per cluster per feature.
d. They must be refreshed monthly.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SVMs use policy-based management for many resources. A policy is a collection of rules or properties that the cluster
administrator or SVM administrator create and manage. Policies are predefined as defaults or created to manage various
resources. By default, a policy applies to the current resources and to newly created resources, unless otherwise specified.
For example, Snapshot policies can be used to schedule automatic controller-based Snapshot copies. The policy includes
such things as the schedule or schedules to use and how many copies to retain. When a volume is created for the SVM, the
policy is applied automatically but can be modified later.
The efficiency policy is used to schedule postprocess deduplication operations. The policy might include when and how
long deduplication runs.
The examples are only two of the policies that you encounter in ONTAP. The advantage of policy-based management is
that when you create a policy you can apply the policy to any appropriate resource, either automatically or manually.
Without policy-based management, you would need to enter the settings separately for each individual resource.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A job is any asynchronous task that Job Manager manages. Jobs are typically long-running volume operations such as
copy, move, and mirror. Jobs are placed in a job queue. Jobs run in the background when resources are available. If a job
consumes too many cluster resources, you can stop or pause the job until there is less demand on the cluster. You can also
monitor, view the history of, and restart jobs.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Many tasks, such as volume Snapshot copies, can be configured to run on specified schedules. Schedules that run at
specific times are called cron schedules. The schedules are similar to UNIX cron schedules. Schedules that run at intervals
are called interval schedules.
To manage schedules in System Manager, on the cluster Configuration tab, click the Schedules link. You can create, edit,
or delete schedules.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
How did the cluster behave after you specified the NTP server?
Did the time synchronize immediately?
Have a roundtable discussion with the class to answer these questions. You should also add any comments about
experiences or “lessons learned” during the exercises that others may find helpful.
If you encounter an issue, notify your instructor immediately so that it can be resolved promptly.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Management network:
Cluster administration
Management and data may be on a shared
Ethernet network
Data network:
Management Network One or more networks that are used for
data access from clients or hosts
Ethernet, FC, or converged network
Data Network
© 2016 NetApp, Inc. All rights reserved. 4
In multinode clusters, nodes need to communicate with each other over a cluster interconnect. In a two-node cluster, the
interconnect can be switchless. When you add more than two nodes to a cluster, a private cluster interconnect using
switches is required.
The management network is used for cluster administration. Redundant connections to the management ports on each
node and management ports on each cluster switch should be provided to the management network. In smaller
environments, the management and data networks may be on a shared Ethernet network.
For clients and host to access data, a data network is also required. The data network can be composed of one or more
networks. Depending on the environment, the network might be an Ethernet, FC, or converged network. Data networks
can consist of one or more switches or redundant networks.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Two ISLs
A NetApp ONTAP software cluster is essentially a cluster of high-availability (HA) pairs. Therefore, you need a cluster
network, or cluster interconnect, for all the nodes to communicate with one another. Keep the following principle in mind:
If a node cannot see the cluster interconnect, then the node is not part of the cluster. Therefore, the cluster interconnect
requires adequate bandwidth and resiliency.
The figure shows a four-node cluster and three distinct networks. ONTAP software requires both data and management
connectivity, which can coexist on the same data network.
In multinode configurations, ONTAP software also requires a cluster interconnect for cluster traffic. In a two-node
configuration, the cluster interconnect can be as simple as cabling the two nodes or using switches if expansion is desired.
In clusters of more than two nodes, switches are required. Single-node clusters do not require a cluster interconnect if the
environment does not require high availability and nondisruptive operations (NDO).
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Logical Interface
Logical (LIF)
blue-mgmt blue-data1
Network
Ports
© 2016 NetApp, Inc. All rights reserved. 7
Nodes have physical ports that are available for cluster traffic, management traffic, and data traffic. The ports need to be
configured appropriately for the environment. The example shows Ethernet ports. Physical ports also include FC ports and
Unified Target Adapter (UTA) ports.
Physical Ethernet ports can be used directly or combined by using interface groups (ifgroups). Also, physical Ethernet
ports and ifgroups can be segmented by using virtual LANs (VLANs). VLANs and ifgroups are considered virtual ports
but are treated like physical ports.
Unless specified, the term network port includes physical ports, ifgroups, and VLANs.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
10-Gigabit Ethernet (10-GbE) and Unified Target Adapter 2 (UTA2) ports use either Management Port:
Twinax copper (Cu) or optical (LC) cables, depending on the personality and type of small e0M or Service Processor (SP)
form-factor pluggable (SFP) module.
Console Port
SAS (Also SP)
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Port names consist of two or three characters that describe the port type and location. You must be aware of certain port-
naming conventions on the network interfaces.
Ethernet ports: The first character describes the port type and is always e to represent Ethernet. The second character is
a numeral that identifies the slot in which the port adapter is located; the numeral 0 (zero) indicates that the port is on the
node's motherboard. The third character indicates the port position on a multiport adapter. For example, the port name e0b
indicates the second Ethernet port on the motherboard, and the port name e3a indicates the first Ethernet port on an
adapter in slot 3.
FC ports: The name consists of two characters (dropping the e) but otherwise follows the same naming convention as
Ethernet ports. For example, the port name 0b indicates the second FC port on the motherboard, and the port name 3a
indicates the first FC port on an adapter in slot 3.
UTA ports: A UTA port is physically one port but can pass either Ethernet traffic or FC traffic. Therefore, UTA ports are
labeled with both the Ethernet name and the FC name. For example, the port name e0b/0b indicates the second UTA port
on the motherboard, and the port name e3a/3a indicates the first UTA port on an adapter in slot 3.
NOTE: UTA adapter ports are listed by the only FC label name when you use the ucadmin command, even when the
personality is configured as 10-GbE.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
UTA ports are managed in a similar way and require a reboot to take effect. The adapter must also be offline before any
changes can be made.
When the adapter type is initiator, use the run local storage disable adapter command to bring the
adapter offline.
When the adapter type is target, use the network fcp adapter modify command to bring the adapter offline.
For more information about configuring FC ports, refer to the ONTAP SAN Administration Guide for your release, or
attend the NetApp University SAN Implementation course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
An ifgroup combines one or more Ethernet interfaces, which can be implemented in one of three ways.
In single-mode, one interface is active and the other interfaces are inactive until the active link goes down. The standby
paths are used only during a link failover.
In static multimode, all links are active. Therefore, static multimode provides link failover and load-balancing features.
Static multimode complies with the IEEE 802.3ad (static) standard and works with any switch that supports the
combination of Ethernet interfaces. However, static multimode does not have control packet exchange.
Dynamic multimode is similar to static multimode but complies with the IEEE 802.3ad (dynamic) standard. When
switches that support Link Aggregation Control Protocol (LACP) are used, the switch can detect a loss of link status and
dynamically route data. NetApp recommends that when you configure ifgroups, you use dynamic multimode with LACP
and compliant switches.
All modes support the same number of interfaces per ifgroup, but the interfaces in the group should always be the same
speed and type. The naming syntax for interface groups is the letter “a”, followed by a number, followed by a letter; for
example, a0a.
Vendors might use terms such as link aggregation, port aggregation, trunking, bundling, bonding, teaming, or
EtherChannel.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can create ifgroups for higher throughput, fault tolerance, and elimination of single points of failure (SPOFs).
Manage ifgroups in a similar way, with the exception of the following:
You must name ifgroups by using the syntax a<number><letter>.
You cannot add a port that is already a member of one ifgroup to another ifgroup.
Multimode load-balancing methods include the following:
– mac: Network traffic is distributed on the basis of MAC addresses.
– ip: Network traffic is distributed on the basis of IP addresses.
– sequential: Network traffic is distributed as it is received.
– port: Network traffic is distributed on the basis of the transport layer (TCP/UDP) ports.
For more information about load balancing, please refer to TR-4182: Ethernet Storage Best Practices for ONTAP
Configurations.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can configure ifgroups to add a layer of redundancy and functionality to an ONTAP software environment. You can
also use ifgroups with a failover group to help protect against Layer 2 and Layer 3 Ethernet failures.
A single-mode ifgroup is an active-passive configuration (one port sits idle, waiting for the active port to fail) and cannot
aggregate bandwidth. NetApp advises against using the single-mode type of ifgroup. To achieve the same level of
redundancy, you can use failover groups or one of the two multimode methods.
You might use a static multimode ifgroup if you want to use all the ports in the group to simultaneously service
connections. Static multimode does differ from the type of aggregation that happens in a dynamic multimode ifgroup, in
that no negotiation or automatic detection happens within the group in regard to the ports. A port sends data when the
node detects a link, regardless of the state of the connecting port on the switch side.
You might use a dynamic multimode ifgroup to aggregate bandwidth of more than one port. LACP monitors the ports
on an ongoing basis to determine the aggregation capability of the various ports and continuously provides the maximum
level of aggregation capability achievable between a given pair of devices. However, all the interfaces in the group are
active, share the same MAC address, and load-balance outbound traffic. A single host does not necessarily achieve larger
bandwidth, exceeding the capabilities of any constituent connections. For example, adding four 10-GbE ports to a
dynamic multimode ifgroup does not result in one 40-GbE link for one host. The situation is because of the way that both
the switch and the node manage the aggregation of the ports in the ifgroup. A recommended best practice is to use the
dynamic multimode type of ifgroup so that you can take advantage of all the performance and resiliency functionality that
the ifgroup algorithm offers.
You can use two methods to achieve path redundancy when using iSCSI in ONTAP software: by using ifgroups or by
configuring hosts to use multipath I/O over multiple distinct physical links. Because multipath I/O is required, ifgroups
might have little value.
For more information, refer to TR-4182: Ethernet Storage Best Practices for ONTAP Configurations.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Switch 1
e0a-170
Switch 2
Router
Mgmt
Switch
A port or ifgroup can be subdivided into multiple VLANs. Each VLAN has a unique tag that is communicated in the
header of every packet. The switch must be configured to support VLANs and the tags that are in use. In ONTAP
software, a VLAN ID is configured into the name. For example, VLAN e0a-70 is a VLAN with tag 70 configured on
physical port e0a. VLANs that share a base port can belong to the same or different IP spaces, and the base port can be in
a different IP space than the VLANs that share the base port.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can create a VLAN for ease of administration, confinement of broadcast domains, reduced network traffic, and
enforcement of security policies.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
port ifgrp
port
port port
LIF LIF
ifgrp
NOTE: VLANs and ifgroups cannot be
created on cluster interconnect ports.
port port
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 20 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Did you anticipate the failure of the ifgroup before you removed ports e0a and
e0b from the broadcast domain?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
IPspace
Broadcast Domain
Storage Virtual
Machine (SVM) Subnet
Port
LIF IP Addresses:
192.168.0.101
192.168.0.1 192.168.0.1 – 192.168.0.100
ONTAP software has a set of features that work together to enable multitenancy. An IPspace is a logical container that is
used to create administratively separate network domains. An IPspace defines a distinct IP address space that contains
storage virtual machines (SVMs). The IPspace contains a broadcast domain, which enables you to group network ports
that belong to the same Layer 2 network. The broadcast domain contains a subnet, which enables you to allocate a pool of
IP addresses for your ONTAP network configuration.
When you create a logical interface (LIF) on the SVM, the LIF represents a network access point to the node. You can
manually assign the IP address for the LIF. If a subnet is specified, the IP address is automatically assigned from the pool
of addresses in the subnet, much like the way a Dynamic Host Configuration Protocol (DHCP) server assigns IP
addresses.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Storage SP
Point of
Presence
Default: 192.168.0.0 Company A: 10.0.0.0 Company B: 10.0.0.0
The IPspace feature enables clients from more than one disconnected network to access a storage system or cluster, even
if the clients use the same IP address.
An IPspace defines a distinct IP address space in which virtual storage systems can participate. IP addresses that are
defined for an IPspace are applicable only within the IPspace. A distinct routing table is maintained for each IPspace. No
cross-IPspace traffic routing occurs. Each IPspace has a unique assigned loopback interface. The loopback traffic on each
IPspace is isolated from the loopback traffic on other IPspaces.
Example
A storage SP needs to connect customers of companies A and B to a storage system on the storage SP premises. The
storage SP creates SVMs on the cluster—one per customer—and provides a dedicated network path from one SVM to A’s
network and one from the other SVM to B’s network.
The deployment should work if both companies use nonprivate IP address ranges. However, because the companies use
the same private addresses, the SVMs on the cluster at the storage SP location have conflicting IP addresses.
To overcome the problem, two IPspaces are defined on the cluster—one per company. Because a distinct routing table is
maintained for each IPspace, and no cross-IPspace traffic is routed, the data for each company is securely routed to the
respective network, even if the two SVMs are configured in the 10.0.0.0 address space.
Also, the IP addresses that various configuration files (the /etc/hosts file, the /etc/hosts.equiv file, the /etc/rc file, and so
on) refer to are relative to the IPspace. Therefore, the IPspaces enable the storage SP to use the same IP address for the
configuration and authentication data for both SVMs, without conflict.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
IPspaces are distinct IP address spaces in which SVMs reside. All IPspace names must be unique within a cluster.
If required, you can change the name of an existing IPspace (except for the two system-created IPspaces) by using the
network ipspace rename command.
If you no longer need an IPspace, you can delete it by using the network ipspace delete command.
NOTE: No broadcast domains, network interfaces, or SVMs can be associated with and IPspace that you want to delete.
You cannot delete the system-defined Default and Cluster IPspaces.
You can display the list of IPspaces that exist in a cluster, and you can view the SVMs, broadcast domains, and ports that
are assigned to each IPspace.
After you create an IPspace but before you create the SVMs in the IPspace, you must create a broadcast domain that
defines the ports that will be part of the IPspace.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Broadcast domains
Default
enable you to group Broadcast Domain
network ports that
belong to the same Company A
Layer 2 network. Broadcast Domain
Broadcast domains are often used when a system administrator wants to reserve specific ports for use by a certain client or
group of clients. A broadcast domain should include ports from many nodes in the cluster, to provide high availability for
the connections to SVMs.
The figure shows the ports that are assigned to three broadcast domains in a four-node cluster:
The Default broadcast domain, which was created automatically during cluster initialization, is configured to contain
a port from each node in the cluster.
The Company A broadcast domain was created manually and contains one port each from the nodes in the first HA
pair.
The Company B broadcast domain was created manually and contains one port each from the nodes in the second HA
pair.
The Cluster broadcast domain was created automatically during cluster initialization but is not shown in the figure.
The system administrator created the two broadcast domains specifically to support the customer IPspaces.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You create a broadcast domain to group network ports in a cluster that belongs to the same Layer 2 network. SVMs can
then use the ports.
NOTE: The ports that you plan to add to the broadcast domain must not belong to another broadcast domain.
All broadcast domain names must be unique within an IPspace.
The ports that you add to a broadcast domain can be network ports, VLANs, or ifgroups.
Add ports by using the network port broadcast-domain add-ports command.
If the ports that you want to use belong to another broadcast domain but are unused, use the network port
broadcast-domain remove-ports command to remove the ports from the existing broadcast domain.
The maximum transmission units (MTU) value of the ports that you add to a broadcast domain are updated to the
MTU value that is set in the broadcast domain.
The MTU value must match all the devices that are connected to the Layer 2 network.
If you do not specify an IPspace name, the broadcast domain is created in the Default IPspace.
You can rename or delete broadcast domains that you create but not the system-created Cluster and Default broadcast
domains.
To make system configuration easier, a failover group of the same name is created automatically and contains the same
ports. All failover groups that relate to the broadcast domain are removed when you delete the broadcast domain.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
LIF creation.
Company A 10.1.2.5
A subnet is created Broadcast Domain subnet
to
10.1.2.20
within a broadcast
domain and contains a 10.1.2.5
Company B to
pool of IP addresses Broadcast Domain subnet
10.1.2.100
that belong to the
same Layer 3 subnet.
Subnets are recommended for easier LIF creation.
© 2016 NetApp, Inc. All rights reserved. 26
Subnets enable you to allocate specific blocks, or pools, of IP addresses for your ONTAP network configuration. The
allocation enables you to create LIFs more easily when you use the network interface create command, by
specifying a subnet name instead of specifying IP address and network mask values.
IP addresses in a subnet are allocated to ports in the broadcast domain when LIFs are created. When LIFs are removed,
the IP addresses are returned to the subnet pool and are available for future LIFs.
You should use subnets because subnets simplify the management of IP addresses and the creation of LIFs. Also, if you
specify a gateway when defining a subnet, a default route to that gateway is added automatically to the SVM when a LIF
is created using that subnet.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You create a subnet to allocate, or reserve, specific blocks of IPv4 or IPv6 addresses for ONTAP network configuration.
When creating subnets, note the following:
When adding IP address ranges to a subnet, no IP addresses in the network can overlap (so that different subnets, or
hosts, do not attempt to use the same IP address).
If you do not use subnets or do not specify a gateway when defining a subnet, then you must use the route
create command to add a route to the SVM manually.
The value true can be set for the -force-update-lif-associations option. The command fails if any SP
or network interfaces currently use the IP addresses in the specified range. Setting the value to true associates any
manually addressed interfaces with the current subnet and enables the command to succeed.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Logical LIF
blue-mgmt blue-data1
Physical
Port e2a e3a
Network
Ports
© 2016 NetApp, Inc. All rights reserved. 32
A LIF is associated with a physical port, an ifgroup, or a VLAN. Virtual storage systems—VLANs and SVMs—own the
LIFs. Multiple LIFs belonging to multiple SVMs can reside on a single port.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Data LIFs can have a many-to-one relationship with network ports: Many data IP addresses can be assigned to a single
network port. If the port becomes overburdened, NAS data LIFs can be transparently migrated to different ports or nodes.
Clients know the data LIF IP address but do not know which node or port hosts the LIF. If a NAS data LIF is migrated,
the client might unknowingly be contacting a different node. The NFS mount point or CIFS share is unchanged.
A node can have a maximum of 128 NAS data LIFs (with failover enabled). For SAN LIF maximums, refer to the SAN
Configuration Guide for your ONTAP release.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A LIF is an IP address or worldwide port name (WWPN) that is associated with a physical port. If any component fails,
most LIF types (excluding SAN) can fail over to or be migrated to a different physical port, thereby continuing to
communicate with the cluster.
The underlying physical network port must be configured to the administrative up status.
If you are planning to use a subnet name to allocate the IP address and network mask value for a LIF, the subnet must
already exist.
You can create IPv4 and IPv6 LIFs on the same network port.
You cannot assign NAS and SAN protocols to a LIF.
The supported protocols are CIFS, NFS, FlexCache, iSCSI, and FC.
The data-protocol parameter must be specified when the LIF is created and cannot be modified later.
If you specify none as the value for the data-protocol parameter, the LIF does not support any data protocol.
The home-node parameter is the node to which the LIF returns when the network interface revert
command is run on the LIF.
The home-port parameter is the port or ifgroup to which the LIF returns when the network interface
revert command is run on the LIF.
All the name mapping and host-name resolution services—such as DNS, Network Information Service (NIS),
Lightweight Directory Access Protocol (LDAP), and Active Directory—must be reachable from the data, cluster-
management, and node-management LIFs of the cluster.
A cluster LIF should not be on the same subnet as a management LIF or a data LIF.
When using a subnet to supply the IP address and network mask, if the subnet was defined with a gateway, a default
route to that gateway is added automatically to the SVM when a LIF is created using that subnet.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Why migrate a LIF? It might be necessary for troubleshooting a faulty port or to offload a node whose data network ports
are saturated with other traffic. The LIF fails over if its current node is rebooted.
Unlike storage failover (SFO), LIF failover or migration does not cause a reboot of the node from which the LIF is
migrating. After a LIF is migrated, the LIF can remain on the new node for as long as the administrator wants.
Failover groups for LIFs can be broadcast domain–based or user-defined. You create a failover group of network ports so
that a LIF can automatically migrate to a different port if a link failure occurs on the LIF's current port. The failover group
enables the system to reroute network traffic to other available ports in the cluster.
The ports that are added to a failover group can be network ports, VLANs, or ifgroups.
All the ports that are added to the failover group must belong to the same broadcast domain.
A single port can reside in multiple failover groups.
If you have LIFs in different VLANs or broadcast domains, you must configure failover groups for each VLAN or
broadcast domain.
Failover groups do not apply in SAN iSCSI or FC environments.
You can configure a LIF to fail over to a specific group of network ports by applying a failover policy and a failover
group to the LIF. You can also disable a LIF from failing over to another port. Failover policies can be:
Broadcast-domain-wide: All ports on all nodes in the failover group
System-defined: Only those ports on the LIF's home node and a non-SFO partner
Local-only: Only those ports on the LIF's home node
SFO-partner-only: Only those ports on the LIF's home node and SFO partner
Disabled: Not configured for failover
NOTE: LIFs for SAN protocols do not support failover and so are always set to disabled.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Two types of failover groups exist: those created automatically by the system when a broadcast domain is created, and
those that a system administrator defines.
The ports in the Cluster broadcast domain are used for cluster communication and include all cluster ports from all nodes
in the cluster.
The ports in the Default broadcast domain are used primarily to serve data, but also for cluster and node management.
Failover groups have the same name as the broadcast domain and contain the same ports as the groups in the broadcast
domain.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Custom failover groups can be created for specific LIF failover functionality in the
following circumstances:
The automatic failover groups do not meet your requirements.
You require only a subset of the ports that are available in the broadcast
domain.
You require consistent performance:
For example, create a failover group that consists of only 10-GbE ports, to enable LIFs
to fail over only to high-bandwidth ports.
For example, create a failover group that consists of a set of ports for SnapMirror
software over a WAN.
User-defined failover groups can be created for special failover situations when the default broadcast domain–based
groups do not meet your needs.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Broadcast-domain-wide The LIF fails over to a port in the same Default for cluster-management LIF
broadcast domain as the home port
(including any port from any node in the
failover group).
System-defined The LIF fails over to a port on the home Default for data LIFs
node or a non-SFO partner only.
Recommended for nondisruptive
software updates
Local-only The LIF fails over to a port on the home Default for cluster LIFs, node
node of the LIF only. management LIFs, and intercluster
LIFs
SFO-partner-only The LIF fails over to a port on the home
node or SFO partner only.
Disabled Failover is disabled for the LIF. LIF not configured for failover
© 2016 NetApp, Inc. All rights reserved. 42
The table shows the default policies that should be used in most cases.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The table shows how failover policies and groups work together. Groups include all possible failover targets, whereas
policies limit targets within the group.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What are the benefits of each type of failover group and failover policy type?
When should I use ifgroups or failover groups—or do I need both?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Route tables: System SVMs can own LIFS, and the system SVMs might need route configurations that differ from the
configurations on data SVMs.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can use the optional –metric parameter with the network route create command to specify a hop count
for the route. The default settings for the parameter are 10 for management interfaces, 20 for data interfaces, and 30 for
cluster interfaces. The parameter is used for source-IP address selection of user-space applications such as Network Time
Protocol (NTP).
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggregate
Physical Layer
RAID Groups of Disks
The NetApp ONTAP software storage architecture uses a dynamic virtualization engine, in which data volumes are
dynamically mapped to physical space.
In ONTAP software, disks are grouped into RAID groups. An aggregate is a collection of physical disk space that
contains one or more RAID groups. Each aggregate has a RAID configuration and a set of assigned disks. The disks,
RAID groups, and aggregates make up the physical storage layer.
Within each aggregate, you can create one or more FlexVol volumes. A FlexVol volume is an allocation of disk space that
is a portion of the available space in the aggregate. A FlexVol volume can contain files or LUNs. The FlexVol volumes,
files, and LUNs make up the logical storage layer.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When a disk is inserted into a storage system disk shelf or when a new shelf is added, the controller takes ownership of
the disk by default. In a high-availability (HA) pair, only one controller can own a particular disk, but ownership can be
manually assigned to either controller.
When an aggregate is created or disks are added to an aggregate, the spare disks are used.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software automatically assigns disks to a controller during the initial disk setup and checks occasionally to
determine whether new disks have been added. When the disk is assigned, the disk ownership information is written to the
disk so that the assignment remains persistent.
Ownership can be modified or removed. The data contents of a disk are not destroyed when the disk is marked as
unowned; only the disk-ownership information is erased. Unowned disks that reside on an FC-AL, where owned disks
exist, have ownership information applied automatically to guarantee that all disks on the same loop have the same owner.
Automatic ownership assignment is enabled by default and is invoked at the following times:
Every 5 minutes during normal system operation
10 minutes after the initial system initialization. (The delay enables the person who is configuring the system enough
time to finish the initial disk assignments so that the results of the automatic ownership assignment are correct.)
Whenever you enable automatic ownership assignment
The automatic ownership assignment can also be initiated manually, by using the disk assign command with the
auto parameter.
If your system is not configured to assign ownership automatically, or if your system contains array LUNs, you must
assign ownership manually.
NOTE: NetApp best practice is to unassign only spare disks.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
If you add a spare disk to an aggregate and the spare is larger than the other data disks, then the spare becomes the parity
disk. However, the spare does not use the excess capacity unless another disk of similar size is added. The second largest
additional disk has full use of additional capacity.
Zeroing Used Disks
After you assign ownership to a disk, you can add the disk to an aggregate on the storage system that owns the disk.
Alternatively, you can leave the disk as a spare disk on the storage system. If the disk has been used previously in another
aggregate, you should use the disk zero spares command to zero the disk, to reduce delays when the disk is used.
Zeroing Disks in ONTAP Software
Use the storage disk zerospares command to zero disks in ONTAP software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggregates provide storage to volumes. Aggregates are composed of RAID groups of disks or array LUNs, but not both.
ONTAP software organizes the disks or array LUNs in an aggregate into one or more RAID groups. RAID groups are
then collected into one or two plexes, depending on whether RAID-level mirroring (SyncMirror technology) is in use.
The ONTAP storage architecture contains the following:
Aggregates: Each aggregate contains a plex or plexes, a RAID configuration, and a set of assigned physical disks to
provide storage to the volumes that the aggregate contains.
Plexes: Each plex is associated with an aggregate and contains RAID groups. Typically, an aggregate has only one
plex. Aggregates that use SyncMirror technology have two plexes (plex0 and plex1); plex1 contains a mirror of the
plex0 data.
RAID groups: Each RAID group contains physical disks and is associated with a plex. A RAID group has either a
RAID 4 or NetApp RAID-DP configuration.
Disks: Disks play different roles at different times, depending on the state of the disk. Potential disk states include the
following
– Data
– Parity
– Double-parity
– Spare
– Broken
– Unowned
– Uninitialized (not zeroed)
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A RAID group consists of one or more data disks or array LUNs, across which client data is striped and stored. A RAID
group includes as many as two parity disks, depending on the RAID level of the aggregate that contains the RAID group.
You change the size of RAID groups on a per-aggregate basis. You cannot change the size of an individual RAID group.
When sizing RAID groups of hard disk drives (HDDs) or solid-state drives (SSDs), observe the following guidelines:
RAID groups are composed of the same disk type.
All RAID groups in an aggregate should have the same number of disks.
If you cannot follow the guideline, any RAID group with fewer disks should have only one disk less than the largest
RAID group.
NOTE: The SSD RAID group size can differ from the RAID group size for the HDD RAID groups in a flash pool
aggregate. Usually, you should verify that you have only one SSD RAID group for a flash pool aggregate, to
minimize the number of SSDs that are required for parity.
The recommended range of RAID group size is as follows:
– Between 12 and 20 for SATA HDDs
– Between 20 and 28 for SAS HDDs and SSDs
The reliability and smaller size (faster rebuild times) of performance HDDs can support a RAID group size of up
to 28, if needed.
NetApp recommends that you do not mix 10K-RPM and 15K-RPM disks in the same aggregate.
Mixing 10K-RPM disks with 15K-RPM disks in the same aggregate effectively throttles all disks down to 10K RPM.
Throttling results in longer times for corrective actions such as RAID reconstructions.
Recommendations about spares vary by configuration and situation. For information about best practices for working with
spares, see Technical Report 3437: Storage Subsystem Resiliency Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
RAID 4 group
Single parity for single-disk failure
Data Parity
Minimum of two disks Disks Disk
RAID-DP group (default) RAID-DP
Double parity for double-disk failure
Minimum of three disks
Data Parity dParity
RAID-TEC group (ONTAP 9 and later) Disks Disk Disk
Triple parity for triple-disk failure RAID-TEC
Minimum of four disks
Shorter failed-disk rebuild time for large disks
Data Parity dParity tParity
© 2016 NetApp, Inc. All rights reserved. Disks Disk Disk Disk 12
Understanding how RAID protects your data and data availability can help you administer your storage systems more
effectively.
For native storage, ONTAP software uses NetApp RAID-DP (double-parity) technology or RAID 4 protection to provide
data integrity within a RAID group, even if one or two of the disks fail. Parity disks provide redundancy for the data that
is stored on the data disks. If a disk fails, the RAID subsystem can use the parity disks to reconstruct the data in the failed
disk.
NOTE: NetApp imposes a five-disk minimum for RAID-DP and a four-disk minimum for RAID 4. The minimum is
enforced at the aggregate level, not at the RAID group level.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
RAID 4
In a RAID 4 group, parity is calculated separately for each row. In the example, the RAID 4 group contains seven disks,
with each row containing six data blocks and one parity block.
RAID-DP Technology
In a RAID-DP group, a diagonal parity set is created in addition to the row parity. Therefore, an extra double-parity disk
must be added. In the example, the RAID-DP group contains eight disks, with the double parity calculated diagonally by
using seven parity blocks.
The number in each block indicates the diagonal parity set to which the block belongs.
Each row parity block contains even parity of data blocks in that row, not including the diagonal parity block.
Each diagonal parity block contains even parity of data and row parity blocks in same diagonal.
RAID-TEC Technology
In a RAID-TEC group, an anti-diagonal parity set is created in addition to both the row parity and diagonal parity sets.
Therefore, an extra third-parity disk must be added. In the example, the RAID-TEC group contains nine disks, with the
triple parity calculated anti-diagonally using seven parity blocks.
Seven diagonals (parity blocks) exist, but ONTAP software stores six diagonals (p-1).
The missed diagonal selection is arbitrary. Here, diagonal 6 is missing and is not stored or calculated.
Regarding diagonal numbers, the following guidelines apply:
The set of diagonals collectively span all the data disks and the row parity disk.
Each diagonal misses only one disk, and each diagonal misses a different disk. Each disk misses a different diagonal.
The diagonal sequencing within a given disk starts with the diagonal number that corresponds with the given disk
number. So the first diagonal on disk number 0 is diagonal 0, and the first diagonal on disk N is diagonal N. The
diagonals on the disk wrap around when the end of the diagonal set is reached.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
aggr1
rg0 rg1
Add three disks.
aggr2
rg0 rg1
rg2
Add six disks.
You can add disks to an aggregate to provide more storage to associated volumes. To do so, add available spare disks to
an existing aggregate. When adding disks, consider the size of your RAID groups. Plan to fill complete RAID groups to
maximize the amount of usable space that is gained in comparison to the number of disks that are used for parity. In the
aggr2 example, six disks are added to the aggregate, but only one more data disk adds capacity to the aggregate, compared
to adding three disks.
When adding disks, also consider the following:
Adding disks that the same system owns
Benefits of keeping your RAID groups homogeneous for disk size and speed
Which types of disks can be used together
Checksum rules when disks of more than one checksum type are in use
Adding the correct disks to the aggregate (the disk addition operation cannot be undone)
How to add disks to aggregates from heterogeneous storage
Minimum number of disks to add for best performance
Number of hot spares to provide for protection against disk failures
Requirements for adding disks from multidisk carrier disk shelves
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What is one alternative to adding a few disks to an aggregate when all current
RAID groups are full?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What is the minimum number of disks that are required to create a RAID-TEC
data aggregate (excluding hot spares)?
a. three
b. four
c. five
d. six
e. seven
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
How does our non-high-availability (non-HA) lab environment affect disks and
disk ownership?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
At the storage level, there are two ways to implement Virtual Storage Tier (VST):
The controller-based Flash Cache feature provides acceleration of random-read operations and generally provides the
highest performance solution for file-services workloads.
The Flash Pool feature is implemented at the disk-shelf level, enabling SSDs and traditional HDDs to be combined in
a single ONTAP aggregate. In addition to read caching, Flash Pool technology also provides write caching and is
well-suited for OLTP workloads, which typically have a higher percentage of write operations.
Both VST technologies improve overall storage performance and efficiency and are simple to deploy and operate.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Flash Cache intelligent caching combines software and hardware within NetApp storage controllers to increase system
performance without increasing the disk count. The Flash Cache plug-and-play Peripheral Component Interconnect
Express (PCIe) module requires no configuration to use the default settings, which are recommended for most workloads.
The original Flash Cache module is available in 256-GB, 51-GB, or 1-TB capacities and accelerates performance on all
supported ONTAP client protocols. The Flash Cache controller-based solution is available to all volumes that are hosted
on the controller. A common use case for Flash Cache is to handle VMware boot storms.
Flash Cache 2 is the second generation of Flash Cache performance accelerators. The new architecture of Flash Cache 2
accelerators enables even higher throughput.
For more information, see TR-3832: Flash Cache Best Practice Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Write-cached blocks:
Are written directly to the SSD tier
Are not yet written to the HDD tier
The following blocks are stored in the SSD tier of the flash pool:
Flash pool metadata: All metadata that is associated with the flash pool is stored in the SSD tier of the aggregate.
Read-cached blocks: Read-cached blocks are stored in the SSD tier. Almost all data from the active file system in a
read/write volume is eligible to be read-cached into the SSD tier.
Write-cached blocks: Write-cached blocks are associated with a FlexVol volume that is written directly to the SSD tier
of the aggregate. Only one copy of the block exists. A hard-disk block is reserved for write-cached blocks for an eventual
move into the HDD tier after access to the block ceases.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The SSD RAID group size can be different from the RAID group size for the HDD RAID groups in a Flash Pool
aggregate. Usually, you should ensure that you have only one SSD RAID group for a Flash Pool aggregate, to minimize
the number of SSDs required for parity.
For information about best practices for working with aggregates, see Technical Report 3437: Storage Subsystem
Resiliency Guide.
To see the physical and usable capacity for a specific disk, see the Hardware Universe at hwu.netapp.com.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SSD partitioning for Flash Pool intelligent caching enables customers to group SSDs into a shared resource, which is
allocated to multiple flash pool aggregates. The feature spreads the cost of the parity SSDs over more aggregates,
increases SSD allocation flexibility, and maximizes SSD performance.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Allocation Unit
SSD storage pools provide SSD caching to two or more flash pool aggregates. Creating an SSD storage pool requires
between 2 and 28 spare SSD disks.
In the example, SSD Disk1 through Disk6 are available as spares. The storage pool create command is used to
create the storage pool. The unit of allocation for an SSD storage pool is equal to a single slice from each SSD disk in the
storage pool. The storage pool create command slices each SSD disk into four equal pieces, making an
allocation unit that equals one fourth of all the SSD disks in the storage pool.
An allocation unit becomes a RAID group when assigned to a flash pool aggregate.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Node1
Node2
1 2 3 4 5
By default, two allocation units are assigned to each node in the HA pair. To change the ownership of one or more
allocation units of a storage pool from one HA partner to the other, use the storage pool reassign command. In
the example, one allocation unit is reassigned from Node1 to Node2.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggr2
By default, two allocation units are assigned to each node in the HA pair. To change the ownership of one or more
allocation units of a storage pool from one HA partner to the other, use the storage pool reassign command. In
the example, one allocation unit is reassigned from Node1 to Node2.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What does the feature do? What does the feature do?
Provides per-controller cache Caches random reads and overwrites
Caches random reads Provides cached data persistence through failovers
Where does the feature fit? Where does the feature fit?
With random-read workloads; for example, file With random-overwrite-heavy workloads; for
services example, OLTP
With volumes that span multiple aggregates With consistent performance (required)
The Flash Cache and Flash Pool features bring flash technology to the ONTAP software. The table compares the primary
uses and benefits of both features.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Will Virtual Storage Tier (VST) help an older system run faster?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
svl-nau-01 DP P D D D D D D D D D S DP P D D D D D D D D D S
svl-nau-02
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
SSDs are partitioned into one small root partition and one large data partition.
Standard aggregate configuration per node is as follows:
A root aggregate RAID group of 8 data + 2 parity partitions, and 2 spare root partitions
A data aggregate RAID group of 9 data + 2 parity partitions, and 1 spare data partition
The figure shows the default configuration for a single-shelf All Flash FAS system in Data ONTAP 8.3.x software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
DP R P R R R R R R S R S R DP R P R R R R S R R S
DP P D D D D D D D D D D D D D D D D D D D D D S
svl-nau-01 svl-nau-02
DP P D D D D D D D D D D D D D D D D D D D D D S
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
SSDs are partitioned into one small root and two data partitions, each of which is half the
size of a root-data partition.
Standard aggregate configuration per node is as follows:
A root aggregate RAID group of 8 data + 2 parity partitions, and 2 spare root partitions (no change
from root-data partition)
A data aggregate RAID group of 21 data + 2 parity partitions, and 1 spare data partition
Total usable capacity is 42 data partitions out of a total of 48—87.5% efficiency, or 16.7%
more usable capacity (0.875 / 0.75)
© 2016 NetApp, Inc. All rights reserved. 40
The figure shows the default configuration for a single-shelf All Flash FAS system in ONTAP 9 software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
rg0
Data partition assignments with two shelves are
similar to root-data partitioning:
DP D D D D D D D D D D D DP D D D D D D D D D D D
rg1 DP D D D D D D D D D D D DP D D D D D D D D D D D rg1
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 Data partitions on an SSD are assigned to same
node.
Twice as many RAID groups are used.
The figures show the default configuration for the two-shelf and half-shelf All Flash FAS systems in ONTAP 9 software.
For root-data partitioning and root-data-data partitioning, RAID uses the partitions in the same way as physical disks. If a
partitioned disk is moved to another node or used in another aggregate, the partitioning persists. You can use the disk only
in RAID groups that are composed of partitioned disks. If you add an unpartitioned drive to a RAID group that consists of
partitioned drives, the unpartitioned drive is partitioned to match the partition size of the drives in the RAID group, and
the rest of the disk is unused.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggregate
Physical Layer
RAID Groups of Disks
The NetApp ONTAP storage architecture uses a dynamic virtualization engine, in which data volumes are dynamically
mapped to physical space.
In ONTAP, disks are grouped into RAID groups. An aggregate is a collection of physical disk space that contains one or
more RAID groups. Each aggregate has a RAID configuration and a set of assigned disks. The disks, RAID groups, and
aggregates make up the physical storage layer.
Within each aggregate, you can create one or more FlexVol volumes. A FlexVol volume is an allocation of disk space that
is a portion of the available space in the aggregate. A FlexVol volume can contain files or LUNs. The FlexVol volumes,
files, and LUNs make up the logical storage layer.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A FlexVol volume is a volume that is loosely coupled to a containing aggregate, which the volume can share with other
FlexVol volumes. Therefore, one aggregate can be the shared source of all the storage that is used by all the FlexVol
volumes that the aggregate contains.
Because a FlexVol volume is managed separately from the aggregate, you can create small FlexVol volumes (20 MB or
larger). You can also increase or decrease the size of FlexVol volumes in increments as small as 4 KB.
FlexVol volumes have one of two formats: 64-bit or 32-bit. A 64-bit volume has a larger maximum size than a 32-bit
volume. A newly created FlexVol volume has the same format as the associated aggregate. However, a volume can have a
different format than the associated aggregate in certain cases. The maximum size of a 64-bit volume is determined by the
size of the associated aggregate, which depends on the storage system model. A 32-bit volume has a maximum size of 16
TB.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Aggregate
Data that is stored in a volume for a NAS environment is stored as files. Files can be documents, database files and logs,
audio and video, or application data. ONTAP software manages the file system operations, and clients access the data.
Data that is stored in a SAN environment is stored in a logical container that represents a SCSI disk. The container is
called a LUN. The LUN is presented to a host, which treats the LUN like a standard SCSI disk and writes data to the LUN
in 512-byte logical blocks. Therefore, SAN is often called block-level storage—because data is stored in 512-byte SCSI
blocks. ONTAP software is “unaware” of the stored files and is “aware” only of the 512-byte blocks that the host is
reading or writing to.
NOTE: Because SAN data (block data) and NAS data (file data) are treated differently, files and LUNs should not be
placed in the same FlexVol volume.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Volume:
Provisioning types:
Thick: Volume guarantee = volume
Thin: Volume guarantee = none 4KB 4KB
10%
Dynamic mapping to RG1 RG2
physical space
Aggregate
© 2016 NetApp, Inc. All rights reserved. 8
One or more FlexVol volumes can be created in an aggregate. To understand how space is managed, examine how space
is reserved in the aggregate.
The WAFL (Write Anywhere File Layout) file system writes data in 4-KB blocks that are contained in the aggregate.
When the aggregate is created, WAFL reserves 10% capacity for overhead. The remainder of the aggregate is available
for volume creation.
A FlexVol volume is a collection of disk space that is provisioned from the available space within an aggregate. FlexVol
volumes are loosely tied to their aggregates. FlexVol volumes are striped across all the disks of the aggregate, regardless
of the volume size. In the example, the blue block that is labeled “vol1” represents the inode file for the volume, and the
other blue blocks contain the user data.
When a volume is created, the volume guarantee setting must be configured. The volume guarantee setting is the same as
the space reservations. If space is reserved for the volume, the volume is thick-provisioned. If space is not reserved during
creation, the volume is thin-provisioned. FlexVol volumes are dynamically mapped to physical space. Whether the
volume is thick-provisioned or thin-provisioned, blocks are not consumed until data is written to the storage system.
A FlexVol volume can be as small as 20 MB or as large as the controller model supports. Also, the volume can grow or
shrink, regardless of the provisioning type.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Create
rtp-nau::> volume create –vserver svm_blue –name blue_vol1 –aggr rtp01_fcal_001
–size 200gb
Resize
rtp-nau::> vol modify –vserver svm_blue –name blue_vol1 –size +10gb
Destroy
rtp-nau::> vol delete –vserver svm1 –name blue_vol1 Must be offline
Volume clustershell options correspond to actions on the volume toolbar in NetApp OnCommand System Manager.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can enable or disable automatic resizing of volumes. If you enable the capability, ONTAP automatically increases the
capacity of the volume up to a predetermined maximum size. Space must be available in the containing aggregate to
support the automatic growth of the volume. Therefore, if you enable automatic resizing, you must monitor the free space
in the containing aggregate and add more when needed.
The capability cannot be triggered to support Snapshot creation. If you attempt to create a Snapshot copy and the volume
has insufficient space, the Snapshot creation fails, even when automatic resizing is enabled.
For more information about using automatic resizing, see the SAN Administration Guide.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
File or A
LUN B
C
Snapshot
Copy 1
Understanding the technology that is used to create a Snapshot copy helps you to understand how space is utilized.
Furthermore, understanding the technology also helps you to understand features such as FlexClone technology,
deduplication, and compression.
A Snapshot copy is a local, read-only, point-in-time image of data. Snapshot copy technology is a built-in feature of
WAFL storage virtualization technology and provides easy access to old versions of files and LUNs.
When ONTAP creates a Snapshot copy, ONTAP starts by creating pointers to physical locations. The system preserves
the inode map at a point in time and then continues to change the inode map on the active file system. ONTAP then
retains the old version of the inode map. No data is moved when the Snapshot copy is created.
Snapshot technology is highly scalable. A Snapshot copy can be created in a few seconds, regardless of the size of the
volume or the level of activity on the NetApp storage system. After the copy is created, changes to data objects are
reflected in updates to the current version of the objects, as if the copy did not exist. Meanwhile, the Snapshot copy of the
data remains stable. A Snapshot copy incurs no performance overhead. Users can store as many as 255 Snapshot copies
per volume. All the Snapshot copies are accessible as read-only and online versions of the data.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A
B
C
Snapshot
Copy 1
When ONTAP writes changes to disk, the changed version of block C is written to a new location. In the example, C’ is
the new location. ONTAP changes the pointers rather than moving data.
The file system avoids the parity update changes that are required if new data is written to the original location. If the
WAFL file system updated the same block, then the system would need to perform multiple parity reads to update both
parity disks. The WAFL file system writes the changed block to a new location, again writing in complete stripes and
without moving or changing the original data blocks.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A A
B B
C C’
Snapshot Snapshot
Copy 1 Copy 2
When ONTAP creates another Snapshot copy, the new Snapshot copy points only to the unchanged blocks A and B and to
block C’. Block C’ is the new location for the changed contents of block C. ONTAP does not move any data; the system
keeps building on the original active file system. The method is simple and so is good for disk use. Only new and updated
blocks use additional block space.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can use OnCommand System Manager or clustershell to create, schedule, and maintain Snapshot copies for volumes
and aggregates.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Snapshot copies are the first line of defense against accidental data loss or inconsistency. Before you implement a
Snapshot copy solution, you should thoroughly understand the customer needs and environment. Each customer has
unique requirements for the recovery time objective (RTO) and recovery point objective (RPO).
RTO
The RTO is the amount of time within which the service, data, or process must be made available again to avoid
undesirable outcomes.
RPO
The RPO is a point to which data must be restored or recovered to be acceptable to the organization’s acceptable data loss
policy.
To provide efficient use of disk space, deploy only the required number of Snapshot copies on each volume. If you deploy
more Snapshot copies than are required, the copies consume more disk space than necessary.
You might need to adjust default settings for Snapshot copy reserve for volumes and aggregates:
Snapshot copy reserve guarantees that you can create Snapshot copies until the reserved space is filled.
When Snapshot copies fill the reserved space, then Snapshot blocks compete for space with the active file system.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
(Prefix) (Timestamp)
By taking advantage of the Snapshot copy prefix, timestamp, and comment features system, administrators can easily
determine why a Snapshot copy was created.
The Prefix or Schedule
The prefix is an optional string of characters that you can specify for an automatic Snapshot copy. If a prefix is
specified, then the Snapshot name is made up of the prefix and timestamp. Prefix names must be unique within a
policy.
A schedule cannot have more than one prefix. The number of characters in the prefix counts toward the 255-character
limit on the Snapshot name.
If a prefix is specified in the Snapshot schedule, then the schedule name is not used. The schedule name is used if the
prefix is not specified for a Snapshot schedule:
volume snapshot policy add-schedule -policy <snapshot policy> -schedule <text> -
count <integer> [-prefix <text>]
The Comment
Use the volume snapshot modify command to change the text comment that is associated with a Snapshot copy.
The Label
The Vaulting subsystem uses the SnapMirror label when you back up Snapshot copies to the Vault Destination. If an
empty label ("") is specified, the existing label is deleted.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Snapshot Policy
Job Schedule
SVM
Cluster
FlexVol Volume
© 2016 NetApp, Inc. All rights reserved. 24
A Snapshot policy enables you to configure the frequency and maximum number of Snapshot copies that are created
automatically:
You can create Snapshot polices as necessary.
You can apply one or more schedules to the Snapshot policy.
The Snapshot policy can have zero schedules.
When you create an SVM, you can specify a Snapshot policy that becomes the default for all FlexVol volumes that are
created for the SVM. When you create a FlexVol volume, you can specify which Snapshot policy you want to use, or you
can enable the FlexVol to inherit the SVM Snapshot policy.
The default Snapshot policy might meet your needs. The default Snapshot copy policy is useful if users rarely lose files.
The default Snapshot policy specifies the following:
Weekly schedule to keep two weekly Snapshot copies
Daily schedule to keep two daily Snapshot copies
Hourly schedule to keep six hourly Snapshot copies
However, if users often lose files, then you should adjust the default policy to keep Snapshot copies longer:
Weekly schedule to keep two weekly Snapshot copies
Daily schedule to keep six daily Snapshot copies
Hourly schedule to keep eight hourly Snapshot copies
For typical systems, only 5% to 10% of the data changes each week: six daily and two weekly Snapshot copies consume
10% to 20% of disk space. Adjust the Snapshot copy reserve for the appropriate amount of disk space for Snapshot
copies.
Each volume on an SVM can use a different Snapshot copy policy. For active volumes, create a Snapshot schedule that
creates Snapshot copies every hour and keeps them for just a few hours, or turn off the Snapshot copy feature.
You back up Snapshot copies to the Vault Destination. If an empty label ("") is specified, the existing label is deleted.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Suppose that after the Snapshot copy is created, the file or LUN becomes corrupted, which affects logical block C’. If the
block is physically bad, RAID can manage the issue without recourse to the Snapshot copies. In the example, block C’
becomes corrupted because part of the file is accidentally deleted. You want to restore the file.
To easily restore data from a Snapshot copy, use the SnapRestore feature. SnapRestore technology does not copy files;
SnapRestore technology moves pointers from files in the good Snapshot copy to the active file system. The pointers from
the good Snapshot copy are promoted to become the active file system pointers. When a Snapshot copy is restored, all
Snapshot copies that were created after that point in time are destroyed. The system tracks links to blocks on the WAFL
system. When no more links to a block exist, the block is available for overwrite and is considered free space.
Because a SnapRestore operation affects only pointers, the operation is quick. No data is updated, nothing is moved, and
the file system frees any blocks that are used after the selected Snapshot copy. SnapRestore operations generally require
less than one second. To recover a single file, the SnapRestore feature might require a few seconds or a few minutes.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Copy data from Snapshot Locate the Snapshot copy. Requires a SnapRestore
data. license
Copy the file to the original
Use SnapRestore data location. Restores entire volumes
recovery software.
Copy the file to a new Quickly restores large files
Use the Windows Previous location.
Versions feature.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
CLI commands are available to control the visibility of Snapshot directories on a volume from NAS clients.
NOTE: Show Hidden Files and Folders must be enabled on your Windows system.
To make the .snapshot directory invisible to NFS clients, turn on the nfs.hide_snapshot option. Turn off the
nosnapdir option for each volume for which you want directories to be visible.
Access to .snapshot and ~snapshot is controlled at the volume level by setting the –snapdir-access switch. In
addition, you can control access to ~snapshot from CIFS clients at the share level with the showsnapshot share
property.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
What are the advantages and disadvantages of enabling clients to restore their
own data?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Every volume in your file system contains a special Snapshot subdirectory that enables users to access earlier versions of
the file system to recover lost or damaged files.
The Snapshot directory appears to NFS clients as .snapshot. The .snapshot directory is usually hidden and is not displayed
in directory listings, unless you use the ls command with the –a option.
When client Snapshot directories are listed, the timestamp is usually the same for all directories. To find the actual date
and time of each Snapshot copy, use the snap list command on the storage system.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
home .snapshot
.snapshot
daily.2014-09-18_0010
daily.2014-09-17_0010
daily.2014-09-18_0010
daily.2014-09-17_0010
Files on home Files on home Files on vol0 Files on vol0
(as of previous midnight) (as of night before last) (as of previous midnight) (as of night before last)
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Snapshot copies
are visible to
Windows clients
that have File
Manager
configured to
display hidden
files.
Snapshot directories are hidden on Windows clients. To view them, you must first configure File Manager to display
hidden files, then navigate to the root of the CIFS share and find the directory folder.
The subdirectory for Snapshot copies appears to CIFS clients as ~snapshot. Both automatic and manually created
Snapshot copies are listed.
To restore a file from the ~snapshot directory, rename or move the original file, then copy the file from the ~snapshot
directory to the original directory.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In Windows, right-click the file, and from the list, select Restore previous versions.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
After you complete the steps to revert a file, ONTAP software displays a warning message and prompts you to confirm
your decision to revert the file. Press Y to confirm that you want to revert the file. If you do not want to proceed, press
Ctrl+C or press N.
If you confirm that you want to revert the file that exists in the active file system, the file is overwritten by the version in
the Snapshot copy.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Whether you restore by copying files from a Snapshot directory or from tape, copying large quantities of data can be time
consuming. Instead, use the SnapRestore function to restore by reverting the volume or file.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Try, Disrupt,
Destroy
Snapshot automatic delete determines when or if Snapshot copies are automatically deleted. The option is set at the
volume level.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
FlexVol volumes can be moved from one aggregate or node to another within the same storage virtual machine (SVM). A
volume move does not disrupt client access during the move.
You can move volumes for capacity use, such as when more space is needed. You can move volumes to change
performance characteristics, such as from a controller with hard disk drives (HDDs) to one that uses SSDs. You can also
move volumes during service periods.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When a volume move is initiated, a Snapshot copy of the source volume is created and is used as the basis to populate the
destination volume. Client systems continue to access the volume from the source destination until all data is moved. At
the end of the move process, client access is temporarily blocked. Meanwhile, the system performs a final replication from
the source volume to the destination volume, swaps the identities of the source and destination volumes, and changes the
destination volume to the source volume. When the move is complete, the system routes client traffic to the new source
volume and resumes client access.
Occasionally, especially when heavy client traffic exists on the source volume, ONTAP software is unable to complete a
replication in a time frame that is transparent to clients. You can specify the –cutover-action option on a volume
move start command to indicate what should happen in such situations:
If the default action, defer_on_failure, is specified, the job tries to cut over until the cutover attempts are
exhausted. If the system fails to cut over, then the system moves into the “cutover deferred state.” The volume move
job waits for the user to issue a volume move trigger-cutover command to restart the cutover process.
If the abort_on_failure action is specified, the job tries to cut over until cutover attempts are exhausted. If the
system fails to cut over, then the system performs a cleanup and ends the operation.
If the force action is specified, the job tries to cut over until the cutover attempts are exhausted, and then forces the
cutover to occur at the expense of disrupting the clients.
If the wait action is specified, then the job does not cut over automatically after reaching the decision point. Instead,
the job waits for the user to issue a volume move trigger-cutover command as the signal to try the cutover.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software enables you to move a volume from one aggregate or node to another within the same SVM to utilize
capacity, improve performance, and satisfy SLAs. The volume move is a nondisruptive operation. During the volume
movement process, the original volume is intact and available for clients to access. You can move a FlexVol volume to a
different aggregate, node, or both within the same SVM. The data is transferred to the destination node through the cluster
interconnect.
Use the volume move start command to initiate the volume transfer. If the cutover action is defer_on_failure, and
the cutover state moves to “cutover deferred”, use the volume move trigger-cutover command to complete the
move. To bypass any confirmation before cutover, use –force true on the volume move start command. The
bypass can cause client I/O disruptions.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The volume rehost command rehosts a volume from a source SVM to a destination SVM. The volume name must be
unique among the other volumes on the destination SVM.
If the volume contains a LUN, you can specify that the LUN needs to be unmapped. In addition, you can specify whether
you want the LUN to be automatically remapped on the destination SVM.
NOTE: Volume rehost is a disruptive operation and requires you to reconfigure access to the volume at the destination.
Access to the volume must be prevented before a rehost to prevent data loss or inconsistency.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 20 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Did your volume move operation disrupt the workload on the volume that was
moved?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Administrators can manage storage systems by allocating volumes in one of two ways:
Thick provisioning of volumes uses a space guarantee for a volume or file. A guarantee of a volume requires reserved
space in the aggregate when the volume is created. A guarantee of file guarantees space for LUNs in the volume.
Thick provisioning is a conservative approach that prevents administrators from overcommitting space to an
aggregate. Thick provisioning simplifies storage management at the risk of wasting unused space.
Thin provisioning of volumes uses a space guarantee of none, meaning that no space within the aggregate is reserved
for the volume when the volume is created.
NOTE: The file guarantee is no longer supported as of NetApp Data ONTAP 8.3 software.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Waste
Save 50% power, cooling, and space.
App 3 8 Spindles
Waste
Shared
App 2 6 Spindles Capacity
App 3
12 Spindles
Waste
App 1
Standard Volume Manager NetApp Thin Provisioning
© 2016 NetApp, Inc. All rights reserved. 5
When you compare the NetApp storage use approach to competitive approaches, one feature stands out. Flexible dynamic
provisioning with FlexVol technology provides high storage use rates and enables customers to increase capacity without
the need to physically reposition or repurpose storage devices. NetApp thin provisioning enables users to overcommit data
volumes, resulting in high use models. You can think of the approach as “just-in-time” storage.
To manage thin provisioning on a cluster, use the volume command.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Thin
Provisioned
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software provides two features that can increase volume efficiency: deduplication and data compression. You can
run deduplication and data compression together or independently on a FlexVol volume to reduce the amount of physical
storage that a volume requires.
To reduce the amount of physical storage that is required, deduplication eliminates the duplicate data blocks and data
compression compresses redundant data blocks. Depending on the version of ONTAP software and the type of disks that
are used for the aggregate, the volume efficiency features can be run inline or postprocess.
Inline deduplication can reduce writes to solid-state drives (SSDs). Starting with Data ONTAP 8.3.2, inline deduplication
is enabled by default on all new volumes that are created on the All Flash FAS systems. Inline deduplication can also be
enabled on new and existing Flash Pool volumes.
Data compression combines multiple 4-KB WAFL (Write Anywhere File Layout) blocks together into compression
groups before compression. Starting with Data ONTAP 8.3.1, two data compression methods can be used: secondary and
adaptive.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Deduplication improves physical storage-space efficiency by eliminating redundant data blocks within a FlexVol volume.
Deduplication works at the block level on an active file system and uses the WAFL block-sharing mechanism. Each block
of data has a digital signature that is compared with all the other blocks in the data volume. If an exact match is identified,
the duplicate block is discarded. A data pointer is modified so that the storage system references the copy of the data
object that is stored on disk. The deduplication feature works well with datasets that have large quantities of duplicated
data or white space. You can configure deduplication operations to run automatically or according to a schedule. You can
run deduplication on new or existing data on any FlexVol volume.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Data compression enables you to reduce the physical capacity that is required to store data on a cluster, by compressing
data blocks within a FlexVol volume. Data compression is available only on FlexVol volumes that are created on 64-bit
aggregates. Data compression optimizes the storage space and bandwidth that are required to replicate data during volume
operations, such as moving volumes and performing SnapMirror transfers. You can compress standard data files, virtual
disks, and LUNs, but not file system internal files, alternate data streams, or metadata.
To manage compression on a cluster, use the volume efficiency command.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct answers
polling panel. have a green
Answer each Raise your hand to
check mark. ask a question or
question. Compare your
When finished, make a comment.
answers to the
click “Submit.” correct answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The figure shows the writes for a host or client and the amount of space on disk without any efficiency features enabled.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Default policy for All Flash FAS systems running Data ONTAP 8.3.1 software and later.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Default policy for All Flash FAS systems running ONTAP 9 software.
Data compaction is an inline operation and occurs after inline compression and inline deduplication. On an AFF system,
the order of execution is as follows:
1. Inline zero-block deduplication. All zero blocks are detected, and no user data is written to physical storage; only
metadata and reference counts are updated.
2. Inline adaptive compression. Compresses 8K logical blocks into 4K physical blocks; very efficient in determining
compressibility of the data and doesn’t waste lot of CPU cycles trying to compress incompressible data.
3. Inline deduplication. Opportunistically deduplicates incoming blocks to already existing blocks on physical storage.
4. Inline adaptive data compaction. Combines multiple <4K logical blocks into a single 4K physical block to maximize
savings. It also tries to compress any 4K logical blocks that are skipped by inline compression to gain additional
compression savings.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
FlexClone volume clones provide an efficient way to copy data for the following purposes:
Manipulation
Projection operations
Upgrade testing
ONTAP software enables you to create a volume duplicate in which the original volume and clone volume share disk
space for storing unchanged data.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
FlexClone volumes are managed similarly to regular FlexVol volumes, with a few key differences:
A FlexClone volume is a point-in-time, writable copy of the parent volume. Changes that are made to the parent
volume after the FlexClone volume is created are not reflected in the FlexClone volume.
You can clone FlexVol volumes. To create a copy of a traditional volume, you must use the vol copy command,
which creates a distinct copy with its own storage.
FlexClone volumes are fully functional volumes that are managed, as is the parent volume, by using the vol
command.
FlexClone volumes always exist in the same aggregate as parent volumes.
FlexClone volumes and parent volumes share disk space for common data. Therefore, creating a FlexClone volume is
instantaneous and requires no additional disk space (until changes are made to the clone or parent).
A FlexClone volume is created with the same space guarantee as the parent.
You can sever the connection between the parent and the clone. The severing is called splitting the FlexClone volume.
Splitting removes all restrictions on the parent volume and causes the FlexClone to use its own storage.
IMPORTANT: Splitting a FlexClone volume from the parent volume deletes all existing Snapshot copies of the
FlexClone volume and disables the creation of new Snapshot copies during the splitting operation.
Quotas that are applied to a parent volume are not automatically applied to the clone.
When a FlexClone volume is created, existing LUNs in the parent volume are also present in the FlexClone volume,
but the LUNs are unmapped and offline.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Use the volume clone split start command to initiate a split of the clone from the parent.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
NFS iSCSI
Corporate CIFS
LAN FCoE
FC
File System
(NAS)
NAS SAN
(File-Level (Block-Level
Access) Access)
NetApp FAS
A SAN is a block-based storage system that uses FC, FCoE, and iSCSI protocols to make data available over the network.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Server
NFS SMB
Volume Volume
NAS is a distributed file system that enables users to access resources, such as volumes, on a remote storage system as if
the resources were located on a local computer system.
NAS provides services through a client-server relationship. Storage systems that enable file systems and other resources to
be available for remote access are called servers. The server is set up with a network address and provides file-based data
storage to other computers, called clients, that use the server resources.
The NetApp ONTAP software supports the NFS and SMB protocols. (SMB is also known as CIFS.)
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
With the NAS protocols, you need to create file systems and other resources that are available to clients through either
NFS or SMB.
Volumes are the highest level of logical storage object. FlexVol volumes are data containers that enable you to partition
and manage your data. In a NAS environment, volumes contain file systems. The first resource to create is the volume.
In ONTAP software, the volume is associated with a storage virtual machine (SVM). The SVM is a virtual management
entity, within which you create a namespace. Volumes are joined to the namespace through junctions. The junctions are
exported.
Qtrees enable you to partition FlexVol volumes into smaller segments that you can manage individually. ONTAP
software creates a default qtree, called qtree0, for each volume. If you do not create and put data in another qtree, all the
data resides in qtree0. Qtrees enable you to partition data without incurring the overhead that is associated with creating
another FlexVol volume. You might create qtrees to organize data or to manage one or more of the following factors:
quotas, security style, or opportunistic lock (oplock) settings.
You can also create a directory or a file on the client in a FlexVol volume, to use as a resource to export or share. A qtree
is a partition that is created on the storage system. A directory is a partition that is created on the client within a FlexVol
volume.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
OR
Create a second named project volume:
rtp-nau::> volume create –vserver svm_blue projects
-aggregate sas_data_18 –volume thesis
-size 10GB –state online –type RW
–policy Default –security-style unix
The following is an abbreviated list of parameters that are used to mount a volume:
Junction path of the mounting volume: -junction-path <junction path>
The junction path name is case insensitive and must be unique within an SVM namespace.
Active junction path: [-active {true|false}]
The optional parameter specifies whether the mounted volume is accessible. The default setting is false. If the
mounted path is inaccessible, then the path does not appear in the SVM namespace.
Override the export policy: [-policy-override {true|false}]
The optional parameter specifies whether the parent volume’s export policy overrides the mounted volume’s export
policy. The default setting is false.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Volume junctions are a way to join individual volumes into a single logical namespace. Volume junctions are transparent
to CIFS and NFS clients. When NAS clients access data by traversing a junction, the junction appears to be an ordinary
directory.
A junction is formed when a volume is mounted to a mount point below the root and is used to create a file-system tree.
The top of a file-system tree is always the root volume, which is represented by a slash mark (/). A junction points from a
directory in one volume to the root directory of another volume.
A volume must be mounted at a junction point in the namespace to enable NAS client access to contained data. Although
specifying a junction point is optional when a volume is created, data in the volume cannot be exported and a share cannot
be created until the volume is mounted to a junction point in the namespace. A volume that was not mounted during
volume creation can be mounted post-creation. New volumes can be added to the namespace at any time by mounting
them to a junction point.
NOTE: Use the storage system to mount volumes to junction paths.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When volumes are created by using the volume create command, a junction path is usually specified. The junction
path is optional; a volume can be created and not mounted into the namespace. To put a volume without a junction path
into use, you must use the volume mount command to assign a junction path to the volume.
A volume can be mounted to the namespace of the SVM in only one place.
When you unmount a volume, you take the volume out of the namespace. An unmounted volume is inaccessible to NFS
and CIFS clients but is still online and can be mirrored, backed up, moved, and so on.
You can then mount the volume again to the same location or to a different location in the namespace and in relation to
other volumes. For example, you can unmount a volume from one parent volume, and then mount the volume to another
parent volume.
Be careful when unmounting and remounting a volume to a new path. Because rejunctioning changes the location of a
flexible volume inside the namespace, the namespace is not transparent to client access. The client now has to access the
data at the new directory location.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
/acct /project2
/project1
pro2
acct pro1
SVM Volume Junction Path
svm_blue blue_root /
svm_blue acct /acct
svm_blue pro1 /project1
svm_blue pro2 /project2
svm_blue pro3 /project3
In an architecture with standalone volumes, every volume has an insertion point to the root of the SVM namespace. No
volume is junctioned below another volume. Each volume has a unique path and is junctioned directly below the root.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
root /project
/acct pro3 /project/pro3
acct
project
/project/pro2
pro2
SVM Volume Junction Path
svm_blue blue_root /
/project/pro1 pro1 svm_blue acct /acct
svm_blue project /project
svm_blue pro1 /project/pro1
svm_blue pro2 /project/pro2
svm_blue pro3 /project/pro3
An architecture with multiple branched trees has multiple insertion points to the root of the SVM namespace. The
insertion points can be junctioned volumes, directories, or qtrees beneath the root. All other volumes are mounted at
junction points beneath the insertion points (which can be volumes, directories, or qtrees).
The figure shows a typical volume junction configuration, with two insertion points to the root volume of the SVM. One
insertion point is a junctioned volume “acct”, and one insertion point is a junctioned volume “project.” The other volumes
are junctioned under the “project” volume.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
root
/acct
project
acct /project/pro3
pro3
The example has two insertion points. One insertion point is from the root to the “acct” volume. The second insertion
point is a directory that was created from either an export of the root volume to a UNIX host or from within a share of the
root volume to a Windows host.
The second insertion point can also be a qtree in place of the directory.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
vol01
NFS is a distributed file system that enables users to access resources, such as volumes, on remote storage systems as if
the resources were located on a local computer system.
NFS provides services through a client-server relationship.
Storage systems that enable the file systems and other resources to be available for remote access are called servers.
The computers that use a server's resources are called clients.
The procedure of making file systems available is called exporting.
The act of a client accessing an exported file system is called mounting.
When a client mounts a file system that a server exports, users on the client machine can view and interact with the
mounted file systems on the server within the permissions granted.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The figure shows the basic process for implementing the NFS protocol between a UNIX host and an ONTAP storage
system. The process consists of several steps.
First, you need to enable the NFS functionality, license NFS, and then enable the feature on the storage system.
Second, you need resources to export, so you create volumes and qtrees.
Third, you determine which clients have which type of access to the resources. You need a way to authenticate client
access and authorize users with appropriate permissions, including read-only or read/write.
Finally, when the client has been granted access to the exported resource, the client mounts the resource and grants access
to the users.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Best practice:
Configure NAS protocols through
NetApp OnCommand System
Manager.
After you license NFS, enable the protocol. You can enable NFS through the CLI or NetApp OnCommand System
Manager. NetApp recommends using the tools and wizards that are available through System Manager.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Create
IPspace
Choose an IP
address from
the subnet? Create an SVM
admin.
Protocols
Network port
Root Create a
aggregate volume to
NIS information
export
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software uses export policies and rules to control host access.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
-clientmatch <text>
The clientmatch parameter specifies the client or clients to which the export rule applies. You can specify the match
in any of the following formats:
As a host name; for example, host1
As an IPv4 address; for example, 10.1.12.24
As an IPv4 address with a subnet mask that is expressed as a number of bits; for example, 10.1.12.10/4
As an IPv4 address with a network mask; for example, 10.1.16.0/255.255.255.0
As a netgroup, with the netgroup name preceded by the @ character; for example, @netgroup
As a network name from files, Network Information System (NIS), or Lightweight Directory Access Protocol
(LDAP), preceded by the = character; for example, =networkname
As a domain name preceded by the .character; for example, .example.com
NOTE: Entering an IP address range, such as 10.1.12.10-10.1.12.70, is not permitted. Entries are interpreted as a text
string and treated as a hostname.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In the example scenario, you create an export policy and an export policy rule, and then you apply the rule to the export.
You create an export rule with index number 1 in an export policy named vs1_pro1 on an SVM named vs1. The rule
matches all clients in the specified subnet. The rule enables read-only access by any matching client and requires
authentication by Kerberos 5 for read-write access.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
To enable an NFS client, mount a remote file system after NFS starts. Usually, only a privileged user can mount file
systems with NFS. However, you can enable users to mount and unmount selected file systems by using the mount and
umount commands, if the user option is set in /etc/fstab. The setting can reduce traffic by having file systems mounted
only when they are needed. To enable user mounting, create an entry in /etc/fstab for each file system to be mounted.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Server
vol01
SMB is an application-layer network file-sharing protocol that the Microsoft Windows operating system uses. SMB
enables users or applications to access, read, and write to files on remote computers just like on a local computer. For the
purposes of this course, the terms SMB and CIFS are used interchangeably (although the definitions of the two terms are
not strictly the same).
A user or application can send network requests to read and write to files on remote computers. Messages travel from the
network interface card (NIC) of the user’s computer, through the Ethernet switch, to the NIC of the remote computer.
SMB provides access to files and directories that are stored on the remote computer, through sharing resources. The
network read and write process, which is also called network I/O, is controlled by the rules of network protocols such as
IPv4 and IPv6.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
To implement SMB, first enable the SMB functionality on the ONTAP storage system. Then share the available
resources. Finally, map the shared resources on the Windows client.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Best practice:
rtp-nau::> vserver cifs create…
Configure NAS protocols through
OnCommand System Manager.
© 2016 NetApp, Inc. All rights reserved. 31
After you license CIFS, enable the protocol. You can enable SMB through the CLI or OnCommand System Manager.
NetApp recommends using the tools and wizards that are available through OnCommand System Manager.
CIFS setup enables you to perform several tasks: create and name a CIFS server that your CIFS clients can access,
join the CIFS server to a domain, and create a default set of local CIFS users and groups.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Create
IPspace Choose an IP
address from
the subnet?
Create an SVM
admin.
Protocols Network port
Create a
Root volume and a
aggregate share
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You assign exports to volumes and qtrees. Which resources can you share
through SMB?
SMB shares are associated with paths within the namespace. Because the namespace is constructed by junctions, qtrees,
and directories, shares can be associated with any of the resources.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
A CIFS share is a named access point in a volume that enables CIFS clients to view, browse, and manipulate files on a file
server. When creating CIFS shares, consider certain guidelines.
When you create a share, you must provide all the following information:
The complete path in a volume to the CIFS share
The name of the share that users enter when they connect to the share
When you create a share, you can optionally specify a description for the share. The share description appears in the
Comment field when you browse the shares on the network.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In the example, on the pro share, you set the share access for the “friends” group to Full Control and delete the “everyone”
access control list (ACL) entry.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Share permissions apply only to users who access the resource over the network. The permissions apply to all files and
folders in the shared resource.
Full Control: Full control is the default permission that is assigned to the Administrators group on the local
computer. Full control permits all Read and Change permissions, plus Changing permissions (NTFS files and folders
only).
Read: Read is the default permission that is assigned to the Everyone group. Read permits the following actions:
– View file names and subfolder names.
– View data in files.
– Run program files.
Change: Change is not a default permission for any group. The Change permission enables all Read permissions, plus
the following actions:
– Add files and subfolders.
– Change data in files.
– Delete subfolders and files.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The net view command displays a list of computers with shared resources that are available on the specified
computer.
To use the net view command, use the following steps:
1. Click the Start button, point to Programs, and then click the MS-DOS prompt.
2. At the command prompt, type net view \\<computername>, where <computername> is the name of a specific
computer whose resources you want to view.
You can connect or disconnect a computer from a shared resource or display information about computer connections.
The command also controls persistent net connections. Used without parameters, the net use command retrieves a list
of network connections.
You can also use Windows to map a share to a client.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Permissions are rules that are associated with objects on a computer or network, such as files and folders. Permissions
determine whether a user can access an object and what the user can do with the object. For example, you might have
access to a document on a shared folder on a network. Even though you can read the document, you might not have
permissions to change the document. Windows file permissions include the following:
Full control: Users can see the contents of a file or folder, change existing files and folders, create new files and
folders, and run programs in a folder.
Modify: Users can change existing files and folders but cannot create new ones.
Read and execute: Users can see the contents of existing files and folders and can run programs in a folder.
Read: Users can see the contents of a folder and open files and folders.
Write: Users can create files and folders and make changes to existing files and folders.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SMBvol SMBvol
You can configure an SVM with a CIFS server. You can create the CIFS server either as a member of a Microsoft Active
Directory domain or in a workgroup.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
CIFS servers in workgroup mode do not support the following CIFS features:
Configuration in OnCommand System Manager
SMB3 witness protocol
SMB3 continuously available shares
SQL over SMB
Folder redirection
Roaming profiles
Group Policy Object (GPO)
Volume Snapshot Service (VSS)
Before creating a CIFS server, be aware that all the CIFS features that require a Windows domain are unsupported by a
CIFS server in workgroup mode.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 45 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Share your experience using the System Manager SVM Creation Wizard to
configure CIFS and NFS. Do you expect your NFS mounts and SMB drive
mappings to be successful?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
DNS load balancing and automatic LIF rebalancing methods can help you select an appropriately loaded data LIF and
balance user network traffic across all available ports (physical or interface groups).
DNS Load Balancing
With DNS load balancing, you can create a DNS load-balancing zone on the SVM that returns the least-loaded LIF, based
on the network traffic and the availability of the port resources (such as CPU usage, throughput, and open connections).
By configuring a DNS load-balancing zone, you can better balance new client connections across available resources.
Balance leads to improved performance for the entire cluster. Also, no manual intervention is required for deciding which
LIFs to use when mounting a particular SVM. You can use the DNS load-balancing method to balance loads for only new
share connections and new mount requests. DNS load balancing cannot be used with existing connections. DNS load
balancing works with NFSv3, NFSv4, NFSv4.1, CIFS, SMB 2.0, SMB 2.1, and SMB 3.0.
Automatic LIF Rebalancing
With automatic load balancing, LIFs are dynamically migrated to ports with low utilization, based on the failover rules.
Automatic LIF rebalancing works only with NFSv3 connections. Automatic LIF rebalancing provides the following
benefits:
Different client connections use different bandwidth. Therefore, LIFs can be migrated based on the load capacity.
When new nodes are added to the cluster, LIFs can be migrated to the new ports.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
name.
3. Configure DNS Server for e0e
LIF2
round-robin load balancing.
e0e
LIF3
1 3 e0e
DNS LIF4
blue.netapp.com IN A <LIF1 IP Address>
blue.netapp.com IN A <LIF2 IP Address>
blue.netapp.com IN A <LIF3 IP Address> Create “A” record for
blue.netapp.com IN A <LIF4 IP Address>
each LIF on DNS server.
With off-box DNS, each data LIF in each SVM that resides in the cluster has a DNS “A” record that is created with the
same name.
1. NFS client makes a request for name resolution to the site-wide DNS server.
2. Site-wide DNS server resolves the request to an IP address, by using a round-robin algorithm.
3. Site-wide DNS server responds to the client with the chosen IP address.
As with the on-box method, if an environment has many SVMs, you must account for each data LIFs of each SVM that
you add to the site-wide DNS server.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 15 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Were you able to use both the SMB and NFS protocols to access the same
volume in the namespace?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
NFS iSCSI
Corporate CIFS
LAN FCoE
FC
File System
(NAS)
NAS SAN
(File-Level (Block-Level
Access) Access)
NetApp FAS
A SAN is a block-based storage system that uses FC, FCoE, and iSCSI protocols to make data available over the network.
SAN is currently supported in clusters of up to eight nodes.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
LUN
LUN
In an application server environment, locally attached hard disks, also called direct-attached storage (DAS), are separately
managed resources. In an environment with more than one application server, each server’s storage resource also needs to
be managed separately.
A SAN provides access to a LUN, which represents a SCSI-attached hard disk. The host operating system partitions,
formats, writes to, and reads from the LUN as if the LUN were any other locally attached disk. The advantages of using
SAN storage include support for clustered hosts, where shared disks are required, and centrally managed resources. In the
example, if a SAN was not used, the administrator would need to manage separate resources for each application server
and host cluster. In addition to centrally managed resources, SAN also enables centrally managed data protection, using
NetApp ONTAP Snapshot copy technology.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
SCSI provides low-level block access to data, typically in 512-byte blocks. Low-level block access requires less overhead
than file-level access. SCSI has a high level of resiliency that is suitable for an enterprise-level protocol.
A client-server service-delivery model describes the relationships between SCSI devices. A relationship between two
SCSI devices is called a nexus. The client, or SCSI initiator, sends a command and the server, or SCSI target, returns a
response. The initiator uses SCSI commands to request a read from a LUN or write to a LUN.
iSCSI does not use the complete SCSI standard.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Up to 8 Nodes in
ONTAP 9.0
ONTAP software continues to support Windows, Red Hat Linux, VMware ESX, HP-UX, and Solaris hosts. ONTAP
software now also supports AIX as a SAN host. To function with scalable SAN, all SAN client stacks must support
asymmetric logical unit access (ALUA).
As of NetApp Data ONTAP 8.3, the maximum number of LUNs per cluster increased to 98,304. The increase enables
12,288 LUNs per node, times the maximum SAN cluster size of eight nodes.
Consult the NetApp Supportability Matrix for details about supported versions of SAN hosts.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
LUN
≈ SCSI Disk
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Disk 1 (C:)
Disk 2 (E:) LUN
Connected
Through a
Switch
(Controller or
SVM)
LUN FlexVol Volume
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Cluster Interconnect
HA
Do not fail over LIF1 LIF1
HA
e1b
Recommendation: Use at least one LIF per
node, per SVM, per network.
Data SVM
© 2016 NetApp, Inc. All rights reserved. 11
SAN data logical interfaces (LIFs) do not migrate or fail over the way that NAS does. However, the logical interfaces
(LIFs) can be moved to another node or port in the SVM.
To move a data LIF with SAN protocols, use the network interface modify command:
1. To view the current status of a LIF, use the network interface show command.
2. Change the admin status of the LIF to down (offline).
3. Change the location of the LIF to a new node or port by using the network interface modify command.
4. Change the admin status of the LIF to up (online).
5. Verify the changes by using the network interface show command.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Data is communicated over ports. In an Ethernet SAN, the data is communicated by means of Ethernet ports. In an FC
SAN, the data is communicated over FC ports. For FCoE, the initiator has a converged network adapter (CNA) and the
target has a unified target adapter (UTA).
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
e0a 0a
HA
LIF1 LIF2
Data SVM
© 2016 NetApp, Inc. All rights reserved. 13
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
vs_iscsi
iqn.1992-08.com.netapp:sn.000…:vs
LUN FlexVol Volume
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Ethernet
LUNa
Microsoft Multipath I/O (MPIO) software is required any time that a Windows host has more than one path to the storage
system. The MPIO software presents a single disk to the operating system for all paths, and an ONTAP device-specific
module (DSM) manages path failover. Without MPIO software, the operating system might see each path as a separate
disk, which can lead to data corruption.
On a Windows system, there are two main components to any MPIO configuration: the Windows MPIO components and
a DSM. MPIO is supported for Windows Server 2003, Windows Server 2008, and Windows Server 2012 systems.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Multipath Driver
E0 E1
LUNa
As paths are added between the storage controllers and the host, the LUN is seen once through each path. When a
multipath driver is added to the host, the multipath driver can present the LUN as a single instance.
The figure illustrates four paths. Two paths are active and optimized. Two paths are active and nonoptimized.
NOTE: The paths in the figure are simplified for conceptual purposes. Depending on the platform and version, paths
might appear differently, physically or logically, but the concept of ALUA states (active/optimized, active/non-optimized,
or unavailable) is the same. ONTAP active/nonoptimized paths are over the cluster interconnect.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Host Utilities is a set of software programs and documentation that enables you to connect host computers to LUNs on
NetApp storage systems. Download Host Utilities from the NetApp Support site for the operating system that runs on
your host.
Host Utilities features for each operating system might differ slightly. Windows Unified Host Utilities, for example,
includes an installation program that sets required parameters on the host computer and on certain host bus adapters
(HBAs). Parameters include setting time-out values to enable proper failover.
The package also includes documentation to describe how to install Host Utilities and troubleshoot typical problems. The
package might also include diagnostic programs to troubleshoot problems with hosts that connect to the storage system.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
This poll has one quick question about Performance Manager. When the instructor begins the polling session, you see the
polling question that you are to answer. After all answers are submitted, the instructor closes the poll.
After the polling session ends, the instructor briefly answers the question for you.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
With which set of protocols does ONTAP software support asymmetric logical unit
access (ALUA)?
a. FC
b. FC and FCoE
c. FC, FCoE, and iSCSI
d. FC, FCoE, iSCSI, and NFS
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The figure shows the basic steps to implement iSCSI for an SVM on ONTAP software. You can enable iSCSI by using
either the CLI or the NetApp OnCommand System Manager UI.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When iSCSI is enabled, the operational status of the iSCSI service on the specified SVM is up and ready to serve data.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
For ONTAP software, to create an iSCSI LIF, you must specify the -role parameter as data and the –protocol parameter as
iscsi. A built-in firewall exists for extra security. A LIF should be associated with the firewall role of the data. To verify
the firewall role for a particular LIF, use the instance switch on the network interface show command.
NOTE: You should create at least one LIF for each node and each network on all SVMs that are serving data with the
iSCSI protocol. NetApp recommends having network redundancy, either through multiple networks or link aggregation.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The iSCSI software initiator creates the iSCSI connection on the Windows host. For Windows Server 2003, you must
download and install the software initiator. The iSCSI software initiator is built in to Windows Server 2008 and Windows
Server 2012.
If the system has not used an iSCSI initiator before, a dialog box appears, which requests that you turn on the service.
Click Yes. The iSCSI Initiator Properties dialog box appears.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Initiator Ethernet
Administrator must tell the host
where to discover the target by
using either the target portal IP
IP SAN
address or an Internet Storage
Name Service (iSNS) server.
LIF LIF
Target Ethernet HA
(Controller or
SVM)
The administrator must direct the software initiator on the host to discover the target. Discovery can be performed through
one of two methods: send targets (entering the target portal IP address) or using an iSNS server. The slides illustrate the
send targets method, which is most widely used and require no other servers to implement.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When a target is discovered, the target appears in the iSCSI Initiator properties as Inactive.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
2. Click Connect.
Select the inactive target and click the Connect button. The Connect To Target dialog box opens. In that dialog box, you
can enable persistent bindings (Favorite Targets), enable multipath, and modify advanced options.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
1. Click Properties.
Use iscsi session show
to verify on storage system.
When a successful connection is made with the storage system, a session is created. If multiple paths are available, a
session needs to be created for each path (or LIF).
You can also display information about sessions or connections on the storage.
The iscsi session show command displays session information, and the vserver iscsi connection
show command displays connection information.
The session information is also available through OnCommand System Manager.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When a session is created between an initiator and a target, a session ID is generated. Sessions IDs are typically a long
string of digits, which can be difficult to use for identification of a particular session. An easier way to identify a session
by LIF is to use the Target Portal Group tag.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
This poll has one quick question about Performance Manager. When the instructor begins the polling session, you see the
polling question that you are to answer. After all answers are submitted, the instructor closes the poll.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 45 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In Windows, a LUN
appears as a disk.
There are many ways to discover and prepare the LUN in Windows. Each version of Windows might have slightly
different tools that you can use. This module illustrates the most often used method. In Windows, a LUN appears as a disk
and and is labeled as a disk.
Open Computer Management. Select Disk Management. If the LUN that you created is not displayed, rescan disks by
right-clicking Disk Management or, from the Action menu, select Rescan Disks.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
1. Right-click the disk, and then 2. Right-click the new disk, and then
select Initialize Disk. select Online.
Depending on how many LUNs you mapped, one or more disks might appear. Identify the disk that you want to prepare.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Now that the disk—the LUN was presented to Windows—is online and initialized, you need to provision a volume on
which to put data. There are many ways to provision a volume in Windows. This module illustrates the most often used
method: the New Simple Volume Wizard from the Disk Management utility.
In the Disk Management utility, launch the New Simple Volume Wizard by right-clicking the disk that you want to
provision and selecting New Simple Volume.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You need to specify the size of the volume—typically equal to the LUN size. (The volume spans the LUN.)
You also need to select a way to access the volume, by assigning a drive letter or a mount point. You can also choose not
to do anything if you do not want the volume to be accessible yet.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Format the volume with a file system, typically NTFS. Now is a good time to label the volume for easier identification.
Verify the settings, and then click Finish to complete the process.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 20 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
How does partitioning and formatting a LUN from the Windows host differ from
partitioning and formatting a physical hard drive in Windows?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Upgrade Advisor is an online tool, available on the NetApp Support Site, that simplifies the process of planning ONTAP
upgrades. When you submit your system identification and target release to Upgrade Advisor, the tool compares
AutoSupport data about your cluster to known requirements and limitations of the target release. Upgrade Advisor then
generates an upgrade plan (and optionally a back-out plan) with recommended preparation and execution procedures.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Rolling upgrades can be performed on clusters of two or more nodes but run on one node of an HA pair at a time.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can perform batch upgrades on clusters of eight or more nodes. Unlike rolling upgrades, batch upgrades can be run
on more than one HA pair at a time.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The automated upgrades that you can perform by using System Manager consist of three stages: Select, Validate, and
Update.
In the first stage, you select the ONTAP software image. The current version details are displayed for each node or HA
pair.
In the second stage, you view and validate the cluster against the software image version for the update. A pre-update
validation helps you determine whether the cluster is ready for an update. If the validation is completed with errors, a
table displays the status of the various components and the required corrective actions. You can perform the update only
when the validation is completed successfully.
In the third and final stage, you update all the nodes in the cluster or an HA pair in the cluster to the selected version of the
software image. While the update is in progress, you can choose to pause and then either cancel or resume the update. If
an error occurs, the update is paused and an error message is displayed with the remedial steps. You can choose to resume
the update after performing the remedial steps or cancel the update. You can view the table with the node name, uptime,
state, and ONTAP software version when the update is successfully completed.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
1. Verify that the nodes are configured as You can enter the following commands at any time:
"help" or "?" - if you want to have a question
HA pairs and connected to the cluster clarified,
interconnect. "back" - if you want to change previously answered
questions, and
"exit" or "quit" - if you want to quit the cluster
2. Power on both nodes of the HA pair. setup wizard.
Any changes you made before quitting will be saved.
3. Start the Cluster Setup wizard on one You can return to cluster setup at any time by typing
of the nodes. "cluster setup".
To accept a default or omit a question, do not enter a
value.
4. Use the join command and follow the
wizard.
Do you want to create a new cluster or join an existing
cluster?
{create, join}: join
5. Repeat Steps 3 and 4 on the partner
node.
© 2016 NetApp, Inc. All rights reserved. 10
You can expand an existing cluster by nondisruptively adding nodes to the cluster.
You must add nodes from HA pairs that are connected to the cluster interconnect. Nodes are joined to the cluster one at a
time.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
1. Which two upgrade types can group HA pairs that are upgraded together?
(Choose two.)
a. rolling
b. batch
c. automated
d. hardware
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
3. Which three protocols can you use to download the ONTAP software image?
(Choose three.)
a. NFS
b. FTP
c. TFTP
d. HTTPS
e. HTTP
f. CIFS
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
!!
© 2016 NetApp, Inc. All rights reserved. 16
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The AutoSupport tool is an integrated and efficient monitoring and reporting technology that checks the health of
AutoSupport-enabled NetApp systems on a continual basis. The AutoSupport tool should be enabled on each node of the
cluster.
To manage AutoSupport in System Manager, on the node’s Configuration tab, click the AutoSupport link. The
AutoSupport tool can be enabled or disabled. To configure AutoSupport, click Edit, and then enter your configuration
information.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
My AutoSupport is a suite of web-based applications hosted on the NetApp Support site and accessible via your web
browser. Using the data from the AutoSupport support tool, My AutoSupport proactively identifies storage infrastructure
issues through a continuous health-check feature and automatically provides guidance on remedial actions that help to
increase uptime and avoid disruptions to your business.
For example, My AutoSupport might find a configuration issue, a bad disk drive, or version incompatibility on your
system. Or My AutoSupport can notify you of end-of-life (EOL) issues or an upcoming support contract expiration date.
If you plan any changes to your controllers, NetApp recommends manually triggering an AutoSupport message before
you make the changes. The message provides a “before” snapshot for comparison, in case a problem arises later.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The event management system (EMS) collects and displays information about events that occur on your cluster. You can
manage the event destination, event route, mail history records, and SNMP trap history records. You can also configure
event notification and logging.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Performance, Capacity,
Complexity of Configuration
Configuration,
Strong ROI Story
Insight
Manage at Scale,
Automate Storage Processes
and Data Protection
Simple, Web-Based,
No Storage Expertise Required
Basic
System Manager
NetApp Storage Multivendor
© 2016 NetApp, Inc. All rights reserved. 21
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The OnCommand System Manager dashboard shows at-a-glance system status for a storage system. It displays vital
storage information, including efficiency and capacity utilization for various storage objects, such as aggregates and
volumes.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
To learn more about OnCommand Unified Manager and how it integrates with OnCommand
Performance Manager and OnCommand Workflow Automation, enroll in the instructor-led course:
Administration of OnCommand Management Solutions
By using OnCommand Unified Manager, you can configure global threshold values for all your aggregates and volumes,
to track any threshold breaches.
Events are notifications that are generated automatically when a predefined condition occurs or when an object crosses a
threshold. The events enable you to take action to prevent issues that can lead to poor performance and system
unavailability. Events include an impact area, severity, and impact level. Events are categorized by the type of impact
area, such as availability, capacity, configuration, or protection.
You can create alerts to notify you when a particular event is generated. You can create alerts for a single resource, group
of resources, and events of a particular severity type. You can specify the frequency with which you want to be notified.
You can integrate OnCommand Workflow Automation with Unified Manager to execute workflows for your storage
classes. You can also monitor SVMs that have an infinite volume but do not have storage classes. When Unified Manager
is integrated with Workflow Automation, the reacquisition of Workflow Automation cached data is triggered.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Storage system performance calculations vary widely based on the kind of operations, or workloads, that are being
managed.
The storage system sends and receives information in the form of I/O operations. l/O operations can be categorized as
either random or sequential. Random operations, such as database operations, are usually small, lack any pattern, and
happen quickly. In contrast, sequential operations, such as video files, are large and have multiple parts that must be
accessed in a particular order.
Some applications have more than one dataset. For example, a database application’s data files and log files might have
different requirements. Data requirements might also change over time. For example, data might start with specific
requirements that change as the data ages.
If more than one application shares the storage resources, each workload might need to have quality of service (QoS)
restrictions imposed. QoS restrictions prevent applications or tenants from being either bullies or victims.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Input/output operations per second (IOPS) is a measurement of how many requests can be managed in one second.
Factors that affect IOPS include the balance of read and write operations in the system and whether traffic is sequential,
random, or mixed. Other factors that affect IOPS include the application type, the operating system, background
operations, and I/O size.
Applications with a random I/O profile, such as databases and email servers, usually have requirements that are based on
an IOPS value.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Throughput is a measurement of the average number of megabytes, that is how much data, can be transferred within a
period for a specific file size. Throughput is measured in megabytes per second.
Applications with a sequential I/O profile, such as video or audio streaming, file servers, and disk backup targets, usually
have requirements that are based on megabytes per second.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Latency is the measurement of how long a storage system takes to process an I/O task. Smaller latency values are better.
Latency for hard disks is typically measured in milliseconds. Because solid-state media is much faster than hard disks, the
latency of the media is measured in submilliseconds or microseconds.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Latency
Headroom assessment varies with workload,
and the optimal operating point is dynamic. Max Optimal
Operating Point
The optimal operating point is determined by
Operating Point
analyzing system behavior under load and
applying a heuristic “knee of the latency curve” Headroom
algorithm. Utilization
Headroom is the difference between the
maximum optimal and current operating points.
Knowing the available performance capacity in the cluster helps you to provision and balance workflows. Performance
capacity is how much work you can place on a node or an aggregate before latency begins to affect the performance of all
workloads. You can use Performance Manager or the CLI to measure performance capacity.
To use CLI to obtain the remaining performance capacity, use the following steps:
1. Change to advanced privilege level:
set -privilege advanced
2. Start the statistics command line prompt:
statistics start –object resource_headroom_cpu
3. Display real-time headroom information:
statistics show -object resource_headroom
4. Return to administrative privilege:
set -privilege admin
Compute the available performance capacity by subtracting the optimal_point_counter from the
current_counter. For example, if the utilization capacity for a particular CPU is -14% (72%-86%) over an hourly
basis, the node's CPU might be overutilized on average for the past one hour. In addition, you can specify daily, weekly,
or monthly to get the same information, but averaged over a longer period.
NOTE: The resource_headroom_cpu CM object in the example represents the entire node (all CPUs collectively).
You can get the available performance capacity on aggregates by using the same stat command syntax but with the
resource_headroom_aggr CM object.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Counter Value
-------------------------------- --------------------------------
ewma_hourly -
ops 4376
latency 37719
utilization 60
optimal_point_ops 2573 72%-60%=12%
optimal_point_latency 3589
optimal_point_utilization 72
optimal_point_confidence_factor 1
Knowing the available performance capacity in the cluster helps you provision workflows and balance them. Performance
capacity is how much work you can place on a node or an aggregate before performance of all workloads begins to be
affected by latency.
You compute the available performance capacity by subtracting the optimal_point_utilization counter from the utilization
counter. In this example, the utilization capacity for this CPU is 12% (72%-60%). This suggests that the node's CPU has
been underutilized on average for the past one hour.
Additional headroom capability is available in OnCommand Performance Manager 7.0.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
With ONTAP 9, workloads can apply both IOPS and megabytes per second controls to storage QoS limits.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
ONTAP software performance is measured at the aggregate level. To support the differing security, backup, performance,
and data sharing needs of your users, you can group the physical data storage resources on your storage system into one or
more aggregates. You can then design and configure the aggregates to provide the appropriate level of performance and
redundancy.
When creating aggregates and the underlying RAID group, you must balance the need for performance and the need for
resilience. By adding more disks per RAID group, you increase performance by spreading the workload across more
disks, but at the cost of resiliency. In contrast, adding fewer disks per RAID group increases the resiliency because the
parity has less data to protect, but at the cost of performance.
By following best practices when you add storage to an aggregate, you optimize aggregate performance. You should also
choose the right disk type for the workload requirements.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Flash
Acceleration Use SAS for
performance.
The proper disk type depends on the performance or capacity requirements of the workload.
When a workload requires the largest capacity at the lowest cost with lower performance, use SATA disks.
When a workload requires the highest performance at the lowest cost with lower capacity, use solid-state drives (SSDs).
When a workload requires a balance of capacity and performance, use SAS disks.
Sometimes, a workload might require large amounts of capacity at the lowest cost but at a higher performance than SATA
or SAS provides. To improve the performance of high-capacity hard disks, you can use Flash Cache or a Flash Pool
technologies.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
As well as discussing performance at the node level, discussing performance at the cluster level is important.
In the example, an administrator creates volumes on a two-node cluster that is used for file services. The system is
configured with SATA disks to meet the workload requirements.
After some time, the administrator needs to add a volume for a database application. The SATA disks do not meet the
requirements for the new workload. The administrator decides, for future growth, to nondisruptively add another HA pair
with SAS disks. With new nodes with SAS disks active in the cluster, the administrator can nondisruptively move the
volume to the faster disks.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The administrator has a new requirement for a workload that requires high-performance requirements. For easier
management of the various workload types, the administrator decides to create in the cluster a new high-performance tier
that uses All Flash FAS controllers.
NetApp ONTAP FlashEssentials is the power behind the performance and efficiency of All Flash FAS. All Flash FAS
uses high-end or enterprise-level controllers with an All Flash personality, which supports SSDs only. For more
information about All Flash FAS and FlashEssentials, see Using All Flash FAS with ONTAP on the NetApp Support site.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
You can use storage QoS to deliver consistent performance by monitoring and managing application workloads.
You can configure the storage QoS feature to prevent user workloads or tenants from affecting one another. The feature
can be configured to isolate and throttle resource-intensive workloads. The feature can also enable critical applications to
achieve consistent performance expectations.
Policies are throughput limits that can be defined in terms of IOPS or megabytes per second.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Start with a properly sized system and follow best practices for ONTAP software, the host operating system, and the
application. Verify and adhere to the supported minimums, maximums, and mixing rules. Use the NetApp Interoperability
Matrix Tool (IMT) to check compatibility.
Situations can change and issues arise over time. Performance issues can occur for many reasons. Performance analysis
can be complex and is beyond the scope of a fundamentals course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
The system log contains information and error messages that the storage system displays on the console and logs in
message files.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
For support information, documentation, software downloads, and access to My AutoSupport, see NetApp Support at
mysupport.netapp.com.
To access AutoSupport for your storage systems, see My AutoSupport at mysupport.netapp.com/myautosupport.
For system configuration information, see the NetApp Hardware Universe at hwu.netapp.com.
To determine the compatibility between various NetApp and officially supported third-party products, see the NetApp
IMT at mysupport.netapp.com/matrix.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 30 minutes
Access your lab
equipment.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When you observed the behavior in Iometer, were the benefits of storage QoS
apparent?
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
When you discuss data and data protection, you must first consider the currency of data. Assign a monetary value to the
data, based on the significance to the organization that owns the data. For example, the video of child's first steps is
important to the child’s family but might be of little value outside the family. However, the medical records of the same
child are of great importance to the health of the child, the family, and possibly many other people. The health records can
be used to identify, heal, or prevent health issues for the child, the family, and possibly other people around the globe.
Protecting a video or picture on a cellphone and protecting health records in a health network with many doctors and
hospitals present very different challenges.
Data currency is important when defining the terms of an SLA between the service provider and the customer. The two
terms most commonly used are recovery point objective (RPO), which is the maximum amount of acceptable data loss
during a failure, and recovery time objective (RTO), which is the maximum acceptable time that is required to make the
data available after a failure. Determining the RTO and RPO help to define the data protection solution or solutions that
meet particular SLA requirements.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Data consistency requirements vary widely depending on the workload requirements. Start by examining a single text file
on a share or volume. When you back up a file—for example, by using a NetApp ONTAP Snapshot copy—the file is
consistent in that point in time. The backup protects the file at a particular point in time, and if needed, you can restore the
file to that exact point. When ONTAP software creates a Snapshot copy, the copy is at the volume level, and so all the
files in a volume are backed up at the same time. As previously stated, for most file shares, this level of consistency is
adequate.
For block-level data from a host using SAN protocols, where the host controls the file system, consistency is required
between the host and the storage system. If the host writes data while the storage system is performing a backup, the data
consistency between the host and storage system might be compromised. The same is true with applications that write
structured data; for example, a database application data. For such workloads, transactional consistency is required.
Transactions must be paused or quiesced while the data is backed up. With ONTAP software, Snapshot copies are nearly
instantaneous and so the pause is brief, but the backup must be orchestrated between the host, application, and storage
system.
Server and desktop virtualization poses a unique challenge as multiple layers of data need to be protected. The host
administrator uses the virtualization software to create storage pools or containers on the storage system. The host
administrator uses the storage pools or containers to create virtual machines and virtual disks to present to the VMs.
Lastly, the administrator installs applications on the virtual machines, which in turn write data to the virtual disks. In a
virtualized environment, you need to consider the host and its data, the virtual machines and their data, and the
applications and their data. For virtual machines in particular, there are two consistency types: crash and application. The
difference between the types is whether only the virtual machine is backup-aware or whether both the virtual machine and
application are backup-aware, respectively.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Consider the different types or categories of data protection and the challenges that each poses.
High availability includes features that provide availability or takeover of resources if a component or controller fails.
High availability typically occurs within a data center.
Backup and archive includes features that back up or archive data locally or remotely.
Disaster recovery includes features that mirror data either locally or remotely. If a failure occurs at the mirror source (or
primary site), the data at the mirror destination (or disaster-recovery site) is made available. Disaster recovery is typically
considered a site-level protection and usually occurs between two data centers.
Compliance includes features that encrypt data or prevent data from being deleted or changed for a specified period.
Compliance features are typically used to comply with a regulation or policy requirement; for example the Sarbanes–
Oxley Act or the Health Insurance Portability and Accountability Act (HIPAA).
Cloud integration includes features that back up, restore, archive, or mirror data to a destination that is either in or near the
cloud.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Protection
NVRAM Write acknowledgement before committing to
disk
High-availability (HA) pairs Data availability if a controller fails
The listed features are part of ONTAP software and require no additional licensing.
The fundamentals of high availability is covered in the ONTAP Cluster Fundamentals course and is not discussed in this
course.
You can learn more about high availability administration in the ONTAP Cluster Administration course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Protection
Snapshot technology Point-in-time volume-level copy
The listed features listed are used to back up and archive data locally, remotely, or to tape. Snapshot copies, NDMP, and
SMTape are part of ONTAP software and require no additional licensing. Enabling SnapRestore and SnapVault software
requires licensing.
The fundamentals of Snapshot technology was covered in the ONTAP Cluster Fundamentals course and only a review is
provided in this course. This course focuses on when to use Snapshot copies or restore from a Snapshot copy by using
SnapRestore software. The course also discusses how SnapVault software can be used as a disk-to-disk backup solution.
You can learn more about Snapshot and SnapRestore administration in the ONTAP Cluster Administration course. In
addition, SnapVault administration and tape backups are covered in the ONTAP Data Protection Administration course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Protection
SnapMirror software Asynchronous volume-level data replication for
data movement and disaster recovery
FlexClone technology Instantaneous, space-efficient copies of
replicated data
Load-sharing mirrors Namespace (SVM root volume) protection
SyncMirror software Synchronous aggregate-level mirror
MetroCluster software Zero RTO/RPO disaster recovery
© 2016 NetApp, Inc. All rights reserved. Note: SVM=storage virtual machine 11
The listed features are used for disaster recovery. Load-sharing mirrors and SyncMirror software are part of ONTAP
software and require no additional licensing. Enabling the SnapMirror and FlexClone software requires licensing.
Flexible clones and load-sharing mirrors are discussed in the ONTAP Cluster Fundamentals and ONTAP NAS
Fundamentals courses, respectively, and are not discussed in this course.
This course focuses on SnapMirror, SnapVault, storage virtual machine (SVM) disaster recovery, NDMP, and tape
backup. We also discuss how SyncMirror and MetroCluster software work and where the technology is used.
You can learn more about FlexClone and load-sharing mirror administration in the ONTAP Cluster Administration
course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Protection
NetApp Storage Encryption (NSE) Full disk encryption (FDE) using a self-
encrypting disk (SED)
SnapLock compliance software WORM solution to meet external and internal
requirements for retaining, protecting, and
accessing regulated and reference data
The listed features are used for comprehensive encryption and retention of data at rest.
Compliance solutions are not covered in this course. You can learn more about compliance in the ONTAP Compliance
Solutions course.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Protection
NetApp Private Storage for Cloud (NPS for Dedicated, private NetApp storage (near-
Cloud solution) cloud)
NetApp Snap-to-Cloud disaster recovery Cloud-integrated data storage for disaster
solution recovery
NetApp AltaVault cloud-integrated storage Cloud-integrated backup and recovery
technology
The listed features are used for backup, archive, or disaster recovery in the cloud.
Although NetApp Snap-to-Cloud disaster recovery solution and NetApp Private Storage for Cloud (NPS for Cloud
solution) are not directly covered in this course, the knowledge you gain in the course can easily be transferred to those
solutions. In addition, this course focuses on ONTAP 9 Data Management Software, so NetApp AltaVault cloud-
integrated storage technology is not discussed. You can find more about AltaVault training by searching the NetApp
Learning Center.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Description
NetApp OnCommand System Manager Provides fast, simple configuration and management
for an ONTAP cluster
NetApp OnCommand Unified Manager Monitors the health and simplifies management of
multiple ONTAP clusters
NetApp OnCommand Workflow Automates storage tasks and data protection
Automation (NetApp WFA) processes
NetApp OnCommand APIs Integrates with third-party management solutions
The listed products are used to manage and monitor data protection solutions.
This course uses NetApp OnCommand System Manager only. You can find training for the other products by searching
the NetApp Learning Center.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Feature Description
SnapDrive data management software Automates storage and data management for
physical and virtual environments
SnapManager software Streamlines storage management and
simplifies configuration, backup, and restore
for enterprise operating environments
NetApp SnapCenter software Centralizes data protection and clone
management with a single interface across all
application environments
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
In addition to the NetApp data protection management software, which is written primarily for application or system
administrators, NetApp partners offer software that is written primarily for backup administrators.
For details on the listed partner products, visit the specific partner websites.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
NetApp Interoperability 15
Matrix Tool (IMT):
mysupport.netapp.com/matrix
Documentation:
mysupport.netapp.com
NetApp provides a variety of tools to help decide on a solution and to search for supported configurations.
The data protection assessment tool can help discover the NetApp data protection solution or solutions that best fit your
requirements. A link to the tool can be found on the data protection products page on the NetApp website.
The NetApp Interoperability Matrix Tool (IMT) is a web-based application that enables you to search for configurations
of NetApp products and components that meet the standards and requirements that NetApp specifies. To find data
protection solutions, click the Solutions Explorer link.
Documentation for the data protections solutions can be found on NetApp Support under the documentation tab.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Duration: 5 minutes
Instructor begins
polling session Instructor ends
polling session
Questions Instructor leads
appear in the debrief discussion
Correct
polling panel. answers have
Answer each Raise your hand
a green check to ask a question
question. mark.
When finished, or make a
Compare your comment.
click “Submit.” answers to the
correct
answers.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
Please take a few minutes to complete the survey for this course.
Your feedback is important for ensuring the quality of NetApp courses. Your instructor will give you instructions about
how to find the survey for this class and about how to use the survey web site.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
To measure your new knowledge of course topics, take the post-class assessment. You access the assessment via the link
that is provided.
https://www.brainshark.com/netapp/CDOTA_posttest
You can compare your pre-assessment score with your post-assessment score to measure how much you have learned. All
scores are private and are not retained or communicated.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.
© 2016 NetApp, Inc. This material is intended only for training. Reproduction is not authorized.