Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Virrrr

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 26

UNIT-4 STORAGE VIRTUALIZATION

Storage virtualization is a major component for storage servers, in the form of functional
RAID levels and controllers. Applications and operating systems on the device can directly
access the discs for writing. Local storage is configured by the controllers in RAID groups,
and the operating system sees the storage based on the configuration. However, the storage is
abstracted and the controller is determining how to write the data or retrieve the requested
data for the operating system. Storage virtualization is important in various other forms:
 File servers: The operating system doesn't need to know how to write to physical
media; it can write data to a remote location.
 WAN Accelerators: WAN accelerators allow you to provide re-requested blocks at
LAN speed without affecting WAN performance. This eliminates the need to transfer
duplicate copies of the same material over WAN environments.
 SAN and NAS: Storage is presented over the Ethernet network of the operating
system. NAS (Network Attached Storage) presents the storage as file operations (like
NFS). SAN technologies present the storage as block level storage (like Fibre
Channel). SAN (Storage Area Network) technologies receive the operating
instructions only when if the storage was a locally attached device.
 Storage Tiering: Analysing the most frequently used data and allocating it to the
best-performing storage pool, storage tiering uses the storage pool concept as an entry
point. The least used data is stored in the storage pool with the lowest performance.

Advantages of Storage Virtualization


 Data is stored in the more convenient locations away from the specific host. In the
case of a host failure, the data is not compromised necessarily
 The storage devices can perform advanced functions like replication, reduplication,
and disaster recovery functionality.
 By doing abstraction of the storage level, IT operations become more flexible in how
storage is provided, partitioned, and protected.

Drawbacks of Storage Virtualization


 Agility and scalability: Storage virtualization cannot always be a smooth
implementation. It comes with few technical hurdles, such as scalability.
 Data security: Data security also remains a concern. Though some may argue that
virtual machines and servers are more secure than physical ones, virtual
environments can attract new kinds of cyber-attacks.
 Manageability and integration: Virtualisation breaks the end-to-end view of your
data. The virtualized storage solution must be capable of integrating with existing
tools and systems.

Memory Virtualization:
 Memory virtualization gathers volatile random access memory (RAM) resources from
many data centre systems, making them accessible to any cluster member machine.
 Software performance issues commonly occur from physical memory limits. Memory
virtualization solves this issue by enabling networked, and hence distributed, servers
to share a pool of memory. Applications can utilise a vast quantity of memory to
boost system utilisation, enhance memory usage efficiency, and open up new use
cases when this feature is integrated into the network.
 Shared memory systems and memory virtualization solutions are different. Because
shared memory systems do not allow memory resources to be abstracted, they can
only be implemented with a single instance of an operating system (that is, not in a
clustered application environment).
 Memory virtualization differs from flash memory-based storage, like solid-state
drives (SSDs), in that the former replaces or enhances regular RAM, while the latter
replaces hard drives (networked or not).
 Products based on Memory Virtualization are: ScaleMP, RNA Networks Memory
Virtualization Platform, Oracle Coherence and GigaSpaces.

Implementaions
Application level integration
In this case, applications running on connected computers connect to the memory pool
directly through an API or the file system.

Operating System Level Integration


In this case, the operating system connects to the memory pool, and makes pooled memory
available to applications.

Features
1. Virtual Address Space: Creating a virtual address space for each programme that
corresponds to a physical memory address is the first stage in memory virtualization.
While physical memory addresses are often bigger than virtual address spaces, numerous
applications can run simultaneously.

2. Page Tables: The operating system keeps track of the memory pages used by each app
and their matching physical memory addresses in order to manage the mapping between
virtual and physical memory addresses. This data structure is known as a page table.

3. Memory Paging: A page fault occurs when an application tries to access a memory page
that is not already in physical memory. The OS reacts to this by loading the requested
page from disc into physical memory and swapping out a page of memory from physical
memory to disc.

4. Memory compression: Different memory compression algorithms, which analyse the


contents of memory pages and compress them to conserve space, are used to make better
use of physical memory. A compressed page is instantly decompressed when a
programme wants to access it.

5. Memory Over commitment: Memory over commitment, in which applications are given
access to more virtual memory than is physically accessible, is made possible by
virtualization. because not all memory pages are actively being used at once, the System
can employ memory paging and compression to release physical memory as needed.

6. Memory Ballooning: Several virtualization technologies utilise a method called


ballooning to further minimise memory use. This entails dynamically modifying the
memory allotted to each virtual machine in accordance with its usage trends. The
hypervisor can reclaim some of a virtual machine’s allocated memory if it is not being
fully utilised and make it accessible to other virtual machines.

Benefits of Memory Virtualization


1. Increased Address Space: It allows processes to utilize a larger address space than what is
physically available, enabling the execution of larger programs or multiple programs
concurrently.

2. Memory Isolation: Each process has its own virtual memory space, which provides
memory isolation and protects processes from interfering with each other’s memory.

3. Simplified Memory Management: Memory virtualization simplifies memory


management for both the operating system and application developers. It abstracts away the
details of physical memory allocation and allows for more flexible memory usage.

4. Efficient Memory Utilization: By using techniques like demand paging and page
replacement, memory virtualization optimizes the usage of physical memory by keeping
frequently accessed pages in memory and swapping out less used pages to disk.

Types Of Storage Virtualization:


1. Block-Level: When you write to a hard drive on your desktop computer, it writes
directly to the hard disk. This is block-level storage. When you use virtualized block
storage, the server acts as a desktop computer and can access virtual disks, which act
like regular hard drives. This gives you benefits such as booting off of a block device
along with increased performance and scalability.

2. Object-Level: Data is not immediately stored on a disc when using object storage.
Data buckets are used to abstract it instead. You can retrieve this data from your
programme using API (programme Programming Interface) calls. This may be a more
scalable option than block storage when dealing with big data volumes. Hence, after
arranging your buckets, you won't need to be concerned about running out of room.

3. File-Level: When someone wants another server to host their data, they use file
server software such as Samba and NFS. The files are kept in directories known as
shares. As a result, this eliminates the requirement for disc space management and
permits numerous users to share a storage device. File servers are useful for desktop
PCs, virtual servers, and servers.

4. Host-based: Access to the host or any connected devices is made possible via host-
based storage virtualization. The server's installed driver intercepts and reroutes the
input and output (IO) requests. These input/output (IO) requests are typically sent
towards a hard disc, but they can also be directed towards other devices, including a
USB flash drive. This kind of storage is mostly used for accessing actual installation
CDs or DVDs, which make it simple to install an operating system on the virtual
computer.

5. Network-based: The host and the storage are separated by a fibre channel switch.
The virtualization takes place and the IO requests are redirected at the switch. No
specific drivers are needed for this approach to function on any operating system.

6. Array-based: All of the arrays' IO requests are handled by a master array. This
makes data migrations easier and permits management from a single location.

Block-Level Storage Virtualization:


 Block-level storage virtualization is a storage service that provides a flexible, logical
arrangement of storage capacity to applications and users while abstracting its
physical location. As a software layer, it intercepts I/O requests to that logical
capacity and maps them to the appropriate physical locations.
 Block-level storage virtualization is a technology that abstracts physical storage
devices into a virtualized layer, providing a unified and simplified view of storage
resources. This virtualization occurs at the block level, where data is divided into
fixed-size blocks, typically ranging from a few kilobytes to several megabytes. Each
block is assigned a unique address, and the storage virtualization layer manages the
mapping of these addresses to physical storage locations.

Key Aspects Of Block-Level Storage Virtualization


1. Abstraction and Pooling:

Abstraction: Block-level virtualization abstracts the underlying physical storage devices,


presenting them as a single, logical storage pool.
Pooling: Multiple storage devices, such as hard disk drives (HDDs) or solid-state drives
(SSDs), can be pooled together to create a larger and more flexible storage resource.

2. Storage Virtualization Layer: A storage virtualization layer sits between the applications
and the physical storage devices. It manages the allocation and retrieval of data blocks,
providing a transparent interface to the applications.

3. Uniform Addressing: Each block of data is assigned a unique address within the
virtualized storage space. This allows for consistent addressing regardless of the physical
location of the data.

4. Dynamic Provisioning: Block-level virtualization enables dynamic provisioning of storage


space. Storage can be allocated or de-allocated on-the-fly without disrupting ongoing
operations.

5. Data Migration and Load Balancing: The virtualization layer can facilitate data migration
across different storage devices without affecting the applications using the data. This helps in
load balancing and optimizing storage performance.

6. Improving Utilization and Efficiency: By pooling and dynamically allocating storage


resources, block-level storage virtualization improves overall storage utilization, ensuring that
available storage capacity is used efficiently.

7. Vendor Independence: Users can often mix and match storage devices from different
vendors within the virtualized storage pool. This promotes vendor independence and
flexibility in choosing hardware components.

8. Snapshot and Backup: Many block-level storage virtualization solutions offer features like
snapshots and backups. Snapshots allow for point-in-time copies of data, and backup
processes can be streamlined through centralized management.

9. Centralized Management: Administrators can centrally manage the storage infrastructure,


monitor performance, and implement policies for data protection and access control.

10. Scalability: Block-level storage virtualization is scalable, allowing organizations to easily


expand their storage infrastructure by adding new devices to the virtualized pool.
Benefits of Block-Level Storage Virtualization:
1. Centralized Management: Administrators can manage storage resources centrally,
streamlining tasks such as provisioning, monitoring, and data migration.

2. Improved Utilization: Virtualization allows for efficient use of storage capacity, as it


enables pooling and dynamic allocation of resources based on demand.

3. Vendor Independence: Users can integrate storage devices from different vendors into a
unified storage pool, promoting flexibility and preventing vendor lock-in.

4. Scalability: Block-level storage virtualization is scalable, enabling organizations to easily


expand their storage infrastructure by adding new devices to the virtualized pool.

5. Data Migration and Load Balancing: The virtualization layer facilitates seamless data
migration across storage devices, aiding in load balancing and optimizing storage
performance.

Drawbacks of Block-Level Storage Virtualization:


1. Complexity: Implementing and managing a block-level storage virtualization solution can
be complex. It may require specialized knowledge and skills.

2. Performance Overhead: Depending on the virtualization implementation, there may be


some level of performance overhead introduced, potentially impacting the speed of data
access.
3. Initial Setup Costs: The initial investment in virtualization infrastructure, including
hardware and software, can be significant.

4. Compatibility Issues: Integrating storage devices from different vendors may lead to
compatibility issues or require additional effort to ensure seamless operation.

5. Security Concerns: Centralized management of storage resources requires robust security


measures to protect against unauthorized access and data breaches.

In summary, while block-level storage virtualization offers numerous benefits in terms of


flexibility, efficiency, and centralized management, organizations should carefully consider
the associated complexities, costs, and potential drawbacks. The decision to implement storage
virtualization should align with the specific needs and goals of the organization.

File-Level Storage Virtualization:


 File-level virtualization is a method that operates at the file system layer, which is the
level that organizes and manages the files and directories on a storage device.
 File-level virtualization allows multiple file systems to be pooled together and
accessed as a single namespace, regardless of their physical location, size, or format.
This can simplify the administration and migration of files, as well as provide load
balancing and fault tolerance
 File-level storage virtualization is a technology that abstracts physical file storage
systems and presents a unified view of files and directories to users and applications.
Unlike block-level storage virtualization, which operates at a lower level with data
blocks, file-level virtualization deals with entire files and the hierarchical structure of
file systems.

Key Aspects Of File-Level Storage Virtualization:


1. Abstraction of Physical Storage: File-level storage virtualization abstracts the
underlying physical storage systems, presenting them as a single, logical file system.

2. Unified Namespace: It provides a unified namespace for files and directories, allowing
users and applications to interact with a centralized and standardized file system.

3. Hierarchical Structure: File-level virtualization maintains the hierarchical structure of


files and directories, resembling traditional file systems.

4. Access Control and Security: Administrators can implement access control and
security policies at the file level, managing permissions for individual files or directories.

5. Dynamic Expansion and Contraction: The virtualization layer allows for dynamic
expansion or contraction of storage resources, making it easier to manage changing storage
requirements.

6. Data Migration: File-level virtualization facilitates the movement of files across


different storage devices or locations without affecting how users or applications access
the data.
7. Centralized Management: Administrators can centrally manage file storage, monitor
usage, and implement policies for data protection and access control.

8. Vendor Independence: Similar to block-level virtualization, file-level virtualization


allows the integration of storage devices from different vendors into a unified file system.

9. Compatibility with Network Attached Storage (NAS): File-level virtualization is


often used in conjunction with Network Attached Storage (NAS) environments, where it
can simplify storage management across multiple NAS devices.

Benefits of File-Level Storage Virtualization:


1. Simplified Data Management: Users and applications interact with a single, unified
file system, simplifying data management and reducing the complexities associated with
multiple storage systems.

2. Improved Scalability: File-level virtualization supports dynamic expansion and


contraction of storage resources, improving scalability to meet changing storage needs.
3. Efficient Data Migration: Files can be migrated between storage devices without
affecting user access, facilitating efficient data movement for load balancing or hardware
upgrades.

4. Centralized Control: Administrators have centralized control over file-level


permissions, security settings, and storage policies.

5. Enhanced Access Control: Access control can be applied at the file level, allowing for
fine-grained permissions management.

Drawbacks of File-Level Storage Virtualization:


1. Performance Overhead: Depending on the implementation, file-level virtualization
may introduce some performance overhead, potentially affecting data access speeds.

2. Complexity: Implementing and managing file-level virtualization solutions can be


complex, requiring specialized knowledge and skills.

3. Compatibility Challenges: Some legacy applications or systems may not fully support
file-level virtualization, leading to compatibility challenges.

4. Initial Setup Costs: There can be significant initial setup costs associated with
implementing file-level virtualization, including hardware and software investments.

5. Learning Curve: Adopting file-level storage virtualization may involve a learning


curve for administrators, especially if they are not familiar with the specific virtualization
solution.

File-level storage virtualization is often employed in scenarios where simplifying storage


management, supporting scalable file systems, and providing centralized control over file
access are critical requirements. Organizations considering file-level virtualization should
carefully evaluate their specific needs and weigh the benefits against potential drawbacks.

Address Space Remapping:


 Virtualization of storage helps achieve location independence by abstracting the
physical location of the data. The virtualization system presents to the user a logical
space for data storage and handles the process of mapping it to the actual physical
location.
 It is possible to have multiple layers of virtualization or mapping. It is then possible
that the output of one layer of virtualization can then be used as the input for a higher
layer of virtualization. Virtualization maps space between back-end resources, to
front-end resources. In this instance, "back-end" refers to a logical unit number (LUN)
that is not presented to a computer, or host system for direct use. A "front-end" LUN
or volume is presented to a host or computer system for use.
 The actual form of the mapping will depend on the chosen implementation. Some
implementations may limit the granularity of the mapping which may limit the
capabilities of the device. Typical granularities range from a single physical disk
down to some small subset (multiples of megabytes or gigabytes) of the physical disk.
 In a block-based storage environment, a single block of information is addressed
using a LUN identifier and an offset within that LUN – known as a logical block
addressing (LBA).

Address space remapping is a technique used in storage virtualization to manage and


control the mapping of logical addresses to physical storage locations. This process
involves dynamically associating logical addresses, which are used by applications or
the operating system, with physical addresses on the storage devices. Address space
remapping plays a crucial role in achieving flexibility, efficiency, and abstraction in
storage environments. Below are key aspects and considerations related to address
space remapping in storage virtualization:

Considerations and Challenges:


1. Performance Impact: The remapping process may introduce some performance
overhead, depending on the implementation. Storage virtualization solutions aim to
minimize this impact to ensure efficient data access.

2. Complexity: Managing the dynamic mapping of addresses adds complexity to


storage virtualization systems. Administrators must carefully configure and monitor
the remapping process.

3. Compatibility: Compatibility with existing applications and systems may be a


consideration. Address space remapping should be transparent to applications to
ensure a smooth integration.

4. Security: Security measures must be in place to protect the mapping information


and prevent unauthorized access or tampering with the address space remapping
process.

5. Data Integrity: Ensuring data integrity during address space remapping is crucial.
The virtualization layer must guarantee that data is correctly mapped to the intended
physical locations.

Risk Of Storage Virtualization:


The major challenges linked with storage virtualisation are as follows:

 Managing the different software and hardware can get difficult when there are several
hardware and software elements.
 Storage systems need frequent upgradation to meet the challenging nature of
applications and huge data.
 Despite the ease of accessing data with storage virtualisation, there is always a risk of
cyber-attacks and various cyber threats in virtual environments. That is, for the data
stored in virtual machines, data security and its governance are the major challenges.
 Amongst the various vendors delivering storage virtualisation solutions, it’s important
to find a reliable one. As many a time, it happens when vendors provide storage
solutions but ignore the complexities of backing up virtual storage pools.
 Similarly, they fall in situations when there is a need for immediate recovery of data
in case of hardware failure or any other issue.
 Storage virtualisation, at times, can lead to access issues. This can be if the LAN
connection is disrupted, or internet access is lost due to some reason.
 There comes a time when there is a need to switch from a smaller network to a larger
one, as the capacity of the current one is insufficient. The migration process is time-
consuming and can even result in downtime.
 Additionally, problems like more significant data analysis, lack of agility, scalability,
and more rapid access to data are the common challenges companies face while
selecting storage solutions.

Storage Area Network(SAN):


 A Storage Area Network (SAN) is a network of storage devices that can be
accessed by multiple servers or computers, providing a shared pool of storage
space. Each computer on the network can access storage on the SAN as though
they were local disks connected directly to the computer.
 A SAN is typically assembled with cabling, host bus adapters, and SAN switches
attached to storage arrays and servers. Each switch and storage system on the SAN
must be interconnected.

SANs are often used to:

 Improve application availability (e.g., multiple data paths),


 Enhance application performance (e.g., off-load storage functions, segregate or zone
networks, etc.),
 Increase storage utilization and effectiveness (e.g., consolidate storage resources,
provide tiered storage, etc.), and improve data protection and security.

 A SAN presents storage devices to a host such that the storage appears to be locally
attached. This simplified presentation of storage to a host is accomplished through the
use of different types of virtualization.
 SANs perform an important role in an organization's Business Continuity Management
(BCM) activities (e.g., by spanning multiple sites).
 SANs are commonly based on a switched fabric technology. Examples include Fibre
Channel (FC), Ethernet, and InfiniBand. Gateways may be used to move data between
different SAN technologies.
 Fibre Channel is commonly used in enterprise environments. Fibre Channel may be used
to transport SCSI, NVMe, FICON, and other protocols.
 Ethernet is commonly used in small and medium sized organizations. Ethernet
infrastructure can be used for SANs to converge storage and IP protocols onto the same
network. Ethernet may be used to transport SCSI, FCoE, NVMe, RDMA, and other
protocols.
 InfiniBand is commonly used in high performance computing environments. InfiniBand
may be used to transport SRP, NVMe, RDMA, and other protocols.

SAN Fabric Architecture and Operation:

 The core of a SAN is its fabric: the scalable, high-performance network that interconnects
hosts -- servers -- and storage devices or subsystems. The design of the fabric is directly
responsible for the SAN's reliability and complexity. At its simplest, an FC SAN can simply
attach HBA ports on servers directly to corresponding ports on SAN storage arrays, often
using optical cables for top speed and support for networking over greater physical
distances.
 But such simple connectivity schemes belay the true power of a SAN. In actual practice, the
SAN fabric is designed to enhance storage reliability and availability by eliminating single
points of failure. A central strategy in creating a SAN is to employ a minimum of two
connections between any SAN elements. The goal is to ensure that at least one working
network path is always available between SAN hosts and SAN storage.

SAN architecture includes host components, fabric components and storage components.

 Consider a simple example in the image above where two SAN hosts must communicate with
two SAN storage subsystems. Each host employs a separate HBA -- not a multiport HBA
because the HBA device itself is a single point of failure. The port from each HBA is
connected to a port on a different SAN switch, such as Fibre Channel switch. Similarly,
multiple ports on the SAN switch connect to different storage target devices or systems.
This is a simple redundant fabric; remove any one connection in the diagram,
and both servers can still communicate with both storage systems to preserve storage access
for the workloads on both servers.
 Consider the basic behaviour of a SAN and its fabric. A host server requires access to SAN
storage; the host will internally create a request to access the storage device. The traditional
SCSI commands used for storage access are encapsulated into packets for the network -- in
this case FC packets -- and the packets are structured according to the rules of the FC
protocol. The packets are delivered to the host's HBA where the packets are placed onto the
network's optical or copper cables. The HBA transmits the request packet(s) to the SAN
where the request will arrive at the SAN switch(s). One of the switches will receive the
request and send it along to the corresponding storage device. In a storage array, the storage
processor will receive the request and interact with storage devices within the array to
accommodate the host's request.

SAN Switches:

 The SAN switch is the focal point of any SAN. As with most network switches, the SAN
switch receives a data packet, determines the source and destination of the packet and
then forwards that packet to the intended destination device. Ultimately, the SAN fabric
topology is defined by number of switches, the type of switches -- such as backbone
switches, or modular or edge switches -- and the way in which the switches are
interconnected. Smaller SANs might use modular switches with 16, 24 or even 32 ports,
while larger SANs might use backbone switches with 64 or 128 ports. SAN switches can
be combined to create large and complex SAN fabrics that connect thousands of servers
and storage devices.

Alternative SAN Approaches:

 Virtual SAN. Virtualization technology was a natural fit for the SAN, encompassing
both storage and storage network resources to add flexibility and scalability to the
underlying physical SAN. A virtual SAN -- denoted with a capital V in VSAN -- is a
form of isolation, reminiscent of traditional SAN zoning, which essentially uses
virtualization to create one or more logical partitions or segments within the physical
SAN. Traditional VSANs can employ such isolation to manage SAN network traffic,
enhance performance and improve security. Thus, VSAN isolation can prevent potential
problems on one segment of the SAN from affecting other SAN segments, and the
segments can be changed logically as needed without the need to touch any physical
SAN components. VMware offers Virtual SAN Technology.

 Unified SAN. A SAN is noted for its support of block storage, which is typical for
enterprise applications. But file, object and other types of storage would traditionally
demand a separate storage system, such as network-attached storage (NAS). A SAN that
supports unified storage is capable of supporting multiple approaches -- such as file,
block and object-based storage -- within the same storage subsystem. Unified storage
provides such capabilities by handling multiple protocols, including file-based SMB and
NFS, as well as block-based, such as FC and iSCSI. By using a single storage platform
for block and file storage, users can take advantage of powerful features that are usually
reserved for traditional block-based SANs, such as storage snapshots, data replication,
storage tiering, data encryption, data compression and data deduplication.
 Converged SAN. One common disadvantage to a traditional FC SAN is the cost and
complexity of a separate network dedicated to storage. ISCSI is one means of
overcoming the cost of a SAN by using common Ethernet networking components rather
than FC components. FCoE supports a converged SAN that can run FC communication
directly over Ethernet network components -- converging both common IP and FC
storage protocols onto a single low-cost network. FCoE works by encapsulating FC
frames within Ethernet frames to route and transport FC data across an Ethernet network.
However, FCoE relies on end-to-end support in network devices, which has been
difficult to achieve on a broad basis, making the choice of vendor limited.

 Hyper-converged infrastructure. The data center use of HCI has grown dramatically in
recent years. HCI combines compute and storage resources into pre-packaged modules,
allowing modules -- also called nodes -- to be added as needed and managed through a
single common utility. HCI employs virtualization, which abstracts and pools all the
compute and storage resources. IT administrators then provision virtual machines and
storage from the available resource pools. The fundamental goal of HCI is to simplify
hardware deployment and management while allowing fast scalability.
SAN Benefits:

 High performance. The typical SAN uses a separate network fabric that is dedicated to
storage tasks. The fabric is traditionally FC for top performance, though iSCSI and
converged networks are also available.
 High scalability. The SAN can support extremely large deployments encompassing
thousands of SAN host servers and storage devices or even storage systems. New hosts
and storage can be added as required to build out the SAN to meet the organization's
specific requirements.
 High availability. A traditional SAN is based on the idea of a network fabric, which --
ideally -- interconnects everything to everything else. This means a full-featured SAN
deployment has no single point of failure between a host and a storage device, and
communication across the fabric can always find an alternative path to maintain storage
availability to the workload.
 Advanced management features. A SAN will support an array of useful enterprise-class
storage features, including data encryption, data deduplication, storage replication and
self-healing technologies intended to maximize storage capacity, security and data
resilience. Features are almost universally centralized and can easily be applied to all the
storage resources on the SAN.

SAN Disadvantages:
 Complexity. Although more convergence options, such as FCoE and unified options,
exist for SANs today, traditional SANs present the added complexity of a second
network -- complete with costly, dedicated HBAs on the host servers, switches and
cabling within a complex and redundant fabric and storage processor ports at the
storage arrays. Such networks must be designed and monitored with care, but the
complexity is increasingly troublesome for IT organizations with fewer staff and
smaller budgets.
 Scale. Considering the cost, a SAN is generally effective only in larger and more
complex environments where there are many servers and significant storage. It's
certainly possible to implement a SAN on a small scale, but the cost and complexity
are difficult to justify. Smaller deployments can often achieve satisfactory results
using an iSCSI SAN, a converged SAN over a single common network -- such as
FCoE -- or an HCI deployment, which is adept at pooling and provisioning resources.
 Management. With the idea of complexity focused on hardware, there is also
significant challenge in SAN management. Configuring features, such as LUN
mapping or zoning, can be problematic for busy organizations. Setting up RAID and
other self-healing technologies as well as corresponding logging and reporting -- not
to mention security -- can be time-consuming but unavoidable to maintain the
organization's compliance, DR and BC postures.

Network Attached Storage (NAS):


 An NAS device is a storage device connected to a network that allows storage and
retrieval of data from a central location for authorised network users and varied
clients. NAS devices are flexible and scale out, meaning that as you need additional
storage, you can add to what you have. NAS is like having a private cloud in the
office. It’s faster, less expensive and provides all the benefits of a public cloud on site,
giving you complete control.
 NAS devices typically don't have a keyboard or display and are configured and
managed with a browser-based utility. Each NAS resides on the LAN as an
independent network node, defined by its own unique IP address.
NAS Uses:

 The purpose of network-attached storage is to enable users to collaborate and


share data more effectively. It is useful to distributed teams that need remote
access or work in different time zones. NAS connects to a wireless router, making
it easy for distributed workers to access files from any desktop or mobile device
with a network connection.
 Some NAS products are designed for use in large enterprises. Others are for home
offices or small businesses. Devices usually contain at least two drive bays,
although single-bay systems are available for noncritical data.
 In addition, most NAS vendors partner with cloud storage providers to give
customers the flexibility of redundant backup.
 Network-attached storage relies on hard disk drives (HDDs) to serve data. I/O
contention can occur when too many users overwhelm the system with requests at
the same time.
 Higher-end NAS products have enough disks to support redundant arrays of
independent disks, or RAID, which is a storage configuration that turns multiple
hard disks into one logical unit to boost performance, high availability and
redundancy.

NAS Components:

 CPU. The heart of every NAS is a computer that includes the central processing
unit (CPU) and memory. The CPU is responsible for running the NAS OS,
reading and writing data against storage, handling user access and even
integrating with cloud storage if so designed. Where typical computers or servers
use a general-purpose CPU, a dedicated device such as NAS might use a
specialized CPU designed for high performance and low power consumption in
NAS use cases.
 Network interface. Small NAS devices designed for desktop or single-user use
might allow for direct computer connections, such as USB or limited wireless
(Wi-Fi) connectivity. But any business NAS intended for data sharing and file
serving will demand a physical network connection, such as a cabled Ethernet
interface, giving the NAS a unique IP address. This is often considered part of the
NAS hardware suite, along with the CPU.
 Storage. Every NAS must provide physical storage, which is typically in the
form of disk drives. The drives might include traditional magnetic HDDs, SSDs
or other non-volatile memory devices, often supporting a mix of different storage
devices. The NAS might support logical storage organization for redundancy and
performance, such as mirroring and other RAID implementations -- but it's the
CPU, not the disks, that handle such logical organization.
 OS. Just as with a conventional computer, the OS organizes and manages the
NAS hardware and makes storage available to clients, including users and other
applications. Simple NAS devices might not highlight a specific OS, but more
sophisticated NAS systems might employ a discrete OS such as Netgear
ReadyNAS, QNAP QTS, Zyxel FW, among others.

Types and Alternatives of NAS:


Scale up and scale out are two versions of NAS. Object storage is an alternative to
NAS for handling unstructured data.
Scale-up NAS
In a network-attached storage deployment, the NAS head is the hardware that
performs the control functions. It provides access to back-end storage through an
internet connection. This configuration is known as scale-up architecture. A two-
controller system expands capacity with the addition of drive shelves, depending on
the scalability of the controllers.
Scale-out NAS
With scale-out systems, the storage administrator installs larger heads and more hard
disks to boost storage capacity. Scaling out provides the flexibility to adapt to an
organization's business needs. Enterprise scale-out systems can store billions of files
without the performance tradeoff of doing metadata searches.

Object storage
Some industry experts speculate that object storage will overtake scale-out NAS.
However, it's possible the two technologies will continue to function side by side.
Both scale-out and object storage methodologies deal with scale, but in different
ways.

NAS files are centrally managed via the Portable Operating System Interface
(POSIX). It provides data security and ensures multiple applications can share a
scale-out device without fear that one application will overwrite a file being
accessed by other users.

Object storage is a new method for easily scalable storage in web-scale


environments. It is useful for unstructured data that is not easily compressible,
particularly large video files.

Object storage does not use POSIX or any file system. Instead, all the objects are
presented in a flat address space. Bits of metadata are added to describe each object,
enabling quick identification within a flat address namespace.

Advantages of NAS:

 Many NAS products are multiprotocol which gives users flexibility


 Multiple users can access at any time.
 Offers data backup capabilities like replication and redundancy.
 Provides 24/7 and remote data availability.
 Easy to operate
 Offers data protection capabilities such as RAID.

Disadvantages of NAS:

 Requires special skills to manage high-end NAS.


 Does not work well with block or object storage.
 Performance can increase with increased traffic.
 Not well suited for highly transactional environments.
 Scalability is limited compared to other storage systems.
 Protocol admins choose may affect performance.
Redudant Array Of Indepemdent Disks:

• Technology used in computer systems to organize and manage multiple physical hard
drives as a single logical unit.
• RAID is designed to improve the reliability, performance, and/or capacity of data
storage systems.
• It achieves this by storing data across multiple disks in a way that provides
redundancy and/or data striping.
• There are different levels of RAID, each with its own set of characteristics and
advantages.

• Some common RAID levels include

 RAID 0
 RAID 1
 RAID 2
 RAID 3
 RAID 4
 RAID 5
 RAID 6
RAID 0:
• RAID 0 implements data striping.

• The data blocks are placed in multiple disks without redundancy. None of the disks
are used for data redundancy so if one disk fails
then all the data in the array is lost. No data block is being repeated
in any disk
Block '10,11,12,13' form a stripe

Instead of placing one block of data in a disk, we can place more than one block of data in a
disk and then move to another disk.
Pros of RAID 0:
• All the disk space is utilized and hence performance is increased.
• Data requests can be on multiple disks and not on a single disk hence improving the
throughput.
Cons of RAID 0:
• Failure of one disk can lead to complete data loss in the respective array.
• No data Redundancy is implemented so one disk failure can lead to system failure.

RAID 1:
• RAID 1 implements mirroring which means the data of one disk is replicated in
another disk.

• This helps in preventing system failure as if one disk fails then the redundant disk
takes over.

Here Disk 0 and Disk 1 have the same data as disk 0 is copied to disk 1. Same is the case
with Disk 2 and Disk 3.

Pros of RAID 1:
• Failure of one Disk does not lead to system failure as there is redundant data in other
disk.

Cons of RAID 1:
• Extra space is required for each disk as each disk data is copied to some other disk
also

RAID 2:

• RAID 2 is used when error in data has to be checked at bit level, which uses a
Hamming code detection method.

• Two disks are used in this technique.

• One is used to store bit of each word in the disk and another is used to store error code
correction (Parity bits) of data words.

• The structure of this RAID is complex, so it is not used commonly.


Here Disk 3, Disk 4 and Disk 5 stores the parity bits of Data stored in Disk 0, Disk 1, and
Disk 2 respectively. Parity bits are used to detect the error in data.

Pros of RAID 2:

• It checks for error at a bit level for every data word.

• One full disk is used to store parity bits which helps in detecting error.

Cons of RAID 2:

• Large extra space is used for parity bit storage.

RAID 3:

• RAID 3 implements byte-level striping of Data.

• Data is stored across disks with their parity bits in a separate disk. The parity bits
helps to reconstruct the data when there is a data loss.

Here Disk 3 contains the Parity bits for Disk 0 Disk 1 and Disk 2. If any one of the Disk's
data is lost the data can be reconstructed using parity bits in Disk 3.
Pros of RAID 3:

 Data can be recovered with the help of parity bits

Cons of RAID 3:

• Extra space for storing parity bits is used.

RAID 4:

• RAID 4 implements block-level striping of data with dedicated parity drive.

• If only one of the data is lost in any disk, then it can be reconstructed with the help of
parity drive.

• Parity is calculated with the help of XOR operation over each data disk block.

Here P0 is calculated using XOR (0,1,0) = 1 and P1 is calculated using XOR (1,1,0) = 0. If
there is even number of 1 then XOR is 0 and for odd number of 1 XOR is 1. If suppose Disk
0 data is lost, by checking parity P0=1 we will know that Disk 0 should have 0 to make the
Parity P0 as 1 whereas if there was 1 in Disk 0 it would have made the parity P0=0 which
contradicts with the current parity value.

Pros of RAID 4:

• Parity bits helps to reconstruct the data if at most one data is lost from the disks.

Cons of RAID 4:

• Extra space for Parity is required.

• If there is more than one data loss from multiple disks then Parity cannot help us
reconstruct the data.

RAID 5:

• RAID 5 is similar to RAID 4 with only one difference.

• The parity Rotates among the Disks.


Pros of RAID 5:

• Parity is distributed over the disk and makes the performance better.

• Data can be reconstructed using parity bits.

Cons of RAID 5:

• Parity bits are useful only when there is data loss in at most one Disk.

• If there is loss in more than one Disk block then parity is of no use.

• Extra space for parity is required.

RAID 6:

• If there is more than one Disk failure, then RAID 6 implementation helps in that case.

• In RAID 6 there are two parity in each array/row. It is similar to RAID 5 with extra
parity.

Here P0,P1,P2,P3 and Q0,Q1,Q2,Q3 are two parity to reconstruct the data if almost two disks
fail.
Pros of RAID 6:

• More parity helps in reconstructing at most 2 Disk data.

Cons of RAID 6:

• Extra space is used for both parities. (P and Q).

• More than 2 disk failures can not be corrected.

In Summary, RAID is used to backup the data when a disk fails for some reason and there are
some levels of RAID.

• RAID 0 implements data striping.

• RAID 1 implements mirroring which creates redundant data.

• RAID 2 uses Hamming code Error Detection method to correct error in data.

• RAID 3 does byte-level data striping and has parity bits for each data word.

• RAID 4 does block-level data striping.

• RAID 5 has rotating parity across the disks.

• RAID 6 has two parity which can handle at most two disk failures.

You might also like