SlideShare una empresa de Scribd logo
1 de 12
Descargar para leer sin conexión
Red Hat Storage
An introduction to
Red Hat Storage architecture

Abstract	2
Red Hat Storage design goals	
2
Elasticity	2
Linear Scaling	2
Scale-Out with Red Hat Storage	4
Technical Differentiators	
7
Software Only	7
Open Source	7
Complete Storage Operating
System Stack 	7
User Space	7
Modular, Stackable
Architecture	7
Data Stored in Native Formats	8
No metadata with the
Elastic Hash Algorithm	8
Red Hat Storage global
namespace technology 	9
Conclusion	11
Glossary	11

www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 1

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

Abstract
Over the past ten years, enterprises have seen enormous gains in scalability, flexibility, and affordability as
they migrated from proprietary, monolithic server architectures to architectures that are virtualized, open
source, standardized, and commoditized.
Unfortunately, storage has not kept pace with computing. The proprietary, monolithic, and scale-up solutions that dominate the storage industry today do not deliver the scalability, flexibility, and economics that
modern datacenter and cloud computing environments need in a hyper-growth, virtualized, and increasingly
cloud-based world. The GlusterFS file system was created to address this gap.
Red Hat Storage is scale-out network attached storage (NAS) for private cloud or datacenter, public cloud,
and hybrid cloud environments. It is software-only, open source, and designed to meet unstructured data
storage requirements. It enables enterprises to combine large numbers of commodity storage and compute
resources into a high-performance, virtualized, and centrally managed storage pool. Both capacity and
performance can scale linearly and independently on-demand, from a few terabytes to petabytes and
beyond, using both on-premise commodity hardware and the public cloud compute and storage infrastructure. By combining commodity economics with a scale-out approach, Red Hat customers can achieve radically better price and performance in an easily deployed and managed solution that can be configured for
increasingly demanding workloads.
Red Hat Storage is built using the GlusterFS open source project as the foundation. This whitepaper
discusses some of the unique technical aspects of the Red Hat Storage architecture, speaking to those
aspects of the system that are designed to provide linear scale-out of both performance and capacity
without sacrificing resiliency.

Red Hat Storage design goals
Red Hat Storage was designed to achieve several major goals:
Elasticity
Elasticity is the notion that an enterprise should be able to flexibly adapt to the growth (or reduction) of data
and to add or remove resources to a storage pool as needed without disrupting the system. Red Hat Storage
was designed to allow enterprises to add or delete users, application data, volumes and storage nodes, etc.,
without disrupting any running functionality within the infrastructure.
Linear scaling
“Linear scaling” is a much-abused phrase within the storage industry. It should mean, for example, that
twice the amount of storage systems will deliver twice the realized performance—twice the throughput (as
measured in gigabytes per second) with the same average response time per external file system I/O event
(i.e., how long an NFS client will wait for the file server to return the information associated with each NFS
client request).
Similarly, if an organization has acceptable levels of performance, but wants to increase capacity, it should
be able to do so without decreasing performance or getting non-linear returns in capacity.

2 www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 2

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

Unfortunately, most storage systems do not demonstrate linear scaling. This seems somewhat counter-intuitive, since it is so easy to purchase another set of disks to double the size of available storage. The caveat in
doing so is that the scalability of storage has multiple dimensions, capacity being only one of them.
Adding capacity is only one dimension; the systems managing the disk storage need to scale as well. There
needs to be enough CPU capacity to drive all of the spindles at their peak capacity. The file system must
scale to support the total size. The metadata telling the system where all the files are located must scale at
the same rate disks are added. The network capacity available must scale to meet the increased number
of clients accessing those disks. In short, it is not storage that needs to scale as much as it is the complete
storage system that needs to scale.
Traditional file system models and architectures are unable to scale in this manner and therefore can never
achieve true linear scaling of performance. For traditional distributed systems, each storage node must
always incur the overhead of interacting with one or more other storage nodes for every file operation, and
that overhead subtracts from the scalability simply by adding to the list of tasks and the amount of work to
be done.
Even if those additional tasks could be done with near-zero effort (in the CPU and other system resources
sense of the term), latency problems remain. Latency results from waiting for the responses across the
networks connecting the distributed storage nodes in those traditional system architectures and nearly
always impacts performance. This type of latency increases proportionally relative to the speed and responsiveness—or lack of—of the networking connecting the nodes to each other. Attempts to minimize coordination overhead often result in unacceptable increases in risk.
This is why claims of linear scalability often break down for traditional distributed architectures.
Instead, as illustrated in Figure 1, most traditional systems demonstrate logarithmic scalability—storage’s
useful capacity grows more slowly as it gets larger. This is due to the increased overhead necessary to
maintain data resiliency. Examining the performance of some storage networks reflects this limitation as
larger units offer slower aggregate performance than their smaller counterparts.

Throughput

FIGURE 1

Performance and Capacity

www.redhat.com 3

RH_Storage_WP_8457567_1211_ma.indd 3

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

scale-out with Red Hat Storage
Red Hat Storage is designed to provide a scale-out architecture for both performance and capacity. This
implies that the system should be able to scale up (or down) along multiple dimensions.
By aggregating the disk, CPU, and I/O resources of large numbers of inexpensive systems, an enterprise
should be able to create one very large and high-performing storage pool. If the enterprise wants to add
more capacity to scale out a system, they can do so by adding more inexpensive disks. If the enterprise
wants to gain performance, they can do so by deploying more inexpensive sever nodes.
Red Hat Storage’s unique architecture is designed to deliver the benefits of scale-out (more units means
more capacity, more CPU, and more I/O), while avoiding the corresponding overhead and risk associated
with keeping large numbers of storage nodes in sync.
In practice, both performance and capacity can be scaled out linearly with Red Hat Storage. We do this by
employing three fundamental techniques:
•	The elimination of metadata
•	Effective distribution of data to achieve scalability and reliability
•	The use of parallelism to maximize performance via a fully distributed architecture
To illustrate how Red Hat Storage scales, Figure 2 shows how a baseline system can be scaled to increase
both performance and capacity. The discussion below uses some illustrative performance and capacity
numbers.
A typical direct-attached Red Hat Storage configuration will have a moderate number of disks attached to
two or more server nodes, which act as NAS heads (or storage nodes). For example, to support a requirement for 24 TB of capacity, a deployment might have two servers, each of which contains a quantity of 12
one-terabyte SATA drives. (See Config A.)
FIGURE 2

4 www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 4

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

If a customer has found that the performance levels are acceptable but wants to increase capacity by 25%,
they could add another four one-terabyte drives to each server and will not generally experience performance degradation. (i.e., each server would have 16 one-terabyte drives). (See Config B.) Note that they do
not need to upgrade to larger or more powerful hardware; they simply add eight more inexpensive SATA
drives.
On the other hand, if the customer is happy with 24 TB of capacity but wants to double performance, they
could distribute the drives among four servers, rather than two (i.e., each server would have six one-terabyte
drives, rather than 12). Note that in this case, they are adding two more low-price servers and can simply
redeploy existing drives. (See Config C.)
If they want to quadruple both performance and capacity, they could distribute among eight servers (i.e.,
each server would have 12 one-terabyte drives). (See Config D.)
Note that by the time a solution has approximately ten drives, the performance bottleneck has generally
already moved to the network. (See Config D.)
In order to maximize performance, we can upgrade from a 1-Gigabit Ethernet network to a 10-Gigabit
Ethernet network. Note that performance in this example is more than 25 times what we saw in the baseline.
This is evidenced by an increase in performance from 200 MB/s in the baseline configuration to 5,000 MB/s.
(See Config E.)
FIGURE 3

www.redhat.com 5

RH_Storage_WP_8457567_1211_ma.indd 5

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

FIGURE 4

As you will note, the power of the scale-out model is that both capacity and performance can scale linearly
to meet requirements. It is not necessary to know what performance levels will be needed two or three years
out. Instead, configurations can be easily adjusted as the need demands.
While the above discussion was using round, theoretical numbers, actual performance tests have proven
this linear scaling. The results, illustrated in Figure 5, show write throughput scaling linearly from 100 MB/s
on one server (e.g. storage node) to 800 MB/s (on eight systems) in a 1 GbE environment. However, on an
Infiniband network, we have seen write throughput scale from 1.5 GB/s (one system) to 12 GB/s (eight
systems).
Figure 5: Linear scaling In Red Hat Storage1

	1	
Results of scalability testing performed at Red Hat Storage running an Iozone test on eight clients connecting to between one and eight
server nodes. Total capacity was 13 TB. Network was 1 GbE. Servers were configured with single quad-core Intel Xeon CPUs and eight GB of
RAM. Clients had two GB of RAM.

6 www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 6

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

We have experience with Red Hat Storage being deployed in a multitude of scale-out scenarios. For example,
Red Hat Storage has been successfully deployed in multi-petabyte archival scenarios, where the goal was
moderate performance in the <$0.25/GB range. Additionally, Red Hat Storage has been deployed in very
high-performance production scenarios and has demonstrated throughput exceeding 22 GB/s

Technical differentiators
There are seven fundamental technical differentiators between Red Hat Storage and traditional storage
systems. These are discussed in brief below.
Software-only
We believe that storage is a software problem—one that cannot be solved by locking customers into a particular storage hardware vendor or a particular hardware configuration. We have designed Red Hat Storage
to work with a wide variety of industry-standard storage, networking, and compute server solutions. For
commercial customers, Red Hat Storage is delivered as virtual appliance, either packaged within an ISO or an
image deployed in a public cloud.
Open source
We believe that the best way to deliver functionality is by embracing the open source model. As a result,
Red Hat Storage users benefit from a worldwide community of thousands of developers who are constantly
testing the product in a wide range of environments and workloads, providing continuous feedback and
support—and providing unbiased feedback to other users. Red Hat development engineering sits at the
center of our broad-based community.
CoMPLETE storage operating system stACK
Our belief is that it’s important not only to deliver a distributed file system, but also to deliver a number of
other important functions in a distributed fashion. Red Hat Storage delivers distributed memory management, I/O scheduling, software RAID, self-healing, local N-way synchronous replication as well as asynchronous long-distance replication via Red Hat Geo-Replication. In essence, by taking a lesson from micro-kernel
architectures, we have designed Red Hat Storage to deliver a complete storage operating system stack in
user space.
User space
Unlike traditional file systems, Red Hat Storage operates in user space. This makes installing and upgrading
Red Hat Storage significantly easier. And it means that users who choose to develop on top of Red Hat
Storage need only have general C programming skills, not specialized kernel expertise.
Modular, stackable architecture
Red Hat Storage is designed using a modular and stackable architecture approach. To configure Red Hat
Storage for highly specialized environments (e.g., large number of large flies, huge numbers of very small
files, environments with cloud storage, various transport protocols, etc.), it is a simple matter of including or
excluding particular modules.
For the sake of stability, certain options should not be changed once the system is in use (for example, one
would not remove a function such as replication if high availability was a desired functionality).

www.redhat.com 7

RH_Storage_WP_8457567_1211_ma.indd 7

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

FIGURE 6: MODULAR AND STACKABLE ARCHITECTURE

•	Red Hat Storage is a complete storage stack in
userspace
•	Everything is a ‘volume’
•	Volume=C shared-object library
•	Stack volumes to create configuration
•	Customize to match unique workload needs
•	Open APIs
•	Volume-to-volume and Platform-to-world
•	Development is like application programming
•	Rapid time to market
•	Third party development

Data stored in native formats
With Red Hat Storage, data is stored on disk using native formats (e.g. EXT3, EXT4, XFS). Red Hat Storage
has implemented various self-healing processes for data. As a result, the system is extremely resilient.
Furthermore, files are naturally readable without Red Hat Storage. If a customer chooses to migrate away
from Red Hat Storage, their data is still completely usable without any required modifications or data
migration.
No metadata with the elastic hash algorithm
In a scale-out system, one of the biggest challenges is keeping track of the logical and physical location of
data (location metadata). Most distributed systems solve this problem by creating a separate index with file
names and location metadata. Unfortunately, this creates both a central point of failure and a huge performance bottleneck. As traditional systems add more files, more servers, or more disks, the central metadata
server becomes a performance chokepoint. This becomes an even bigger challenge if the workload consists
primarily of small files and the ratio of metadata to data increases.
Unlike other storage systems with a distributed file system, Red Hat Storage does not create, store, or use
a separate index of metadata in any way. Instead, Red Hat Storage places and locates files algorithmically.
All storage node servers in the cluster have the intelligence to locate any piece of data without looking it up
in an index or querying another server. All a storage node server needs to do to locate a file is to know the
pathname and filename and apply the algorithm. This fully parallelizes data access and ensures linear performance scaling. The performance, availability, and stability advantages of not using metadata are significant
and, in some cases, dramatic.

8 www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 8

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

Red Hat Storage global namespace technology
While many extol the virtues of their namespace capability as enabling easier management of network
storage, Red Hat Storage global namespace technology enables an even greater capability and has enabled
innovate IT solutions that are changing the way Red Hat customers leverage cloud technology and legacy
applications.
In the private cloud or datacenter
Red Hat Storage global namespace technology enables Red Hat customers who linearly scale their Red Hat
Storage NAS environment to tie together hundreds of storage nodes and associated files into one global
namespace. The result is one common mount point for a large pool of network attached storage. In some
cases, where the Red Hat Storage native access client is used in place of NFSv3, multiple parallel access is
supported by the file system.
In the public cloud, Red Hat Storage global namespace technology enables multiple compute instances and
storage to be configured in a massively scaleable pool of network attached storage. For example, within
the AWS cloud, Red Hat Storage global namespace enables the pooling of large quantities of Elastic Cloud
Compute (EC2) instances and Elastic Block Storage (EBS) to form NAS in the AWS cloud. EC2 instances
(storage nodes) and EBS can be added non-disruptively resulting in linear scaling of both performance and
capacity. Being able to deploy NAS in the cloud and run POSIX-compliant applications within the cloud using
Red Hat Storage file storeage accelerates cloud adoption and enables new and creative enterprise customer
business solutions.
Red Hat Storage global namespace technology goes one step further—it enables tying together both private
cloud/datacenters and public cloud NAS resources and enables enterprises to leverage a single mount point
across the private cloud/data center and public cloud environments. This innovative hybrid cloud capability
enables Red Hat customers to leverage the POSIX-compliant aspects of Red Hat Storage and their legacy
applications across vast storage resources.
Red Hat Storage Advanced topics
Elastic volume management
Given the elastic hashing approach assigns files to logical volumes, a question often arises: How do you
assign logical volumes to physical volumes?
In versions 3.1 and later of Red Hat Storage, volume management is elastic. Storage volumes are
abstracted from the underlying hardware and can grow, shrink, or be migrated across physical storage
nodes as systems as necessary. Storage node servers can be added or removed on-the-fly with data automatically rebalanced across the cluster. Data is always online, and there is no application downtime. File
system configuration changes are accepted at runtime and propagated throughout the cluster allowing
changes to be made dynamically as workloads fluctuate or for performance tuning.
Renaming or moving files
If a file is renamed, the hashing algorithm will obviously result in a different value, which will frequently
result in the file being assigned to a different logical volume, which might itself be located in a different
physical location.

www.redhat.com 9

RH_Storage_WP_8457567_1211_ma.indd 9

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

Since files can be large and rewriting and moving files is generally not a real-time operation, Red Hat
Storage solves this problem by creating a pointer at the time a file (or set of files) are renamed. Thus, a client
looking for a file under the new name would look in a logical volume and be redirected to the old logical
volume location. As background processes result in files ultimately being moved, the pointers are then
removed.
Similarly, if files need to be moved or reassigned (e.g. if a disk becomes hot or degrades in performance),
reassignment decisions can be made in real-time, while the physical migration of files can happen as a background process.
High availability
N-way Local Synchronous Replication
Generally speaking, we recommend the use of mirroring (2, 3, or n-way) to ensure availability. In this
scenario, each storage node server’s file data is replicated to another storage node server using synchronous writes. The benefits of this strategy are full fault-tolerance; failure of a single storage server is
completely transparent to Red Hat Storage clients. In addition, reads are spread across all members of the
mirror. Using Red Hat Storage, there can be an unlimited number of storage node members in a mirror.
While the elastic hashing algorithm assigns files to unique logical volumes, Red Hat Storage ensures that
every file is located on at least two different storage system server nodes.
Geo-rep long distance asynchronous replication
For long distance data replication requirements, Red Hat Storage supports Geo-Rep long distance replication. Customers can configure storage server nodes and Red Hat Storage to asynchronously replication data
over vast geographical distances.
Replication in the private cloud/datacenter, public cloud, and hybrid cloud environments
Both n-way synchronous and Geo-Rep asynchronous data replication are supported in the private cloud/
datacenter, public cloud, and hybrid cloud environments.
Within the AWS cloud, Red Hat Storage supports n-way synchronous replication across availability zones and
Geo-Rep asynchronous replication across AWS Regions. In fact, Red Hat Storage is the only way to ensure
high availability for NAS storage within the AWS infrastructure.
While Red Hat Storage offers software-level disk and server redundancy at the storage node server level,
in some cases we also recommend the use of hardware RAID (e.g., RAID 5 or 6) within individual storage
system servers to provide an additional level of protection where required. Red Hat solutions architects can
advise you on the best form of storage level and storage node level data protection and replication strategies for your specific requirements.

10 www.redhat.com

RH_Storage_WP_8457567_1211_ma.indd 10

12/6/11 4:58 PM
Red Hat Storage
An introduction to Red Hat Storage architecture

Conclusion
By delivering increased scalability, flexibility, affordability, performance, scalability, and ease-of-use in
concert with reduced cost of acquisition and maintenance, Red Hat Storage is a revolutionary step forward
in data management. Multiple advanced architectural design decisions make it possible for Red Hat
Storage to deliver great performance, greater flexibility, greater manageability, and greater resilience at a
significantly reduced overall cost. The complete elimination of location metadata via the use of the Elastic
Hashing Algorithm is at the heart of many of Red Hat Storage’s fundamental advantages, including its
remarkable resilience, which dramatically reduces the risk of data loss, data corruption, or data becoming
unavailable.
Red Hat Storage can be deployed in the private could or datacenter via Red Hat Storage Software Appliance
(SSA), an ISO image installed on commodity server and storage hardware, resulting in a powerful, turn-key,
massively scalable, and highly available NAS environment. Additionally, Red Hat Storage can be deployed
in the public cloud via Red Hat Virtual Storage Appliance, for example, within the AWS cloud, and delivers
all the features and functionally possible in the private cloud/datacenter to the public cloud—essentially
massively scalable and highly available NAS in the cloud.
To start a functional trial of Red Hat Storage, visit redhat.com. To speak with a Red Hat representative about
how to solve your particular storage challenges, call +1 (800) 805-5215.

Glossary
Block storage: Block special files or block devices correspond to devices through which the system moves
data in the form of blocks. These device nodes often represent addressable devices such as hard disks,
CD-ROM drives, or memory-regions. 2 Red Hat Storage supports most POSIX-compliant block level file
systems with extended attributes. Examples include ext3, ext4, and ZFS.
Distributed file system: Any file system that allows access to files from multiple hosts sharing via a
computer network. 3
Metadata: Data providing information about one or more other pieces of data.4
Namespace: An abstract container or environment created to hold a logical grouping of unique identifiers
or symbols. 5 Each Red Hat Storage cluster exposes a single namespace as a POSIX mount point that contains
every file in the cluster.

	2	
http://en.wikipedia.org/wiki/Device_file_system#Block_devices
	3	
http://en.wikipedia.org/wiki/Distributed_file_system
	4	
http://en.wikipedia.org/wiki/Metadata#Metadata_definition
	 5	 http://en.wikipedia.org/wiki/Namespace_%28computer_science%29

www.redhat.com 11

RH_Storage_WP_8457567_1211_ma.indd 11

12/6/11 4:58 PM
POSIX (Portable Operating System Interface [for UNIX]): A family of related standards specified by the
IEEE to define the application programming interface (API), along with shell and utilities interfaces for
software compatible with variants of the UNIX operating system. Red Hat Storage exports a fully POSIXcompliant file system.
RAID (Redundant Array of Inexpensive Disks): A technology that provides increased storage reliability
through redundancy, combining multiple low-cost, less-reliable disk drives components into a logical unit
where all drives in the array are interdependent.
Userspace: Applications running in user space don’t directly interact with hardware, instead using the kernel
to moderate access. Userspace applications are generally more portable than applications in kernel space.
Red Hat Storage is a user space application.
N-way replication: Local synchronous data replication typically deployed across campus or Amazon Web
Services Availability Zones.
Geo-Rep (Replication): Long-distance replication typically deployed from one private cloud/datacenter
to another, or one cloud region (i.e. AWS Region) to another located further than a 50 mile radius of the
primary data location.
CTDB: CTDB is primarily developed around the concept of having a shared cluster file system across all the
nodes in the cluster to provide the features required for building a NAS cluster. http://ctdb.samba.org/

Red Hat Sales and Inquiries
NORTH AMERICA
1–888–REDHAT1
www.redhat.com
sales@redhat.com

EUROPE, MIDDLE EAST
AND AFRICA
00800 7334 2835
www.europe.redhat.com
europe@redhat.com

ASIA PACIFIC
+65 6490 4200
www.apac.redhat.com
apac@redhat.com

Copyright © 2011 Red Hat, Inc. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, MetaMatrix,
and RHCE are trademarks of Red Hat, Inc., registered in the U.S. and other countries. Linux® is the registered
trademark of Linus Torvalds in the U.S. and other countries.

RH_Storage_WP_8457567_1211_ma.indd 12

LATIN AMERICA
+54 11 4329 7300
www.latam.redhat.com
info-latam@redhat.com

www.redhat.com

#8457567_1211

12/6/11 4:58 PM

Más contenido relacionado

La actualidad más candente

Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...
Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...
Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...EMC
 
EMC Isilon Best Practices for Hadoop Data Storage
EMC Isilon Best Practices for Hadoop Data StorageEMC Isilon Best Practices for Hadoop Data Storage
EMC Isilon Best Practices for Hadoop Data StorageEMC
 
Netezza Deep Dives
Netezza Deep DivesNetezza Deep Dives
Netezza Deep DivesRush Shah
 
FAQ on Dedupe NetApp
FAQ on Dedupe NetAppFAQ on Dedupe NetApp
FAQ on Dedupe NetAppAshwin Pawar
 
Massive parallel processing database systems mpp
Massive parallel processing database systems mppMassive parallel processing database systems mpp
Massive parallel processing database systems mppDiana Patricia Rey Cabra
 
White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...
 White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian... White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...
White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...EMC
 
Acunu Whitepaper v1
Acunu Whitepaper v1Acunu Whitepaper v1
Acunu Whitepaper v1Acunu
 
Data Domain Architecture
Data Domain ArchitectureData Domain Architecture
Data Domain Architecturekoesteruk22
 
The Database Environment Chapter 13
The Database Environment Chapter 13The Database Environment Chapter 13
The Database Environment Chapter 13Jeanie Arnoco
 
The Importance of Fast, Scalable Storage for Today’s HPC
The Importance of Fast, Scalable Storage for Today’s HPCThe Importance of Fast, Scalable Storage for Today’s HPC
The Importance of Fast, Scalable Storage for Today’s HPCIntel IT Center
 
How to choose a server for your data center's needs
How to choose a server for your data center's needsHow to choose a server for your data center's needs
How to choose a server for your data center's needsIT Tech
 
Hadoop World Vertica
Hadoop World VerticaHadoop World Vertica
Hadoop World VerticaOmer Trajman
 
EOUG95 - Client Server Very Large Databases - Paper
EOUG95 - Client Server Very Large Databases - PaperEOUG95 - Client Server Very Large Databases - Paper
EOUG95 - Client Server Very Large Databases - PaperDavid Walker
 
Ed richard resume
Ed richard resumeEd richard resume
Ed richard resumeed_richard
 
Ibm db2 analytics accelerator high availability and disaster recovery
Ibm db2 analytics accelerator  high availability and disaster recoveryIbm db2 analytics accelerator  high availability and disaster recovery
Ibm db2 analytics accelerator high availability and disaster recoverybupbechanhgmail
 
VNX with the Cloud Tiering Appliance
VNX with the Cloud Tiering Appliance VNX with the Cloud Tiering Appliance
VNX with the Cloud Tiering Appliance EMC
 

La actualidad más candente (20)

Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...
Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...
Data Warehouse Scalability Using Cisco Unified Computing System and Oracle Re...
 
IBM GPFS
IBM GPFSIBM GPFS
IBM GPFS
 
EMC Isilon Best Practices for Hadoop Data Storage
EMC Isilon Best Practices for Hadoop Data StorageEMC Isilon Best Practices for Hadoop Data Storage
EMC Isilon Best Practices for Hadoop Data Storage
 
Netezza Deep Dives
Netezza Deep DivesNetezza Deep Dives
Netezza Deep Dives
 
Whitepaper
WhitepaperWhitepaper
Whitepaper
 
GPFS Solution Brief
GPFS Solution BriefGPFS Solution Brief
GPFS Solution Brief
 
FAQ on Dedupe NetApp
FAQ on Dedupe NetAppFAQ on Dedupe NetApp
FAQ on Dedupe NetApp
 
Massive parallel processing database systems mpp
Massive parallel processing database systems mppMassive parallel processing database systems mpp
Massive parallel processing database systems mpp
 
White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...
 White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian... White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...
White Paper: Backup and Recovery of the EMC Greenplum Data Computing Applian...
 
Acunu Whitepaper v1
Acunu Whitepaper v1Acunu Whitepaper v1
Acunu Whitepaper v1
 
Data Domain Architecture
Data Domain ArchitectureData Domain Architecture
Data Domain Architecture
 
The Database Environment Chapter 13
The Database Environment Chapter 13The Database Environment Chapter 13
The Database Environment Chapter 13
 
The Importance of Fast, Scalable Storage for Today’s HPC
The Importance of Fast, Scalable Storage for Today’s HPCThe Importance of Fast, Scalable Storage for Today’s HPC
The Importance of Fast, Scalable Storage for Today’s HPC
 
How to choose a server for your data center's needs
How to choose a server for your data center's needsHow to choose a server for your data center's needs
How to choose a server for your data center's needs
 
Hadoop World Vertica
Hadoop World VerticaHadoop World Vertica
Hadoop World Vertica
 
EOUG95 - Client Server Very Large Databases - Paper
EOUG95 - Client Server Very Large Databases - PaperEOUG95 - Client Server Very Large Databases - Paper
EOUG95 - Client Server Very Large Databases - Paper
 
Ed richard resume
Ed richard resumeEd richard resume
Ed richard resume
 
Ibm db2 analytics accelerator high availability and disaster recovery
Ibm db2 analytics accelerator  high availability and disaster recoveryIbm db2 analytics accelerator  high availability and disaster recovery
Ibm db2 analytics accelerator high availability and disaster recovery
 
VNX with the Cloud Tiering Appliance
VNX with the Cloud Tiering Appliance VNX with the Cloud Tiering Appliance
VNX with the Cloud Tiering Appliance
 
Hpc compass 2013-final_web
Hpc compass 2013-final_webHpc compass 2013-final_web
Hpc compass 2013-final_web
 

Destacado

The big data_computing_architecture-graph500
The big data_computing_architecture-graph500The big data_computing_architecture-graph500
The big data_computing_architecture-graph500Accenture
 
Using splunk for_big_data
Using splunk for_big_dataUsing splunk for_big_data
Using splunk for_big_dataAccenture
 
Practical guide to cc
Practical guide to ccPractical guide to cc
Practical guide to ccAccenture
 
Ibm 1129-the big data zoo
Ibm 1129-the big data zooIbm 1129-the big data zoo
Ibm 1129-the big data zooAccenture
 
Big data 20120327
Big data 20120327Big data 20120327
Big data 20120327Accenture
 
Reporting demo
Reporting demoReporting demo
Reporting demoAccenture
 

Destacado (7)

The big data_computing_architecture-graph500
The big data_computing_architecture-graph500The big data_computing_architecture-graph500
The big data_computing_architecture-graph500
 
Using splunk for_big_data
Using splunk for_big_dataUsing splunk for_big_data
Using splunk for_big_data
 
Active dw
Active dwActive dw
Active dw
 
Practical guide to cc
Practical guide to ccPractical guide to cc
Practical guide to cc
 
Ibm 1129-the big data zoo
Ibm 1129-the big data zooIbm 1129-the big data zoo
Ibm 1129-the big data zoo
 
Big data 20120327
Big data 20120327Big data 20120327
Big data 20120327
 
Reporting demo
Reporting demoReporting demo
Reporting demo
 

Similar a 1 rh storage - architecture whitepaper

Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdf
Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdfDell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdf
Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdfhellobank1
 
NetApp Unified Scale-Out/Clustered Storage
NetApp Unified Scale-Out/Clustered StorageNetApp Unified Scale-Out/Clustered Storage
NetApp Unified Scale-Out/Clustered StorageNetApp
 
Run more applications without expanding your datacenter
Run more applications without expanding your datacenterRun more applications without expanding your datacenter
Run more applications without expanding your datacenterPrincipled Technologies
 
clusterstor-hadoop-data-sheet
clusterstor-hadoop-data-sheetclusterstor-hadoop-data-sheet
clusterstor-hadoop-data-sheetAndrei Khurshudov
 
Storage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 NotesStorage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 NotesSudarshan Dhondaley
 
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandWhy is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandINFINIDAT
 
SSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-ProcessingSSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-ProcessingSamsung Business USA
 
Red Hat Ceph Storage: Past, Present and Future
Red Hat Ceph Storage: Past, Present and FutureRed Hat Ceph Storage: Past, Present and Future
Red Hat Ceph Storage: Past, Present and FutureRed_Hat_Storage
 
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-SeagateLowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-SeagateMichael K. Connolly, MBATM
 
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storageI-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storageredpel dot com
 
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...Principled Technologies
 
What is Scalability and How can affect on overall system performance of database
What is Scalability and How can affect on overall system performance of databaseWhat is Scalability and How can affect on overall system performance of database
What is Scalability and How can affect on overall system performance of databaseAlireza Kamrani
 
From Rack scale computers to Warehouse scale computers
From Rack scale computers to Warehouse scale computersFrom Rack scale computers to Warehouse scale computers
From Rack scale computers to Warehouse scale computersRyousei Takano
 
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 Technology
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 TechnologyAdd Memory, Improve Performance, and Lower Costs with IBM MAX5 Technology
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 TechnologyIBM India Smarter Computing
 
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...IBM India Smarter Computing
 
Cluster Computers
Cluster ComputersCluster Computers
Cluster Computersshopnil786
 
Optimization on Key-value Stores in Cloud Environment
Optimization on Key-value Stores in Cloud EnvironmentOptimization on Key-value Stores in Cloud Environment
Optimization on Key-value Stores in Cloud EnvironmentFei Dong
 

Similar a 1 rh storage - architecture whitepaper (20)

Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdf
Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdfDell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdf
Dell_R730xd_RedHat_Ceph_Performance_SizingGuide_WhitePaper.pdf
 
NetApp Unified Scale-Out/Clustered Storage
NetApp Unified Scale-Out/Clustered StorageNetApp Unified Scale-Out/Clustered Storage
NetApp Unified Scale-Out/Clustered Storage
 
Hadoop Research
Hadoop Research Hadoop Research
Hadoop Research
 
Run more applications without expanding your datacenter
Run more applications without expanding your datacenterRun more applications without expanding your datacenter
Run more applications without expanding your datacenter
 
clusterstor-hadoop-data-sheet
clusterstor-hadoop-data-sheetclusterstor-hadoop-data-sheet
clusterstor-hadoop-data-sheet
 
Storage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 NotesStorage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 Notes
 
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandWhy is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
 
SSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-ProcessingSSDs Deliver More at the Point-of-Processing
SSDs Deliver More at the Point-of-Processing
 
Red Hat Ceph Storage: Past, Present and Future
Red Hat Ceph Storage: Past, Present and FutureRed Hat Ceph Storage: Past, Present and Future
Red Hat Ceph Storage: Past, Present and Future
 
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-SeagateLowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
 
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storageI-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
 
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...
Dell PowerEdge R750 servers featuring Dell PowerEdge RAID Controllers (PERC 1...
 
What is Scalability and How can affect on overall system performance of database
What is Scalability and How can affect on overall system performance of databaseWhat is Scalability and How can affect on overall system performance of database
What is Scalability and How can affect on overall system performance of database
 
From Rack scale computers to Warehouse scale computers
From Rack scale computers to Warehouse scale computersFrom Rack scale computers to Warehouse scale computers
From Rack scale computers to Warehouse scale computers
 
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 Technology
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 TechnologyAdd Memory, Improve Performance, and Lower Costs with IBM MAX5 Technology
Add Memory, Improve Performance, and Lower Costs with IBM MAX5 Technology
 
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
 
Ceph as software define storage
Ceph as software define storageCeph as software define storage
Ceph as software define storage
 
8 Strategies For Building A Modern DataCenter
8 Strategies For Building A Modern DataCenter8 Strategies For Building A Modern DataCenter
8 Strategies For Building A Modern DataCenter
 
Cluster Computers
Cluster ComputersCluster Computers
Cluster Computers
 
Optimization on Key-value Stores in Cloud Environment
Optimization on Key-value Stores in Cloud EnvironmentOptimization on Key-value Stores in Cloud Environment
Optimization on Key-value Stores in Cloud Environment
 

Más de Accenture

Certify 2014trends-report
Certify 2014trends-reportCertify 2014trends-report
Certify 2014trends-reportAccenture
 
Calabrio analyze
Calabrio analyzeCalabrio analyze
Calabrio analyzeAccenture
 
Tier 2 net app baseline design standard revised nov 2011
Tier 2 net app baseline design standard   revised nov 2011Tier 2 net app baseline design standard   revised nov 2011
Tier 2 net app baseline design standard revised nov 2011Accenture
 
Perf stat windows
Perf stat windowsPerf stat windows
Perf stat windowsAccenture
 
Performance problems on ethernet networks when the e0m management interface i...
Performance problems on ethernet networks when the e0m management interface i...Performance problems on ethernet networks when the e0m management interface i...
Performance problems on ethernet networks when the e0m management interface i...Accenture
 
NetApp system installation workbook Spokane
NetApp system installation workbook SpokaneNetApp system installation workbook Spokane
NetApp system installation workbook SpokaneAccenture
 
Migrate volume in akfiler7
Migrate volume in akfiler7Migrate volume in akfiler7
Migrate volume in akfiler7Accenture
 
Migrate vol in akfiler7
Migrate vol in akfiler7Migrate vol in akfiler7
Migrate vol in akfiler7Accenture
 
Data storage requirements AK
Data storage requirements AKData storage requirements AK
Data storage requirements AKAccenture
 
C mode class
C mode classC mode class
C mode classAccenture
 
Akfiler upgrades providence july 2012
Akfiler upgrades providence july 2012Akfiler upgrades providence july 2012
Akfiler upgrades providence july 2012Accenture
 
Net app virtualization preso
Net app virtualization presoNet app virtualization preso
Net app virtualization presoAccenture
 
Providence net app upgrade plan PPMC
Providence net app upgrade plan PPMCProvidence net app upgrade plan PPMC
Providence net app upgrade plan PPMCAccenture
 
WSC Net App storage for windows challenges and solutions
WSC Net App storage for windows challenges and solutionsWSC Net App storage for windows challenges and solutions
WSC Net App storage for windows challenges and solutionsAccenture
 
50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deploymentAccenture
 
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...Accenture
 
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11Accenture
 
Snap mirror source to tape to destination scenario
Snap mirror source to tape to destination scenarioSnap mirror source to tape to destination scenario
Snap mirror source to tape to destination scenarioAccenture
 
Ref arch for ve sg248155
Ref arch for ve sg248155Ref arch for ve sg248155
Ref arch for ve sg248155Accenture
 

Más de Accenture (20)

Certify 2014trends-report
Certify 2014trends-reportCertify 2014trends-report
Certify 2014trends-report
 
Calabrio analyze
Calabrio analyzeCalabrio analyze
Calabrio analyze
 
Tier 2 net app baseline design standard revised nov 2011
Tier 2 net app baseline design standard   revised nov 2011Tier 2 net app baseline design standard   revised nov 2011
Tier 2 net app baseline design standard revised nov 2011
 
Perf stat windows
Perf stat windowsPerf stat windows
Perf stat windows
 
Performance problems on ethernet networks when the e0m management interface i...
Performance problems on ethernet networks when the e0m management interface i...Performance problems on ethernet networks when the e0m management interface i...
Performance problems on ethernet networks when the e0m management interface i...
 
NetApp system installation workbook Spokane
NetApp system installation workbook SpokaneNetApp system installation workbook Spokane
NetApp system installation workbook Spokane
 
Migrate volume in akfiler7
Migrate volume in akfiler7Migrate volume in akfiler7
Migrate volume in akfiler7
 
Migrate vol in akfiler7
Migrate vol in akfiler7Migrate vol in akfiler7
Migrate vol in akfiler7
 
Data storage requirements AK
Data storage requirements AKData storage requirements AK
Data storage requirements AK
 
C mode class
C mode classC mode class
C mode class
 
Akfiler upgrades providence july 2012
Akfiler upgrades providence july 2012Akfiler upgrades providence july 2012
Akfiler upgrades providence july 2012
 
NA notes
NA notesNA notes
NA notes
 
Net app virtualization preso
Net app virtualization presoNet app virtualization preso
Net app virtualization preso
 
Providence net app upgrade plan PPMC
Providence net app upgrade plan PPMCProvidence net app upgrade plan PPMC
Providence net app upgrade plan PPMC
 
WSC Net App storage for windows challenges and solutions
WSC Net App storage for windows challenges and solutionsWSC Net App storage for windows challenges and solutions
WSC Net App storage for windows challenges and solutions
 
50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment
 
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...
Tr 3998 -deployment_guide_for_hosted_shared_desktops_and_on-demand_applicatio...
 
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11
Tr 3749 -net_app_storage_best_practices_for_v_mware_vsphere,_dec_11
 
Snap mirror source to tape to destination scenario
Snap mirror source to tape to destination scenarioSnap mirror source to tape to destination scenario
Snap mirror source to tape to destination scenario
 
Ref arch for ve sg248155
Ref arch for ve sg248155Ref arch for ve sg248155
Ref arch for ve sg248155
 

Último

Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)wesley chun
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdflior mazor
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024The Digital Insurer
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Tech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfTech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfhans926745
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...apidays
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?Antenna Manufacturer Coco
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 

Último (20)

Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Tech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfTech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdf
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 

1 rh storage - architecture whitepaper

  • 1. Red Hat Storage An introduction to Red Hat Storage architecture Abstract 2 Red Hat Storage design goals 2 Elasticity 2 Linear Scaling 2 Scale-Out with Red Hat Storage 4 Technical Differentiators 7 Software Only 7 Open Source 7 Complete Storage Operating System Stack 7 User Space 7 Modular, Stackable Architecture 7 Data Stored in Native Formats 8 No metadata with the Elastic Hash Algorithm 8 Red Hat Storage global namespace technology 9 Conclusion 11 Glossary 11 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 1 12/6/11 4:58 PM
  • 2. Red Hat Storage An introduction to Red Hat Storage architecture Abstract Over the past ten years, enterprises have seen enormous gains in scalability, flexibility, and affordability as they migrated from proprietary, monolithic server architectures to architectures that are virtualized, open source, standardized, and commoditized. Unfortunately, storage has not kept pace with computing. The proprietary, monolithic, and scale-up solutions that dominate the storage industry today do not deliver the scalability, flexibility, and economics that modern datacenter and cloud computing environments need in a hyper-growth, virtualized, and increasingly cloud-based world. The GlusterFS file system was created to address this gap. Red Hat Storage is scale-out network attached storage (NAS) for private cloud or datacenter, public cloud, and hybrid cloud environments. It is software-only, open source, and designed to meet unstructured data storage requirements. It enables enterprises to combine large numbers of commodity storage and compute resources into a high-performance, virtualized, and centrally managed storage pool. Both capacity and performance can scale linearly and independently on-demand, from a few terabytes to petabytes and beyond, using both on-premise commodity hardware and the public cloud compute and storage infrastructure. By combining commodity economics with a scale-out approach, Red Hat customers can achieve radically better price and performance in an easily deployed and managed solution that can be configured for increasingly demanding workloads. Red Hat Storage is built using the GlusterFS open source project as the foundation. This whitepaper discusses some of the unique technical aspects of the Red Hat Storage architecture, speaking to those aspects of the system that are designed to provide linear scale-out of both performance and capacity without sacrificing resiliency. Red Hat Storage design goals Red Hat Storage was designed to achieve several major goals: Elasticity Elasticity is the notion that an enterprise should be able to flexibly adapt to the growth (or reduction) of data and to add or remove resources to a storage pool as needed without disrupting the system. Red Hat Storage was designed to allow enterprises to add or delete users, application data, volumes and storage nodes, etc., without disrupting any running functionality within the infrastructure. Linear scaling “Linear scaling” is a much-abused phrase within the storage industry. It should mean, for example, that twice the amount of storage systems will deliver twice the realized performance—twice the throughput (as measured in gigabytes per second) with the same average response time per external file system I/O event (i.e., how long an NFS client will wait for the file server to return the information associated with each NFS client request). Similarly, if an organization has acceptable levels of performance, but wants to increase capacity, it should be able to do so without decreasing performance or getting non-linear returns in capacity. 2 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 2 12/6/11 4:58 PM
  • 3. Red Hat Storage An introduction to Red Hat Storage architecture Unfortunately, most storage systems do not demonstrate linear scaling. This seems somewhat counter-intuitive, since it is so easy to purchase another set of disks to double the size of available storage. The caveat in doing so is that the scalability of storage has multiple dimensions, capacity being only one of them. Adding capacity is only one dimension; the systems managing the disk storage need to scale as well. There needs to be enough CPU capacity to drive all of the spindles at their peak capacity. The file system must scale to support the total size. The metadata telling the system where all the files are located must scale at the same rate disks are added. The network capacity available must scale to meet the increased number of clients accessing those disks. In short, it is not storage that needs to scale as much as it is the complete storage system that needs to scale. Traditional file system models and architectures are unable to scale in this manner and therefore can never achieve true linear scaling of performance. For traditional distributed systems, each storage node must always incur the overhead of interacting with one or more other storage nodes for every file operation, and that overhead subtracts from the scalability simply by adding to the list of tasks and the amount of work to be done. Even if those additional tasks could be done with near-zero effort (in the CPU and other system resources sense of the term), latency problems remain. Latency results from waiting for the responses across the networks connecting the distributed storage nodes in those traditional system architectures and nearly always impacts performance. This type of latency increases proportionally relative to the speed and responsiveness—or lack of—of the networking connecting the nodes to each other. Attempts to minimize coordination overhead often result in unacceptable increases in risk. This is why claims of linear scalability often break down for traditional distributed architectures. Instead, as illustrated in Figure 1, most traditional systems demonstrate logarithmic scalability—storage’s useful capacity grows more slowly as it gets larger. This is due to the increased overhead necessary to maintain data resiliency. Examining the performance of some storage networks reflects this limitation as larger units offer slower aggregate performance than their smaller counterparts. Throughput FIGURE 1 Performance and Capacity www.redhat.com 3 RH_Storage_WP_8457567_1211_ma.indd 3 12/6/11 4:58 PM
  • 4. Red Hat Storage An introduction to Red Hat Storage architecture scale-out with Red Hat Storage Red Hat Storage is designed to provide a scale-out architecture for both performance and capacity. This implies that the system should be able to scale up (or down) along multiple dimensions. By aggregating the disk, CPU, and I/O resources of large numbers of inexpensive systems, an enterprise should be able to create one very large and high-performing storage pool. If the enterprise wants to add more capacity to scale out a system, they can do so by adding more inexpensive disks. If the enterprise wants to gain performance, they can do so by deploying more inexpensive sever nodes. Red Hat Storage’s unique architecture is designed to deliver the benefits of scale-out (more units means more capacity, more CPU, and more I/O), while avoiding the corresponding overhead and risk associated with keeping large numbers of storage nodes in sync. In practice, both performance and capacity can be scaled out linearly with Red Hat Storage. We do this by employing three fundamental techniques: • The elimination of metadata • Effective distribution of data to achieve scalability and reliability • The use of parallelism to maximize performance via a fully distributed architecture To illustrate how Red Hat Storage scales, Figure 2 shows how a baseline system can be scaled to increase both performance and capacity. The discussion below uses some illustrative performance and capacity numbers. A typical direct-attached Red Hat Storage configuration will have a moderate number of disks attached to two or more server nodes, which act as NAS heads (or storage nodes). For example, to support a requirement for 24 TB of capacity, a deployment might have two servers, each of which contains a quantity of 12 one-terabyte SATA drives. (See Config A.) FIGURE 2 4 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 4 12/6/11 4:58 PM
  • 5. Red Hat Storage An introduction to Red Hat Storage architecture If a customer has found that the performance levels are acceptable but wants to increase capacity by 25%, they could add another four one-terabyte drives to each server and will not generally experience performance degradation. (i.e., each server would have 16 one-terabyte drives). (See Config B.) Note that they do not need to upgrade to larger or more powerful hardware; they simply add eight more inexpensive SATA drives. On the other hand, if the customer is happy with 24 TB of capacity but wants to double performance, they could distribute the drives among four servers, rather than two (i.e., each server would have six one-terabyte drives, rather than 12). Note that in this case, they are adding two more low-price servers and can simply redeploy existing drives. (See Config C.) If they want to quadruple both performance and capacity, they could distribute among eight servers (i.e., each server would have 12 one-terabyte drives). (See Config D.) Note that by the time a solution has approximately ten drives, the performance bottleneck has generally already moved to the network. (See Config D.) In order to maximize performance, we can upgrade from a 1-Gigabit Ethernet network to a 10-Gigabit Ethernet network. Note that performance in this example is more than 25 times what we saw in the baseline. This is evidenced by an increase in performance from 200 MB/s in the baseline configuration to 5,000 MB/s. (See Config E.) FIGURE 3 www.redhat.com 5 RH_Storage_WP_8457567_1211_ma.indd 5 12/6/11 4:58 PM
  • 6. Red Hat Storage An introduction to Red Hat Storage architecture FIGURE 4 As you will note, the power of the scale-out model is that both capacity and performance can scale linearly to meet requirements. It is not necessary to know what performance levels will be needed two or three years out. Instead, configurations can be easily adjusted as the need demands. While the above discussion was using round, theoretical numbers, actual performance tests have proven this linear scaling. The results, illustrated in Figure 5, show write throughput scaling linearly from 100 MB/s on one server (e.g. storage node) to 800 MB/s (on eight systems) in a 1 GbE environment. However, on an Infiniband network, we have seen write throughput scale from 1.5 GB/s (one system) to 12 GB/s (eight systems). Figure 5: Linear scaling In Red Hat Storage1 1 Results of scalability testing performed at Red Hat Storage running an Iozone test on eight clients connecting to between one and eight server nodes. Total capacity was 13 TB. Network was 1 GbE. Servers were configured with single quad-core Intel Xeon CPUs and eight GB of RAM. Clients had two GB of RAM. 6 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 6 12/6/11 4:58 PM
  • 7. Red Hat Storage An introduction to Red Hat Storage architecture We have experience with Red Hat Storage being deployed in a multitude of scale-out scenarios. For example, Red Hat Storage has been successfully deployed in multi-petabyte archival scenarios, where the goal was moderate performance in the <$0.25/GB range. Additionally, Red Hat Storage has been deployed in very high-performance production scenarios and has demonstrated throughput exceeding 22 GB/s Technical differentiators There are seven fundamental technical differentiators between Red Hat Storage and traditional storage systems. These are discussed in brief below. Software-only We believe that storage is a software problem—one that cannot be solved by locking customers into a particular storage hardware vendor or a particular hardware configuration. We have designed Red Hat Storage to work with a wide variety of industry-standard storage, networking, and compute server solutions. For commercial customers, Red Hat Storage is delivered as virtual appliance, either packaged within an ISO or an image deployed in a public cloud. Open source We believe that the best way to deliver functionality is by embracing the open source model. As a result, Red Hat Storage users benefit from a worldwide community of thousands of developers who are constantly testing the product in a wide range of environments and workloads, providing continuous feedback and support—and providing unbiased feedback to other users. Red Hat development engineering sits at the center of our broad-based community. CoMPLETE storage operating system stACK Our belief is that it’s important not only to deliver a distributed file system, but also to deliver a number of other important functions in a distributed fashion. Red Hat Storage delivers distributed memory management, I/O scheduling, software RAID, self-healing, local N-way synchronous replication as well as asynchronous long-distance replication via Red Hat Geo-Replication. In essence, by taking a lesson from micro-kernel architectures, we have designed Red Hat Storage to deliver a complete storage operating system stack in user space. User space Unlike traditional file systems, Red Hat Storage operates in user space. This makes installing and upgrading Red Hat Storage significantly easier. And it means that users who choose to develop on top of Red Hat Storage need only have general C programming skills, not specialized kernel expertise. Modular, stackable architecture Red Hat Storage is designed using a modular and stackable architecture approach. To configure Red Hat Storage for highly specialized environments (e.g., large number of large flies, huge numbers of very small files, environments with cloud storage, various transport protocols, etc.), it is a simple matter of including or excluding particular modules. For the sake of stability, certain options should not be changed once the system is in use (for example, one would not remove a function such as replication if high availability was a desired functionality). www.redhat.com 7 RH_Storage_WP_8457567_1211_ma.indd 7 12/6/11 4:58 PM
  • 8. Red Hat Storage An introduction to Red Hat Storage architecture FIGURE 6: MODULAR AND STACKABLE ARCHITECTURE • Red Hat Storage is a complete storage stack in userspace • Everything is a ‘volume’ • Volume=C shared-object library • Stack volumes to create configuration • Customize to match unique workload needs • Open APIs • Volume-to-volume and Platform-to-world • Development is like application programming • Rapid time to market • Third party development Data stored in native formats With Red Hat Storage, data is stored on disk using native formats (e.g. EXT3, EXT4, XFS). Red Hat Storage has implemented various self-healing processes for data. As a result, the system is extremely resilient. Furthermore, files are naturally readable without Red Hat Storage. If a customer chooses to migrate away from Red Hat Storage, their data is still completely usable without any required modifications or data migration. No metadata with the elastic hash algorithm In a scale-out system, one of the biggest challenges is keeping track of the logical and physical location of data (location metadata). Most distributed systems solve this problem by creating a separate index with file names and location metadata. Unfortunately, this creates both a central point of failure and a huge performance bottleneck. As traditional systems add more files, more servers, or more disks, the central metadata server becomes a performance chokepoint. This becomes an even bigger challenge if the workload consists primarily of small files and the ratio of metadata to data increases. Unlike other storage systems with a distributed file system, Red Hat Storage does not create, store, or use a separate index of metadata in any way. Instead, Red Hat Storage places and locates files algorithmically. All storage node servers in the cluster have the intelligence to locate any piece of data without looking it up in an index or querying another server. All a storage node server needs to do to locate a file is to know the pathname and filename and apply the algorithm. This fully parallelizes data access and ensures linear performance scaling. The performance, availability, and stability advantages of not using metadata are significant and, in some cases, dramatic. 8 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 8 12/6/11 4:58 PM
  • 9. Red Hat Storage An introduction to Red Hat Storage architecture Red Hat Storage global namespace technology While many extol the virtues of their namespace capability as enabling easier management of network storage, Red Hat Storage global namespace technology enables an even greater capability and has enabled innovate IT solutions that are changing the way Red Hat customers leverage cloud technology and legacy applications. In the private cloud or datacenter Red Hat Storage global namespace technology enables Red Hat customers who linearly scale their Red Hat Storage NAS environment to tie together hundreds of storage nodes and associated files into one global namespace. The result is one common mount point for a large pool of network attached storage. In some cases, where the Red Hat Storage native access client is used in place of NFSv3, multiple parallel access is supported by the file system. In the public cloud, Red Hat Storage global namespace technology enables multiple compute instances and storage to be configured in a massively scaleable pool of network attached storage. For example, within the AWS cloud, Red Hat Storage global namespace enables the pooling of large quantities of Elastic Cloud Compute (EC2) instances and Elastic Block Storage (EBS) to form NAS in the AWS cloud. EC2 instances (storage nodes) and EBS can be added non-disruptively resulting in linear scaling of both performance and capacity. Being able to deploy NAS in the cloud and run POSIX-compliant applications within the cloud using Red Hat Storage file storeage accelerates cloud adoption and enables new and creative enterprise customer business solutions. Red Hat Storage global namespace technology goes one step further—it enables tying together both private cloud/datacenters and public cloud NAS resources and enables enterprises to leverage a single mount point across the private cloud/data center and public cloud environments. This innovative hybrid cloud capability enables Red Hat customers to leverage the POSIX-compliant aspects of Red Hat Storage and their legacy applications across vast storage resources. Red Hat Storage Advanced topics Elastic volume management Given the elastic hashing approach assigns files to logical volumes, a question often arises: How do you assign logical volumes to physical volumes? In versions 3.1 and later of Red Hat Storage, volume management is elastic. Storage volumes are abstracted from the underlying hardware and can grow, shrink, or be migrated across physical storage nodes as systems as necessary. Storage node servers can be added or removed on-the-fly with data automatically rebalanced across the cluster. Data is always online, and there is no application downtime. File system configuration changes are accepted at runtime and propagated throughout the cluster allowing changes to be made dynamically as workloads fluctuate or for performance tuning. Renaming or moving files If a file is renamed, the hashing algorithm will obviously result in a different value, which will frequently result in the file being assigned to a different logical volume, which might itself be located in a different physical location. www.redhat.com 9 RH_Storage_WP_8457567_1211_ma.indd 9 12/6/11 4:58 PM
  • 10. Red Hat Storage An introduction to Red Hat Storage architecture Since files can be large and rewriting and moving files is generally not a real-time operation, Red Hat Storage solves this problem by creating a pointer at the time a file (or set of files) are renamed. Thus, a client looking for a file under the new name would look in a logical volume and be redirected to the old logical volume location. As background processes result in files ultimately being moved, the pointers are then removed. Similarly, if files need to be moved or reassigned (e.g. if a disk becomes hot or degrades in performance), reassignment decisions can be made in real-time, while the physical migration of files can happen as a background process. High availability N-way Local Synchronous Replication Generally speaking, we recommend the use of mirroring (2, 3, or n-way) to ensure availability. In this scenario, each storage node server’s file data is replicated to another storage node server using synchronous writes. The benefits of this strategy are full fault-tolerance; failure of a single storage server is completely transparent to Red Hat Storage clients. In addition, reads are spread across all members of the mirror. Using Red Hat Storage, there can be an unlimited number of storage node members in a mirror. While the elastic hashing algorithm assigns files to unique logical volumes, Red Hat Storage ensures that every file is located on at least two different storage system server nodes. Geo-rep long distance asynchronous replication For long distance data replication requirements, Red Hat Storage supports Geo-Rep long distance replication. Customers can configure storage server nodes and Red Hat Storage to asynchronously replication data over vast geographical distances. Replication in the private cloud/datacenter, public cloud, and hybrid cloud environments Both n-way synchronous and Geo-Rep asynchronous data replication are supported in the private cloud/ datacenter, public cloud, and hybrid cloud environments. Within the AWS cloud, Red Hat Storage supports n-way synchronous replication across availability zones and Geo-Rep asynchronous replication across AWS Regions. In fact, Red Hat Storage is the only way to ensure high availability for NAS storage within the AWS infrastructure. While Red Hat Storage offers software-level disk and server redundancy at the storage node server level, in some cases we also recommend the use of hardware RAID (e.g., RAID 5 or 6) within individual storage system servers to provide an additional level of protection where required. Red Hat solutions architects can advise you on the best form of storage level and storage node level data protection and replication strategies for your specific requirements. 10 www.redhat.com RH_Storage_WP_8457567_1211_ma.indd 10 12/6/11 4:58 PM
  • 11. Red Hat Storage An introduction to Red Hat Storage architecture Conclusion By delivering increased scalability, flexibility, affordability, performance, scalability, and ease-of-use in concert with reduced cost of acquisition and maintenance, Red Hat Storage is a revolutionary step forward in data management. Multiple advanced architectural design decisions make it possible for Red Hat Storage to deliver great performance, greater flexibility, greater manageability, and greater resilience at a significantly reduced overall cost. The complete elimination of location metadata via the use of the Elastic Hashing Algorithm is at the heart of many of Red Hat Storage’s fundamental advantages, including its remarkable resilience, which dramatically reduces the risk of data loss, data corruption, or data becoming unavailable. Red Hat Storage can be deployed in the private could or datacenter via Red Hat Storage Software Appliance (SSA), an ISO image installed on commodity server and storage hardware, resulting in a powerful, turn-key, massively scalable, and highly available NAS environment. Additionally, Red Hat Storage can be deployed in the public cloud via Red Hat Virtual Storage Appliance, for example, within the AWS cloud, and delivers all the features and functionally possible in the private cloud/datacenter to the public cloud—essentially massively scalable and highly available NAS in the cloud. To start a functional trial of Red Hat Storage, visit redhat.com. To speak with a Red Hat representative about how to solve your particular storage challenges, call +1 (800) 805-5215. Glossary Block storage: Block special files or block devices correspond to devices through which the system moves data in the form of blocks. These device nodes often represent addressable devices such as hard disks, CD-ROM drives, or memory-regions. 2 Red Hat Storage supports most POSIX-compliant block level file systems with extended attributes. Examples include ext3, ext4, and ZFS. Distributed file system: Any file system that allows access to files from multiple hosts sharing via a computer network. 3 Metadata: Data providing information about one or more other pieces of data.4 Namespace: An abstract container or environment created to hold a logical grouping of unique identifiers or symbols. 5 Each Red Hat Storage cluster exposes a single namespace as a POSIX mount point that contains every file in the cluster. 2 http://en.wikipedia.org/wiki/Device_file_system#Block_devices 3 http://en.wikipedia.org/wiki/Distributed_file_system 4 http://en.wikipedia.org/wiki/Metadata#Metadata_definition 5 http://en.wikipedia.org/wiki/Namespace_%28computer_science%29 www.redhat.com 11 RH_Storage_WP_8457567_1211_ma.indd 11 12/6/11 4:58 PM
  • 12. POSIX (Portable Operating System Interface [for UNIX]): A family of related standards specified by the IEEE to define the application programming interface (API), along with shell and utilities interfaces for software compatible with variants of the UNIX operating system. Red Hat Storage exports a fully POSIXcompliant file system. RAID (Redundant Array of Inexpensive Disks): A technology that provides increased storage reliability through redundancy, combining multiple low-cost, less-reliable disk drives components into a logical unit where all drives in the array are interdependent. Userspace: Applications running in user space don’t directly interact with hardware, instead using the kernel to moderate access. Userspace applications are generally more portable than applications in kernel space. Red Hat Storage is a user space application. N-way replication: Local synchronous data replication typically deployed across campus or Amazon Web Services Availability Zones. Geo-Rep (Replication): Long-distance replication typically deployed from one private cloud/datacenter to another, or one cloud region (i.e. AWS Region) to another located further than a 50 mile radius of the primary data location. CTDB: CTDB is primarily developed around the concept of having a shared cluster file system across all the nodes in the cluster to provide the features required for building a NAS cluster. http://ctdb.samba.org/ Red Hat Sales and Inquiries NORTH AMERICA 1–888–REDHAT1 www.redhat.com sales@redhat.com EUROPE, MIDDLE EAST AND AFRICA 00800 7334 2835 www.europe.redhat.com europe@redhat.com ASIA PACIFIC +65 6490 4200 www.apac.redhat.com apac@redhat.com Copyright © 2011 Red Hat, Inc. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, MetaMatrix, and RHCE are trademarks of Red Hat, Inc., registered in the U.S. and other countries. Linux® is the registered trademark of Linus Torvalds in the U.S. and other countries. RH_Storage_WP_8457567_1211_ma.indd 12 LATIN AMERICA +54 11 4329 7300 www.latam.redhat.com info-latam@redhat.com www.redhat.com #8457567_1211 12/6/11 4:58 PM