Hitachi Dynamic Tiering (HDT) simplifies storage administration by automatically optimizing data placement in 1, 2 or 3 tiers of storage that can be defined and used within a single virtual volume. Tiers of storage can be made up of internal or external (virtualized) storage, and use of HDT can lower capital costs. Simplified and unified management of HDT allows for lower operational costs and reduces the challenges of ensuring applications are placed on the appropriate classes of storage.
2. WEBTECH EDUCATIONAL SERIES
OVERVIEW OF HITACHI DYNAMIC TIERING, PART 1 OF 2
Hitachi Dynamic Tiering (HDT) simplifies storage administration by automatically
optimizing data placement in 1, 2 or 3 tiers of storage that can be defined and used within
a single virtual volume. Tiers of storage can be made up of internal or external
(virtualized) storage, and use of HDT can lower capital costs. Simplified and unified
management of HDT allows for lower operational costs and reduces the challenges of
ensuring applications are placed on the appropriate classes of storage.
By attending this webcast, you will
•
Hear about what makes Hitachi Dynamic Tiering a unique storage management tool
that enables storage administrators to meet performance requirements at lower costs
than traditional tiering methods.
•
Understand various strategies to consider when monitoring application performance
and relocating pages to appropriate tiers without manual intervention.
•
Learn how to use Hitachi Command Suite (HCS) to manage, monitor and report on an
HDT environment, and how HCS manages related storage environments.
3. UPCOMING WEBTECHS
WebTechs
‒ Hitachi Dynamic Tiering: An In-Depth Look at Managing HDT and
Best Practices, Part 2, November 13, 9 a.m. PT, noon ET
‒ Best Practices for Virtualizing Exchange for Microsoft Private
Cloud, December 4, 9 a.m. PT, noon ET
Check www.hds.com/webtech for
Links to the recording, the presentation, and Q&A (available next
week)
Schedule and registration for upcoming WebTech sessions
Questions will be posted in the HDS Community:
http://community.hds.com/groups/webtech
5. AGENDA
Hitachi Dynamic Tiering
‒ Relation to Hitachi Dynamic Provisioning
‒ Monitoring I/O activity
‒ Relocating pages (data)
‒ Tiering policies
‒ Managing and monitoring HDT environments with Hitachi
Command Suite
6. HITACHI DYNAMIC PROVISIONING
MAINFRAME AND OPEN SYSTEMS
Virtualize devices into a pool of capacity
and allocate by pages
Dynamically provision new servers in
seconds
HDP Volume
(Virtual LUN)
Eliminate allocated but unused waste by
allocating only the pages that are used
Extend Dynamic Provisioning to external
virtualized storage
HDP Pool
LDEVs
Convert fat volumes into thin volumes by
moving them into the pool
Optimize storage performance by
spreading the I/O across more arms
Up to 62,000 LUNs in a single pool
Up to 5PB support
Dynamically expand or shrink pool
Zero page reclaim
LDEV LDEV LDEV LDEV LDEV LDEVLDEV LDEV
7. VIRTUAL STORAGE PLATFORM:
PAGE-LEVEL TIERING
POOL A
Different tiers of storage are now in
EFD/SSD
TIER 1
1 pool of pages
Data is written to the highest-
Least
Refer
enced
performance tier first
SAS 2
TIER
As data becomes less active, it
migrates to lower-level tiers
If activity increases, data will be
promoted back to a higher tier
Since 20% of data accounts for
80% of the activity, only the active
part of a volume will reside on the
higher-performance tiers
Least
Refer
enced
SATA3
TIER
8. VIRTUAL STORAGE PLATFORM:
PAGE-LEVEL TIERING
Automatically detects and assigns
POOL A
EFD/SSD
TIER 1
tiers based on media type
Dynamically
Add or remove tiers
Least Referenced
SAS 2
TIER
Expand or shrink tiers
Expand LUNs
Move LUNs between pools
Automatically adjust sub-LUN
42MB pages between tiers based
on captured metadata
Supports virtualized storage and all
replication/DR solutions
Least Referenced
SATA3
TIER
10. HDT: POLICY-BASED MONITORING
AND RELOCATION
Manual mode
‒ Monitoring and relocation separately controlled
‒ Can set complex schedules to custom fit to
priority work periods
Media Groupings
Supported by VSP*
Order of
Grouping
SSD
1
SAS 15K RPM
2
SAS 10K RPM
3
SAS 7.2K RPM
4
SATA
5
‒ Sampling at ½-, 1-, 2-, 4-, or 8-hour intervals
External #1
6
‒ All aligned to midnight
External #2
7
‒ May select automatic monitoring of I/O intensity
and automatic data relocation
External #3
8
Automatic mode
‒ Customer defines strategy; it is then executed
automatically
‒ 24-hour sampling
‒ Allows for custom selection of partial day periods
* VSP = Hitachi Virtual Storage Platform
11. PERIOD AND CONTINUOUS MONITORING
Impacts Relocation Decisions and How Tier Properties Are Displayed
Period mode
Continuous mode
Relocation uses just the I/O load measurements from the last
completed monitor cycle.
Relocation uses a weighted average of previous cycles. Shortterm I/O load increases or decreases have less influence on
relocation
Period Mode
Continuous Mode
Weighted calculation
Load
Load
105
10
95
93
100
I/O load info
91
per monitoring
cycle
100
Actual I/O load
105
81
84
87
I/O load by weighted
calculation
Time
Relocation executed
based on current I/O load
86
I/O load information
per monitoring cycle
by weighted
calculation
Time
Relocation executed
based on
weighted calculation
12. MONITORING AND RELOCATION OPTIONS
Execution
mode
Cycle
duration
Monitoring
Start
End
Start
End
Auto
24 hours
After setting
auto
execution to
ON, next 0:00
is reached
After
monitoring
started, the
next 0:00 is
reached
Starts
immediately
after
monitoring
data is fixed
One of the following
• Relocation of entire
pool is finished
• Next relocation is
started
• Auto execution is set to
OFF
execution
Time of
day not
specified
Relocation
24 hours
with
time of
day
specified
execution
See
RAIDCOM
command
Specified end
time is reached
30 min.
1 hour
2 hours
4 hours
8 hours
Manual
After setting
auto
execution to
ON, the
specified start
time is
reached
After setting
auto
execution to
ON, cycle
time begins
when
0:00 is
reached
After
monitoring
started, cycle
time is reached
Variable
Request to
start
monitoring is
received
SN2,
RAIDCOM, or
HCS
Request to end
monitoring is
received
Above
Monitoring/relocation cycle
Above
1/1
00:00
1/2
00:00
t
Monitoring
monitor data
for relocate
Relocation
[Ex.] Monitoring period 9:00-17:00
1/1
1/2
9 17
1/3
9 17
t
Above
Above
[Ex.] Monitoring period 8h
1/1
0
Request to
start
relocation is
received
SN2,
RAIDCOM, or
HCS
One of the following
• Relocation of entire
pool finished
• Request to stop
relocation is received
• Auto execution is set to
ON
• Subsequent manual
monitoring is stopped
8
1/2
16 0
8
1/3
16 0
t
Request to Request to Request to
stop
start
start
monitoring monitoring relocation
t
13. HDT PERFORMANCE MONITORING
Back-end I/O (read plus write) counted per
page during the monitor period
IOPH
25
Monitor ignores “RAID I/O” (parity I/O)
Monitoring
20
Count of IOPH for the cycle (period mode)
or a weighted average (continuous mode)
DP-VOLs
15
HDT orders pages by counts high to low
to create a distribution function
‒ IOPH vs. GB
Monitor analysis is performed to determine the
IOPH values that separate the tiers
10
5
0
Page 1
25
Page 999
Pool
20
Aggregate the data
15
Analysis
10
5
0
Capacity 1
Capacity nnn
14. POOL TIER PROPERTIES
What is being used
now in the pool in
terms of capacity
and performance
Can display just the
performance graph
for a tiering policy
The I/O
distribution across
all pages in the
pool. Combined
with the tier
range, HDT
decides where the
pages should go
15. HITACHI DYNAMIC TIERING
HDT Pool
TIER 1
SSD
Frequent Accesses
Dynamic
Provisioning
Virtual
Volume
Infrequent References
TIER 2
SAS
TIER 3
SATA
What determines if a page moves up or down?
When does the relocation happen?
16. PAGE RELOCATION
At the end of a monitor cycle the counters are recalculated
‒ Either IOPH (period) or weighted average (continuous)
Page counters with similar IOPH values are grouped together
IOPH groupings are ordered from highest to lowest
Tier capacity is overlaid on the IOPH groupings to decide on values for
tier ranges
‒ Tier range is the “break point” in IOPH between tiers
Relocation processes DP-VOLs page by page looking for pages on the
“wrong” side of a tier range value
‒ For example, high IOPH in a lower tier
‒ Relocation will perform a ZPR test on a page it moves
You can see the IOPH groupings and tier range values in SN2 “Pool Tier
Properties”
‒ Tier range stops being reported if any tier policy is specified
17. RELOCATION
Standard relocation throughput is about 3TB/day
Write pending and MP utilization rates influence the
pace of page relocation
‒ I/O priority is always given to the host(s)
Relocation statistics are logged
18. TIERING POLICIES
All
2-Tier
Pool
Any Tier
3-Tier
Pool
Any Tier
Level 1
Tier 1
Tier 1
Level 2
Tier 1 > 2
Tier 1 > 2
Level 3
Tier 2
Tier 2
Level 4
Tier 1 > 2
Tier 2 > 3
Level 5
Tier 2
Tier 3
Policy
Purpose
Most flexible
High response but sacrifice Tier 1
space efficiency
Similar to level 1 after level 1
relocates
Useful to reset tiering to a middle
state
Similar to level 3 after level 3
relocates
Useful if dormant volumes are known
Level1,
2-Tier All 2, 4
T1 > T2 > T3
T1 > T2 > T3
T1 > T2 > T3
T2 > T1 > T3
T2 > T3 > T1
T3 > T2 > T1
Level 5
Level 3
Level 1
3-Tier All
Level 4
Level 2
Tier1
Tier1
Tier2
Tier2
Level 3, 5
Default New
Page Assignment
Tier3
19. AVOIDING THRASHING
The bottom of the IOPH range for a tier is the “Tier Range” line
The top of the next tier is slightly higher than the bottom of the higher tier!
The overlap between tiers is called the “delta” and is used to help avoid
thrashing between the low end of 1 tier and the top of the next tier
Tier1
Tier2
To avoid pages “bouncing in
and out of a tier” the pages
in the “grey zone” are left
where they are, unless the
difference is 2 tiers
Tier3
Delta or grey zone
20. HDT USAGE CONSIDERATIONS
Application profiling is important (performance
requirements, sizing)
‒ Not all applications are appropriate for HDT. Sometimes HDP
will be more suitable
Consider
‒ 3TB/day is the average pace of relocation
Will relocations complete if the entire DB is active?
‒ Is disk sizing of pool appropriate?
If capacity is full on 1 tier type, the other tiers may take a
performance hit or page relocations may stop
Pace of relocation is dependent on array processor
utilization
22. HITACHI DYNAMIC TIERING: SUMMARY
Solution capabilities
Automated data placement for higher performance
and lower costs
Simplified ability to manage multiple storage tiers as
a single entity
Self-optimized for higher performance and space
efficiency
Page-based granular data movement for highest
efficiency and throughput
Storage Tiers
Data Heat Index
High
Activity
Set
Normal
Working
Set
Business value
Capex and opex savings by moving data to lowercost tiers
Increase storage utilization up to 50%
Quiet
Data Set
Easily align business application needs to the right
cost infrastructure
AUTOMATE AND ELIMINATE THE COMPLEXITIES OF EFFICIENT TIERED STORAGE
24. UPCOMING WEBTECHS
WebTechs
‒ Hitachi Dynamic Tiering: An In-Depth Look at Managing HDT and
Best Practices, Part 2, November 13, 9 a.m. PT, noon ET
‒ Best Practices for Virtualizing Exchange for Microsoft Private
Cloud, December 4, 9 a.m. PT, noon ET
Check www.hds.com/webtech for
Links to the recording, the presentation, and Q&A (available next
week)
Schedule and registration for upcoming WebTech sessions
Questions will be posted in the HDS Community:
http://community.hds.com/groups/webtech
Cycles – Note:24 hour auto has some start/stop time controlsManual can disconnect Monitor time from Relocation timefocus on last column
At the end of a monitor cycle the counters are recalculatedEither IOPH (Period) or weighted average (Continuous)Page counters with similar IOPH values are grouped togetherIOPH groupings are ordered from highest to lowestTier capacity is overlaid upon the IOPH groupings to decide on values for Tier Ranges Tier Range is the ‘break point’ in IOPH between tiersRelocation processes DP-VOLs page by page looking for pages on the ‘wrong’ side of a Tier Range valuei.e. high IOPH in a lower tierRelocation will perform a ZPR test on a page as it moves itYou can see the IOPH groupings and Tier Range values in SN2 “Pool Tier Properties”
This all leads up to relocation
The high boundary for the tier is 10% above the bottom of the prior tier….