SlideShare una empresa de Scribd logo
1 de 45
When IaaS Meets DFS
IaaS平台中
儲存元件的考量與其需求
Huang Chih-Chieh (soem) @ NEET
Outline
• What is IaaS
• What is OpenStack
• Storage Types in IaaS
• Ceph
– Issues
• Summary
2
WHAT IS IAAS
3
Cloud Service Models Overview
• What if you want to have an IT department ?
– Similar to build a new house in previous analogy
• You can rent some virtualized infrastructure and build up your own
IT system among those resources, which may be fully controlled.
• Technical speaking, use the Infrastructure as a Service (IaaS)
solution.
– Similar to buy an empty house in previous analogy
• You can directly develop your IT system through one cloud
platform, and do not care about any lower level resource
management.
• Technical speaking, use the Platform as a Service (PaaS) solution.
– Similar to live in a hotel in previous analogy
• You can directly use some existed IT system solutions, which were
provided by some cloud application service provider, without
knowing any detail technique about how these service was
achieved.
• Technical speaking, use the Software as a Service (SaaS) solution.
4
From IaaS to PaaS
Traditional IT
Networking
Storage
Servers
Virtualization
OS
Middleware
Runtime
Data
Applications
YouManage
IaaS
Networking
Storage
Servers
Virtualization
OS
Middleware
Runtime
Data
Applications
YouManage
ProviderManage
PaaS
Networking
Storage
Servers
Virtualization
OS
Middleware
Runtime
Data
Applications
YouManage
ProviderManage
5
Service Model Overview
6
WHAT IS OPENSTACK
7
8
9
OpenStack Storage
10
cinder-volume
STORAGE TYPES IN IAAS
11
OpenStack Storage
• Instance Storage Provider
– Off Compute Node Storage
• Shared File System
– On Compute Node Storage
• Shared File System
– On Compute Node Storage
• Non-shared File System
• Image Repository
– Glance
12
OpenNebula Storage
13
• Properties
– File System vs. Block Device
– Shared vs. Non-shared
• Four types:
– Shared File System
– Non-Shared File System
– Shared Block Device
– Non-Shared Block Device
SSCloud Storage
14
• File System
– Bootable image / Small image
• Small image
– FS cache in host memory
– Random accessing
– Type
• Shared
– DFS (ceph), NFS (nfsd)
• Non-Shared
– Local filesystem + scp
SSCloud Storage
15
• Block device (with LVM)
– Additional space / large image
– Heavy accessing image
• Large image
– No FS cache in host memory => save memory
– Large chunk access
» Hadoop (64MB~128MB per file)
– Type
• Shared
– iSCSI + LVM
• Non-Shared
– LVM
SSCloud Storage
16
Storage Systems
• File Based
– NFS
– DFS
• Lustre
• GlusterFS
• MooseFS
• Ceph
17
Storage Systems
• Block Based
– iSCSI + LVM
– DRBD
– VastSky
– KPS: Kernel-based Programmable Storage System
– Ceph
18
Storage Systems
• Object Based
– Openstack Swift
– Hadoop HDFS
• with WebHDFS (1.0.4-stable) or HTTPFS (2.0.3-alpha)
– Ceph
19
CEPH
CEPH: THE FUTURE OF STORAGE™
20
Ceph
• Overview
– Ceph is a free software distributed file system.
– Ceph's main goals are to be POSIX-compatible, and
completely distributed without a single point of failure.
– The data is seamlessly replicated, making it fault
tolerant.
• Release
– On July 3, 2012, the Ceph development team released
Argonaut, the first release of Ceph with long-term
support.
21
Ceph
• Introduction
– Ceph is a distributed file system that provides
excellent performance ,reliability and scalability.
– Objected-based Storage.
– Ceph separates data and metadata operations by
eliminating file allocation tables and replacing them
with generating functions.
– Ceph utilizes a highly adaptive distributed metadata
cluster, improving scalability.
– Using OSD to directly access data, high performance.
22
Ceph
• Objected-based Storage
23
Ceph
• Goal
– Scalability
• Storage capacity, throughput, client performance.
Emphasis on HPC.
– Reliability
• Failures are the norm rather than the exception, so the
system must have fault detection and recovery
mechanism.
– Performance
• Dynamic workloads  Load balance.
24
Ceph
• Ceph Filesystem
– POSIX
• File based
• Ceph Block Device
– RBD
• Block based
• Ceph Object Gateway
– Swift / S3 Restful API
• Object based
25
Ceph
• Three main components
– Clients : Near-POSIX file system interface.
– Cluster of OSDs : Store all data and metadata.
– Metadata Cluster : Manage namespace(file name)
26
Three Fundamental Design
1. Separating Data and Metadata
– Separation of file metadata management from the
storage.
– Metadata operations are collectively managed by
a metadata server cluster.
– User can direct access OSDs to get data by
metadata.
– Ceph remove data allocation lists entirely.
– Using CRUSH assigns objects to storage devices.
27
Separating Data and Metadata
• Ceph separates data and metadata operations
28
Separating Data and Metadata
• Data Distribution with CRUSH
– In order to avoid imbalance(OSD idle, empty) or
load asymmetries(hot data on new device).
→distributing new data randomly.
– Ceph maps ojects into Placement groups(PGs)PGs
are assigned to OSDs by CRUSH.
29
Dynamic Distributed Metadata
Management
2. Dynamic Distributed Metadata Management
 Ceph utilizes a metadata cluster architecture based on Dynamic
Subtree Partitioning.(workload balance)
– Dynamic Subtree Partitioning
• Most FS, use static subtree partitioning
→imbalance workloads and easy hash function.
• Ceph’s MDS cluster is based on a dynamic subtree
partitioning. →balance workloads
30
Reliable Distributed Object Storage
3. Reliable Autonomic Distributed Object
Storage
– Replica.
– Failure Detection and Recovery.
31
Client
• Client Operation
– File I/O and Capabilities
Request
Client
(open file)
MDS
Translate file
name into
inode(inode
number, file
owner, mode,
size, …)Check OK, return
Return inode number,
map file data into
objects(CRUSH)
OSD
Direct
access
32
Client
• Client Synchronization
– If Multiple clients(readers and writers) use same
file, cancel any previously read and write
capability until OSD check OK.
• Traditional: Update serialization. →Bad performance
• Ceph: Use HPC(high-performance computing
community) can read and write different parts of same
file(diff bojects).
→increase performance
33
Metadata
• Dynamically Distributed Metadata
– MDSs use journaling
• Repetitive metadata updates handled in memory.
• Optimizes on-disk layout for read access.
– Per-MDS has journal, when MDS failure another
node can quickly recover with journal.
– Inodes are embedded directly within directories.
– Each directory’s content is written to the OSD
cluster using the same striping and distribution
strategy as metadata journals and file data.
34
Replica
• Replication
– Data is replicated in terms of PGs.
– Clients send all writes to the first non-failed OSD
in an object’s PG (the primary), which assigns a
new version number for the object and PG and
forwards the write to any additional replica OSDs.
35
Failure detection
• Failure detection
– When OSD not response → sign “down”
– Pass to the next OSD.
– If first OSD doesn’t recover →sign “out”
– Another OSD join.
36
Failure Recovery
• Recovery and Cluster Updates
– If an OSD1 crashes → sign “down”
– The OSD2 take over as primary.
– If OSD1 recovers → sign “up”
– The OSD2 receives update request, sent new
version data to OSD1.
37
EVERYTHING LOOKS GOOD, BUT…
38
Issues
• Highly developed
– 0.48
• Monitor waste CPUs
• Recovery into un-consistency state
– 0.56
• Bugs in file extend behavior
– Qcow2 images have got IO errors in VMs kernel,
» but things are going well in the log of Ceph.
– 0.67
• ceph-deploy
39
Issues
• Correct the time
– 0.56
• OSDs waste CPUs
– ntpdate tock.stdtime.gov.tw
– 0.67
• health HEALTH_WARN clock skew detected on mon.1
– ntpdate tock.stdtime.gov.tw
– ntpserver
40
Issues
• CephFS is not statble
– Newly system can use ceph RBD
– Traditional system could only use the POSIX
interface
• 0.56
– Ceph’s operation in a folder would be frozen,
» if that folder is getting heavy loading.
– Bugs in file extend behavior
REF: http://www.sebastien-han.fr/blog/2013/06/24/what-i-think-about-cephfs-in-openstack/
41
Issues
• Mount ceph with
– Kernel module
• mount –t ceph …
– FUSE
• ceph-fuse -c /etc/ceph/ceph.conf …
42
Issues
43
root@SSCloud-01:/# cephfs /mnt/dev set_layout -p 5
Segmentation fault
cephfs is not a super-friendly tool right now — sorry! :(
I believe you will find it works correctly if you specify all the layout parameters,
not just one of them.
root@SSCloud-01:/# cephfs -h
not enough parameters!
usage: cephfs path command [options]*
Commands:
show_layout -- view the layout information on a file or dir
set_layout -- set the layout on an empty file, or the default layout on a directory
show_location -- view the location information on a file
map -- display file objects, pgs, osds
Options:
Useful for setting layouts:
--stripe_unit, -u: set the size of each stripe
--stripe_count, -c: set the number of objects to stripe across
--object_size, -s: set the size of the objects to stripe across
--pool, -p: set the pool to use
Useful for getting location data:
--offset, -l: the offset to retrieve location data for
root@SSCloud-01:/# cephfs /mnt/dev set_layout -u 4194304 -c 1 -s 4194304 -p 5
root@SSCloud-01:/# cephfs /mnt/dev show_layout
layout.data_pool: 5
layout.object_size: 4194304
layout.stripe_unit: 4194304
layout.stripe_count: 1
SUMMARY
44
Summary
• There are three type of storage in IaaS
– File-based, block-based, object-based
• Ceph is a good choice for IaaS
– OpenStack store images in Ceph Block Device
– Cinder or nova-volume to boot a VM
• using a copy-on-write clone of an image
• CephFS is still highly developed
– However, newer version is better.
45

Más contenido relacionado

La actualidad más candente

Hadoop distributed file system
Hadoop distributed file systemHadoop distributed file system
Hadoop distributed file systemAnshul Bhatnagar
 
Hadoop HDFS by rohitkapa
Hadoop HDFS by rohitkapaHadoop HDFS by rohitkapa
Hadoop HDFS by rohitkapakapa rohit
 
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014 WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014 Chris Almond
 
NoSQL overview implementation free
NoSQL overview implementation freeNoSQL overview implementation free
NoSQL overview implementation freeBenoit Perroud
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File SystemVaibhav Jain
 
Hadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldHadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldDataWorks Summit
 
Solving Hadoop Replication Challenges with an Active-Active Paxos Algorithm
Solving Hadoop Replication Challenges with an Active-Active Paxos AlgorithmSolving Hadoop Replication Challenges with an Active-Active Paxos Algorithm
Solving Hadoop Replication Challenges with an Active-Active Paxos AlgorithmDataWorks Summit
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File Systemelliando dias
 
Distributed Computing with Apache Hadoop: Technology Overview
Distributed Computing with Apache Hadoop: Technology OverviewDistributed Computing with Apache Hadoop: Technology Overview
Distributed Computing with Apache Hadoop: Technology OverviewKonstantin V. Shvachko
 
Spectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN CachingSpectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN CachingSandeep Patil
 
Optimizing your Infrastrucure and Operating System for Hadoop
Optimizing your Infrastrucure and Operating System for HadoopOptimizing your Infrastrucure and Operating System for Hadoop
Optimizing your Infrastrucure and Operating System for HadoopDataWorks Summit
 
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...inside-BigData.com
 
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use Case
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use CaseOracle Solaris 11 as a BIG Data Platform Apache Hadoop Use Case
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use CaseOrgad Kimchi
 
02.28.13 WANdisco ApacheCon 2013
02.28.13 WANdisco ApacheCon 201302.28.13 WANdisco ApacheCon 2013
02.28.13 WANdisco ApacheCon 2013WANdisco Plc
 
Apache Hadoop YARN, NameNode HA, HDFS Federation
Apache Hadoop YARN, NameNode HA, HDFS FederationApache Hadoop YARN, NameNode HA, HDFS Federation
Apache Hadoop YARN, NameNode HA, HDFS FederationAdam Kawa
 
Ozone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityOzone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityDinesh Chitlangia
 
Architecture of a Next-Generation Parallel File System
Architecture of a Next-Generation Parallel File System	Architecture of a Next-Generation Parallel File System
Architecture of a Next-Generation Parallel File System Great Wide Open
 

La actualidad más candente (20)

Hadoop HDFS
Hadoop HDFSHadoop HDFS
Hadoop HDFS
 
Hadoop distributed file system
Hadoop distributed file systemHadoop distributed file system
Hadoop distributed file system
 
Big data- HDFS(2nd presentation)
Big data- HDFS(2nd presentation)Big data- HDFS(2nd presentation)
Big data- HDFS(2nd presentation)
 
Hadoop HDFS by rohitkapa
Hadoop HDFS by rohitkapaHadoop HDFS by rohitkapa
Hadoop HDFS by rohitkapa
 
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014 WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014
WANdisco Non-Stop Hadoop: PHXDataConference Presentation Oct 2014
 
NoSQL overview implementation free
NoSQL overview implementation freeNoSQL overview implementation free
NoSQL overview implementation free
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
 
Hadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldHadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the Field
 
Solving Hadoop Replication Challenges with an Active-Active Paxos Algorithm
Solving Hadoop Replication Challenges with an Active-Active Paxos AlgorithmSolving Hadoop Replication Challenges with an Active-Active Paxos Algorithm
Solving Hadoop Replication Challenges with an Active-Active Paxos Algorithm
 
Hadoop Distributed File System
Hadoop Distributed File SystemHadoop Distributed File System
Hadoop Distributed File System
 
Distributed Computing with Apache Hadoop: Technology Overview
Distributed Computing with Apache Hadoop: Technology OverviewDistributed Computing with Apache Hadoop: Technology Overview
Distributed Computing with Apache Hadoop: Technology Overview
 
Spectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN CachingSpectrum Scale Unified File and Object with WAN Caching
Spectrum Scale Unified File and Object with WAN Caching
 
Optimizing your Infrastrucure and Operating System for Hadoop
Optimizing your Infrastrucure and Operating System for HadoopOptimizing your Infrastrucure and Operating System for Hadoop
Optimizing your Infrastrucure and Operating System for Hadoop
 
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
Big Data Meets HPC - Exploiting HPC Technologies for Accelerating Big Data Pr...
 
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use Case
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use CaseOracle Solaris 11 as a BIG Data Platform Apache Hadoop Use Case
Oracle Solaris 11 as a BIG Data Platform Apache Hadoop Use Case
 
02.28.13 WANdisco ApacheCon 2013
02.28.13 WANdisco ApacheCon 201302.28.13 WANdisco ApacheCon 2013
02.28.13 WANdisco ApacheCon 2013
 
Hdfs architecture
Hdfs architectureHdfs architecture
Hdfs architecture
 
Apache Hadoop YARN, NameNode HA, HDFS Federation
Apache Hadoop YARN, NameNode HA, HDFS FederationApache Hadoop YARN, NameNode HA, HDFS Federation
Apache Hadoop YARN, NameNode HA, HDFS Federation
 
Ozone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityOzone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalability
 
Architecture of a Next-Generation Parallel File System
Architecture of a Next-Generation Parallel File System	Architecture of a Next-Generation Parallel File System
Architecture of a Next-Generation Parallel File System
 

Similar a Dfs in iaa_s

What's new in hadoop 3.0
What's new in hadoop 3.0What's new in hadoop 3.0
What's new in hadoop 3.0Heiko Loewe
 
HDFS_architecture.ppt
HDFS_architecture.pptHDFS_architecture.ppt
HDFS_architecture.pptvijayapraba1
 
Storage solutions for High Performance Computing
Storage solutions for High Performance ComputingStorage solutions for High Performance Computing
Storage solutions for High Performance Computinggmateesc
 
hadoop distributed file systems complete information
hadoop distributed file systems complete informationhadoop distributed file systems complete information
hadoop distributed file systems complete informationbhargavi804095
 
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.MaharajothiP
 
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and Hadoop
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and HadoopIOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and Hadoop
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and HadoopLeons Petražickis
 
Hadoop Distributed file system.pdf
Hadoop Distributed file system.pdfHadoop Distributed file system.pdf
Hadoop Distributed file system.pdfvishal choudhary
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storageAndrew Underwood
 
Quick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterQuick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterPatrick Quairoli
 
Gpfs introandsetup
Gpfs introandsetupGpfs introandsetup
Gpfs introandsetupasihan
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCeph Community
 
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMF
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMFGestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMF
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMFSUSE Italy
 
Rapid Cluster Computing with Apache Spark 2016
Rapid Cluster Computing with Apache Spark 2016Rapid Cluster Computing with Apache Spark 2016
Rapid Cluster Computing with Apache Spark 2016Zohar Elkayam
 
Topic 9a-Hadoop Storage- HDFS.pptx
Topic 9a-Hadoop Storage- HDFS.pptxTopic 9a-Hadoop Storage- HDFS.pptx
Topic 9a-Hadoop Storage- HDFS.pptxDanishMahmood23
 

Similar a Dfs in iaa_s (20)

What's new in hadoop 3.0
What's new in hadoop 3.0What's new in hadoop 3.0
What's new in hadoop 3.0
 
HDFS_architecture.ppt
HDFS_architecture.pptHDFS_architecture.ppt
HDFS_architecture.ppt
 
Chapter2.pdf
Chapter2.pdfChapter2.pdf
Chapter2.pdf
 
Storage solutions for High Performance Computing
Storage solutions for High Performance ComputingStorage solutions for High Performance Computing
Storage solutions for High Performance Computing
 
hadoop distributed file systems complete information
hadoop distributed file systems complete informationhadoop distributed file systems complete information
hadoop distributed file systems complete information
 
11. dfs
11. dfs11. dfs
11. dfs
 
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.
P.Maharajothi,II-M.sc(computer science),Bon secours college for women,thanjavur.
 
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and Hadoop
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and HadoopIOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and Hadoop
IOD 2013 - Crunch Big Data in the Cloud with IBM BigInsights and Hadoop
 
DAS RAID NAS SAN
DAS RAID NAS SANDAS RAID NAS SAN
DAS RAID NAS SAN
 
Hadoop data management
Hadoop data managementHadoop data management
Hadoop data management
 
HADOOP.pptx
HADOOP.pptxHADOOP.pptx
HADOOP.pptx
 
Hadoop Distributed file system.pdf
Hadoop Distributed file system.pdfHadoop Distributed file system.pdf
Hadoop Distributed file system.pdf
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storage
 
Quick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage ClusterQuick-and-Easy Deployment of a Ceph Storage Cluster
Quick-and-Easy Deployment of a Ceph Storage Cluster
 
Gpfs introandsetup
Gpfs introandsetupGpfs introandsetup
Gpfs introandsetup
 
Kfs presentation
Kfs presentationKfs presentation
Kfs presentation
 
CephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at LastCephFS in Jewel: Stable at Last
CephFS in Jewel: Stable at Last
 
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMF
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMFGestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMF
Gestione gerarchica dei dati con SUSE Enterprise Storage e HPE DMF
 
Rapid Cluster Computing with Apache Spark 2016
Rapid Cluster Computing with Apache Spark 2016Rapid Cluster Computing with Apache Spark 2016
Rapid Cluster Computing with Apache Spark 2016
 
Topic 9a-Hadoop Storage- HDFS.pptx
Topic 9a-Hadoop Storage- HDFS.pptxTopic 9a-Hadoop Storage- HDFS.pptx
Topic 9a-Hadoop Storage- HDFS.pptx
 

Último

Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxBkGupta21
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 

Último (20)

Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptx
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 

Dfs in iaa_s

  • 1. When IaaS Meets DFS IaaS平台中 儲存元件的考量與其需求 Huang Chih-Chieh (soem) @ NEET
  • 2. Outline • What is IaaS • What is OpenStack • Storage Types in IaaS • Ceph – Issues • Summary 2
  • 4. Cloud Service Models Overview • What if you want to have an IT department ? – Similar to build a new house in previous analogy • You can rent some virtualized infrastructure and build up your own IT system among those resources, which may be fully controlled. • Technical speaking, use the Infrastructure as a Service (IaaS) solution. – Similar to buy an empty house in previous analogy • You can directly develop your IT system through one cloud platform, and do not care about any lower level resource management. • Technical speaking, use the Platform as a Service (PaaS) solution. – Similar to live in a hotel in previous analogy • You can directly use some existed IT system solutions, which were provided by some cloud application service provider, without knowing any detail technique about how these service was achieved. • Technical speaking, use the Software as a Service (SaaS) solution. 4
  • 5. From IaaS to PaaS Traditional IT Networking Storage Servers Virtualization OS Middleware Runtime Data Applications YouManage IaaS Networking Storage Servers Virtualization OS Middleware Runtime Data Applications YouManage ProviderManage PaaS Networking Storage Servers Virtualization OS Middleware Runtime Data Applications YouManage ProviderManage 5
  • 8. 8
  • 9. 9
  • 11. STORAGE TYPES IN IAAS 11
  • 12. OpenStack Storage • Instance Storage Provider – Off Compute Node Storage • Shared File System – On Compute Node Storage • Shared File System – On Compute Node Storage • Non-shared File System • Image Repository – Glance 12
  • 14. • Properties – File System vs. Block Device – Shared vs. Non-shared • Four types: – Shared File System – Non-Shared File System – Shared Block Device – Non-Shared Block Device SSCloud Storage 14
  • 15. • File System – Bootable image / Small image • Small image – FS cache in host memory – Random accessing – Type • Shared – DFS (ceph), NFS (nfsd) • Non-Shared – Local filesystem + scp SSCloud Storage 15
  • 16. • Block device (with LVM) – Additional space / large image – Heavy accessing image • Large image – No FS cache in host memory => save memory – Large chunk access » Hadoop (64MB~128MB per file) – Type • Shared – iSCSI + LVM • Non-Shared – LVM SSCloud Storage 16
  • 17. Storage Systems • File Based – NFS – DFS • Lustre • GlusterFS • MooseFS • Ceph 17
  • 18. Storage Systems • Block Based – iSCSI + LVM – DRBD – VastSky – KPS: Kernel-based Programmable Storage System – Ceph 18
  • 19. Storage Systems • Object Based – Openstack Swift – Hadoop HDFS • with WebHDFS (1.0.4-stable) or HTTPFS (2.0.3-alpha) – Ceph 19
  • 20. CEPH CEPH: THE FUTURE OF STORAGE™ 20
  • 21. Ceph • Overview – Ceph is a free software distributed file system. – Ceph's main goals are to be POSIX-compatible, and completely distributed without a single point of failure. – The data is seamlessly replicated, making it fault tolerant. • Release – On July 3, 2012, the Ceph development team released Argonaut, the first release of Ceph with long-term support. 21
  • 22. Ceph • Introduction – Ceph is a distributed file system that provides excellent performance ,reliability and scalability. – Objected-based Storage. – Ceph separates data and metadata operations by eliminating file allocation tables and replacing them with generating functions. – Ceph utilizes a highly adaptive distributed metadata cluster, improving scalability. – Using OSD to directly access data, high performance. 22
  • 24. Ceph • Goal – Scalability • Storage capacity, throughput, client performance. Emphasis on HPC. – Reliability • Failures are the norm rather than the exception, so the system must have fault detection and recovery mechanism. – Performance • Dynamic workloads  Load balance. 24
  • 25. Ceph • Ceph Filesystem – POSIX • File based • Ceph Block Device – RBD • Block based • Ceph Object Gateway – Swift / S3 Restful API • Object based 25
  • 26. Ceph • Three main components – Clients : Near-POSIX file system interface. – Cluster of OSDs : Store all data and metadata. – Metadata Cluster : Manage namespace(file name) 26
  • 27. Three Fundamental Design 1. Separating Data and Metadata – Separation of file metadata management from the storage. – Metadata operations are collectively managed by a metadata server cluster. – User can direct access OSDs to get data by metadata. – Ceph remove data allocation lists entirely. – Using CRUSH assigns objects to storage devices. 27
  • 28. Separating Data and Metadata • Ceph separates data and metadata operations 28
  • 29. Separating Data and Metadata • Data Distribution with CRUSH – In order to avoid imbalance(OSD idle, empty) or load asymmetries(hot data on new device). →distributing new data randomly. – Ceph maps ojects into Placement groups(PGs)PGs are assigned to OSDs by CRUSH. 29
  • 30. Dynamic Distributed Metadata Management 2. Dynamic Distributed Metadata Management  Ceph utilizes a metadata cluster architecture based on Dynamic Subtree Partitioning.(workload balance) – Dynamic Subtree Partitioning • Most FS, use static subtree partitioning →imbalance workloads and easy hash function. • Ceph’s MDS cluster is based on a dynamic subtree partitioning. →balance workloads 30
  • 31. Reliable Distributed Object Storage 3. Reliable Autonomic Distributed Object Storage – Replica. – Failure Detection and Recovery. 31
  • 32. Client • Client Operation – File I/O and Capabilities Request Client (open file) MDS Translate file name into inode(inode number, file owner, mode, size, …)Check OK, return Return inode number, map file data into objects(CRUSH) OSD Direct access 32
  • 33. Client • Client Synchronization – If Multiple clients(readers and writers) use same file, cancel any previously read and write capability until OSD check OK. • Traditional: Update serialization. →Bad performance • Ceph: Use HPC(high-performance computing community) can read and write different parts of same file(diff bojects). →increase performance 33
  • 34. Metadata • Dynamically Distributed Metadata – MDSs use journaling • Repetitive metadata updates handled in memory. • Optimizes on-disk layout for read access. – Per-MDS has journal, when MDS failure another node can quickly recover with journal. – Inodes are embedded directly within directories. – Each directory’s content is written to the OSD cluster using the same striping and distribution strategy as metadata journals and file data. 34
  • 35. Replica • Replication – Data is replicated in terms of PGs. – Clients send all writes to the first non-failed OSD in an object’s PG (the primary), which assigns a new version number for the object and PG and forwards the write to any additional replica OSDs. 35
  • 36. Failure detection • Failure detection – When OSD not response → sign “down” – Pass to the next OSD. – If first OSD doesn’t recover →sign “out” – Another OSD join. 36
  • 37. Failure Recovery • Recovery and Cluster Updates – If an OSD1 crashes → sign “down” – The OSD2 take over as primary. – If OSD1 recovers → sign “up” – The OSD2 receives update request, sent new version data to OSD1. 37
  • 39. Issues • Highly developed – 0.48 • Monitor waste CPUs • Recovery into un-consistency state – 0.56 • Bugs in file extend behavior – Qcow2 images have got IO errors in VMs kernel, » but things are going well in the log of Ceph. – 0.67 • ceph-deploy 39
  • 40. Issues • Correct the time – 0.56 • OSDs waste CPUs – ntpdate tock.stdtime.gov.tw – 0.67 • health HEALTH_WARN clock skew detected on mon.1 – ntpdate tock.stdtime.gov.tw – ntpserver 40
  • 41. Issues • CephFS is not statble – Newly system can use ceph RBD – Traditional system could only use the POSIX interface • 0.56 – Ceph’s operation in a folder would be frozen, » if that folder is getting heavy loading. – Bugs in file extend behavior REF: http://www.sebastien-han.fr/blog/2013/06/24/what-i-think-about-cephfs-in-openstack/ 41
  • 42. Issues • Mount ceph with – Kernel module • mount –t ceph … – FUSE • ceph-fuse -c /etc/ceph/ceph.conf … 42
  • 43. Issues 43 root@SSCloud-01:/# cephfs /mnt/dev set_layout -p 5 Segmentation fault cephfs is not a super-friendly tool right now — sorry! :( I believe you will find it works correctly if you specify all the layout parameters, not just one of them. root@SSCloud-01:/# cephfs -h not enough parameters! usage: cephfs path command [options]* Commands: show_layout -- view the layout information on a file or dir set_layout -- set the layout on an empty file, or the default layout on a directory show_location -- view the location information on a file map -- display file objects, pgs, osds Options: Useful for setting layouts: --stripe_unit, -u: set the size of each stripe --stripe_count, -c: set the number of objects to stripe across --object_size, -s: set the size of the objects to stripe across --pool, -p: set the pool to use Useful for getting location data: --offset, -l: the offset to retrieve location data for root@SSCloud-01:/# cephfs /mnt/dev set_layout -u 4194304 -c 1 -s 4194304 -p 5 root@SSCloud-01:/# cephfs /mnt/dev show_layout layout.data_pool: 5 layout.object_size: 4194304 layout.stripe_unit: 4194304 layout.stripe_count: 1
  • 45. Summary • There are three type of storage in IaaS – File-based, block-based, object-based • Ceph is a good choice for IaaS – OpenStack store images in Ceph Block Device – Cinder or nova-volume to boot a VM • using a copy-on-write clone of an image • CephFS is still highly developed – However, newer version is better. 45