SlideShare a Scribd company logo
1 of 26
Download to read offline
Deploying Ceph in the wild
Who am I? 
● Wido den Hollander (1986) 
● Co-owner and CTO of a PCextreme B.V., a 
dutch hosting company 
● Ceph trainer and consultant at 42on B.V. 
● Part of the Ceph community since late 2009 
– Wrote the Apache CloudStack integration 
– libvirt RBD storage pool support 
– PHP and Java bindings for librados
What is 42on? 
● Consultancy company focused on Ceph and 
it's eco-system 
● Founded in 2012 
● Based in the Netherlands 
● I'm the only employee 
– My consultancy company
Deploying Ceph 
● As a consultant I see a lot of different organizations 
– From small companies to large governments 
– I see Ceph being used in all kinds of deployments 
● It starts with gathering information about the use-case 
– Deployment application: RBD? Objects? 
– Storage requirements: TBs or PBs? 
– I/O requirements
I/O is EXPENSIVE 
● Everybody talks about storage capacity, 
almost nobody talks about Iops 
● Think about IOps first and then about 
TerraBytes 
Storage type € per I/O Remark 
HDD € 1,60 Seagate 3TB drive for €150 with 90 IOps 
SSD € 0,01 Intel S3500 480GB with 25k iops for €410
Design for I/O 
● Use more, but smaller disks 
– More spindles means more I/O 
– Can go for consumer drives, cheaper 
● Maybe deploy SSD-only 
– Intel S3500 or S3700 SSDs are reliable and fast 
● You really want I/O during recovery operations 
– OSDs replay PGLogs and scan directories 
– Recovery operations require a lot of I/O
Deployments 
● I've done numerous Ceph deployments 
– From tiny to large 
● Want to showcase two of the deployments 
– Use cases 
– Design principles
Ceph with CloudStack 
● Location: Belgium 
● Organization: Government 
● Use case: 
– RBD for CloudStack 
– S3 compatible storage 
● Requirements: 
– Storage for ~1000 Virtual Machines 
● Including PostgreSQL databases 
– TBs of S3 storage 
● Actual data is unknown to me
Ceph with CloudStack 
● Cluster: 
– 16 nodes with 24 drives 
● 19 1TB 7200RPM 2.5” 
● 2 Intel S3700 200GB SSDs for journaling 
● 2 Intel S3700 480GB SSDs for SSD-only storage 
● 64GB of memory 
● Xeon E5-2609 2.5Ghz CPU 
– 3x replication and 80% rounding provides: 
● 81TB HDD storage 
● 8TB SSD storage 
– 3 small nodes as monitors 
● SSD for Operating System and monitor data
Ceph with CloudStack
Ceph with CloudStack 
ROTATIONAL=$(cat /sys/block/$DEV/queue/rotational) 
if [ $ROTATIONAL -eq 1 ]; then 
echo "root=hdd rack=${RACK}-hdd host=${HOST}-hdd" 
else 
echo "root=ssd rack=${RACK}-ssd host=${HOST}-ssd" 
fi 
● If we detect the OSD is running on a SSD it goes into 
a different 'host' in the CRUSH Map 
– Rack is encoded in hostname (dc2-rk01) 
-48 2.88 rack rk01-ssd 
-33 0.72 host dc2-rk01-osd01-ssd 
252 0.36 osd.252 up 1 
253 0.36 osd.253 up 1 
-41 69.16 rack rk01-hdd 
-10 17.29 host dc2-rk01-osd01-hdd 
20 0.91 osd.20 up 1 
19 0.91 osd.19 up 1 
17 0.91 osd.17 up 1
Ceph with CloudStack 
● Download the script on my Github page: 
– Url: https://gist.github.com/wido 
– Place it in /usr/local/bin 
● Configure it in your ceph.conf 
– Push the config to your nodes using Puppet, Chef, 
Ansible, ceph-deploy, etc 
[osd] 
osd_crush_location_hook = /usr/local/bin/crush-location-looukp
Ceph with CloudStack 
● Highlights: 
– Automatic assignment of OSDs to right type 
– Designed for IOps. More, smaller drives 
● SSD for the real high I/O applications 
– RADOS Gateway for object storage 
● Trying to push developers towards objects instead of 
shared filesystems. A challenge! 
● Future: 
– Double cluster size within 6 months
RBD with OCFS2 
● Location: Netherlands 
● Organization: ISP 
● Use case: 
– RBD for OCFS2 
● Requirements: 
– Shared filesystem between webservers 
● Until CephFS is stable
RBD with OCFS2 
● Cluster: 
– 9 nodes with 8 drives 
● 1 SSD for Operating System 
● 7 Samsung 840 Pro 512GB SSDs 
● 10Gbit network (20Gbit LACP) 
– At 3x replication and 80% filling it provides 8.6TB 
of storage 
– 3 small nodes as monitors
RBD with OCFS2
RBD with OCFS2 
● “OCFS2 is a general-purpose shared-disk 
cluster file system for Linux capable of 
providing both high performance and high 
availability.” 
– RBD disks are shared 
– ext4 or XFS can't be mounted on multiple 
locations at the same time
RBD with OCFS2 
● All the challenges were in OCFS2, not in Ceph 
nor RBD 
– Running 3.14.17 kernel due to OCFS2 issues 
– Limited OCFS2 volumes to 200GB to minimize 
impact in case of volume corruption 
– Done multiple hardware upgrades without any 
service interruption 
● Runs smoothly while waiting for CephFS to 
mature
RBD with OCFS2 
● 10Gbit network for lower latency: 
– Lower network latency provides more performance 
– Lower latency means more IOps 
● Design for I/O! 
● 16k packet roundtrip times: 
– 1GbE: 0.8 ~ 1.1ms 
– 10GbE: 0.3 ~ 0.4ms 
● It's not about the bandwidth, it's about latency!
RBD with OCFS2 
● Highlights: 
– Full SSD cluster 
– 10GbE network for lower latency 
– Replaced all hardware since cluster was build 
● From 8 to 16 bays machines 
● Future: 
– Expand when required. No concrete planning
DO and DON'T 
● DO 
– Design for I/O, not raw TerraBytes 
– Think about network latency 
● 1GbE vs 10GbE 
– Use small(er) machines 
– Test recovery situations 
● Pull the plug out of those machines! 
– Reboot your machines regularly to verify it all works 
● So do update those machines! 
– Use dedicated hardware for your monitors 
● With a SSD for storage
DO and DON'T
DO and DON'T 
● DON'T 
– Create to many Placement Groups 
● It might overload your CPUs during recovery situations 
– Fill your cluster over 80% 
– Try to be smarter then Ceph 
● It's auto-healing. Give it some time. 
– Buy the most expensive machines 
● Better to have two cheap(er) ones 
– Use RAID-1 for journaling SSDs 
● Spread your OSDS over them
DO and DON'T
REMEMBER 
● Hardware failure is the rule, not the exception! 
● Consistency goes over availability 
● Ceph is designed to run on commodity 
hardware 
● There is no more need for RAID 
– forget it ever existed
Questions? 
● Twitter: @widodh 
● Skype: @widodh 
● E-Mail: wido@42on.com 
● Github: github.com/wido 
● Blog: http://blog.widodh.nl/

More Related Content

What's hot

Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Community
 
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsCeph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsColleen Corrice
 
Ceph for Big Science - Dan van der Ster
Ceph for Big Science - Dan van der SterCeph for Big Science - Dan van der Ster
Ceph for Big Science - Dan van der SterCeph Community
 
Red Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewRed Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewMarcel Hergaarden
 
Designing for High Performance Ceph at Scale
Designing for High Performance Ceph at ScaleDesigning for High Performance Ceph at Scale
Designing for High Performance Ceph at ScaleJames Saint-Rossy
 
SUSE Storage: Sizing and Performance (Ceph)
SUSE Storage: Sizing and Performance (Ceph)SUSE Storage: Sizing and Performance (Ceph)
SUSE Storage: Sizing and Performance (Ceph)Lars Marowsky-Brée
 
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing GuideCeph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing GuideKaran Singh
 
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Ceph Community
 
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLES
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLESQuick-and-Easy Deployment of a Ceph Storage Cluster with SLES
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLESJan Kalcic
 
Ceph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross TurkCeph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross Turkbuildacloud
 
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake SolutionCeph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake SolutionKaran Singh
 
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...Ceph Community
 
BlueStore: a new, faster storage backend for Ceph
BlueStore: a new, faster storage backend for CephBlueStore: a new, faster storage backend for Ceph
BlueStore: a new, faster storage backend for CephSage Weil
 
Tutorial ceph-2
Tutorial ceph-2Tutorial ceph-2
Tutorial ceph-2Tommy Lee
 
OpenStack and Ceph case study at the University of Alabama
OpenStack and Ceph case study at the University of AlabamaOpenStack and Ceph case study at the University of Alabama
OpenStack and Ceph case study at the University of AlabamaKamesh Pemmaraju
 
Ceph - A distributed storage system
Ceph - A distributed storage systemCeph - A distributed storage system
Ceph - A distributed storage systemItalo Santos
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storageAndrew Underwood
 
DataStax: Extreme Cassandra Optimization: The Sequel
DataStax: Extreme Cassandra Optimization: The SequelDataStax: Extreme Cassandra Optimization: The Sequel
DataStax: Extreme Cassandra Optimization: The SequelDataStax Academy
 
Journey to Stability: Petabyte Ceph Cluster in OpenStack Cloud
Journey to Stability: Petabyte Ceph Cluster in OpenStack CloudJourney to Stability: Petabyte Ceph Cluster in OpenStack Cloud
Journey to Stability: Petabyte Ceph Cluster in OpenStack CloudPatrick McGarry
 

What's hot (20)

Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons LearnedCeph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
Ceph Day Chicago - Ceph Deployment at Target: Best Practices and Lessons Learned
 
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsCeph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
 
Ceph for Big Science - Dan van der Ster
Ceph for Big Science - Dan van der SterCeph for Big Science - Dan van der Ster
Ceph for Big Science - Dan van der Ster
 
Red Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) OverviewRed Hat Storage 2014 - Product(s) Overview
Red Hat Storage 2014 - Product(s) Overview
 
Designing for High Performance Ceph at Scale
Designing for High Performance Ceph at ScaleDesigning for High Performance Ceph at Scale
Designing for High Performance Ceph at Scale
 
SUSE Storage: Sizing and Performance (Ceph)
SUSE Storage: Sizing and Performance (Ceph)SUSE Storage: Sizing and Performance (Ceph)
SUSE Storage: Sizing and Performance (Ceph)
 
Ceph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing GuideCeph Object Storage Reference Architecture Performance and Sizing Guide
Ceph Object Storage Reference Architecture Performance and Sizing Guide
 
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
Common Support Issues And How To Troubleshoot Them - Michael Hackett, Vikhyat...
 
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLES
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLESQuick-and-Easy Deployment of a Ceph Storage Cluster with SLES
Quick-and-Easy Deployment of a Ceph Storage Cluster with SLES
 
Ceph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross TurkCeph Intro and Architectural Overview by Ross Turk
Ceph Intro and Architectural Overview by Ross Turk
 
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake SolutionCeph Object Storage Performance Secrets and Ceph Data Lake Solution
Ceph Object Storage Performance Secrets and Ceph Data Lake Solution
 
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
CEPH DAY BERLIN - DISK HEALTH PREDICTION AND RESOURCE ALLOCATION FOR CEPH BY ...
 
BlueStore: a new, faster storage backend for Ceph
BlueStore: a new, faster storage backend for CephBlueStore: a new, faster storage backend for Ceph
BlueStore: a new, faster storage backend for Ceph
 
Tutorial ceph-2
Tutorial ceph-2Tutorial ceph-2
Tutorial ceph-2
 
Block Storage For VMs With Ceph
Block Storage For VMs With CephBlock Storage For VMs With Ceph
Block Storage For VMs With Ceph
 
OpenStack and Ceph case study at the University of Alabama
OpenStack and Ceph case study at the University of AlabamaOpenStack and Ceph case study at the University of Alabama
OpenStack and Ceph case study at the University of Alabama
 
Ceph - A distributed storage system
Ceph - A distributed storage systemCeph - A distributed storage system
Ceph - A distributed storage system
 
2015 open storage workshop ceph software defined storage
2015 open storage workshop   ceph software defined storage2015 open storage workshop   ceph software defined storage
2015 open storage workshop ceph software defined storage
 
DataStax: Extreme Cassandra Optimization: The Sequel
DataStax: Extreme Cassandra Optimization: The SequelDataStax: Extreme Cassandra Optimization: The Sequel
DataStax: Extreme Cassandra Optimization: The Sequel
 
Journey to Stability: Petabyte Ceph Cluster in OpenStack Cloud
Journey to Stability: Petabyte Ceph Cluster in OpenStack CloudJourney to Stability: Petabyte Ceph Cluster in OpenStack Cloud
Journey to Stability: Petabyte Ceph Cluster in OpenStack Cloud
 

Similar to Ceph Day London 2014 - Deploying ceph in the wild

Ceph Day Chicago - Ceph at work at Bloomberg
Ceph Day Chicago - Ceph at work at Bloomberg Ceph Day Chicago - Ceph at work at Bloomberg
Ceph Day Chicago - Ceph at work at Bloomberg Ceph Community
 
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageCeph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageSage Weil
 
Red hat open stack and storage presentation
Red hat open stack and storage presentationRed hat open stack and storage presentation
Red hat open stack and storage presentationMayur Shetty
 
20121102 ceph-in-the-cloud
20121102 ceph-in-the-cloud20121102 ceph-in-the-cloud
20121102 ceph-in-the-cloudCeph Community
 
Ceph as storage for CloudStack
Ceph as storage for CloudStack Ceph as storage for CloudStack
Ceph as storage for CloudStack Ceph Community
 
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...Ceph Community
 
Red Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed_Hat_Storage
 
Cephalocon apac china
Cephalocon apac chinaCephalocon apac china
Cephalocon apac chinaVikhyat Umrao
 
Deploying OpenStack with Ansible
Deploying OpenStack with AnsibleDeploying OpenStack with Ansible
Deploying OpenStack with AnsibleKevin Carter
 
Boosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringBoosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringShapeBlue
 
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farm
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farmKernel Recipes 2016 - Speeding up development by setting up a kernel build farm
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farmAnne Nicolas
 
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraBackup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraCeph Community
 
Your 1st Ceph cluster
Your 1st Ceph clusterYour 1st Ceph cluster
Your 1st Ceph clusterMirantis
 
Ceph Day Amsterdam 2015 - Ceph over IPv6
Ceph Day Amsterdam 2015 - Ceph over IPv6 Ceph Day Amsterdam 2015 - Ceph over IPv6
Ceph Day Amsterdam 2015 - Ceph over IPv6 Ceph Community
 
Lightweight Virtualization with Linux Containers and Docker | YaC 2013
Lightweight Virtualization with Linux Containers and Docker | YaC 2013Lightweight Virtualization with Linux Containers and Docker | YaC 2013
Lightweight Virtualization with Linux Containers and Docker | YaC 2013dotCloud
 
Lightweight Virtualization with Linux Containers and Docker I YaC 2013
Lightweight Virtualization with Linux Containers and Docker I YaC 2013Lightweight Virtualization with Linux Containers and Docker I YaC 2013
Lightweight Virtualization with Linux Containers and Docker I YaC 2013Docker, Inc.
 
Open Source Data Deduplication
Open Source Data DeduplicationOpen Source Data Deduplication
Open Source Data DeduplicationRedWireServices
 
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes][BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]Wong Hoi Sing Edison
 
Presentation for IGDCloud meetup: The clouds arena AWS ver. others
Presentation for IGDCloud meetup: The clouds arena AWS ver. othersPresentation for IGDCloud meetup: The clouds arena AWS ver. others
Presentation for IGDCloud meetup: The clouds arena AWS ver. othersForthscale
 

Similar to Ceph Day London 2014 - Deploying ceph in the wild (20)

Ceph Day Chicago - Ceph at work at Bloomberg
Ceph Day Chicago - Ceph at work at Bloomberg Ceph Day Chicago - Ceph at work at Bloomberg
Ceph Day Chicago - Ceph at work at Bloomberg
 
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud StorageCeph, Now and Later: Our Plan for Open Unified Cloud Storage
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
 
Red hat open stack and storage presentation
Red hat open stack and storage presentationRed hat open stack and storage presentation
Red hat open stack and storage presentation
 
20121102 ceph-in-the-cloud
20121102 ceph-in-the-cloud20121102 ceph-in-the-cloud
20121102 ceph-in-the-cloud
 
Ceph as storage for CloudStack
Ceph as storage for CloudStack Ceph as storage for CloudStack
Ceph as storage for CloudStack
 
Ceph barcelona-v-1.2
Ceph barcelona-v-1.2Ceph barcelona-v-1.2
Ceph barcelona-v-1.2
 
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...
 
Red Hat Gluster Storage Performance
Red Hat Gluster Storage PerformanceRed Hat Gluster Storage Performance
Red Hat Gluster Storage Performance
 
Cephalocon apac china
Cephalocon apac chinaCephalocon apac china
Cephalocon apac china
 
Deploying OpenStack with Ansible
Deploying OpenStack with AnsibleDeploying OpenStack with Ansible
Deploying OpenStack with Ansible
 
Boosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringBoosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uring
 
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farm
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farmKernel Recipes 2016 - Speeding up development by setting up a kernel build farm
Kernel Recipes 2016 - Speeding up development by setting up a kernel build farm
 
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraBackup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
 
Your 1st Ceph cluster
Your 1st Ceph clusterYour 1st Ceph cluster
Your 1st Ceph cluster
 
Ceph Day Amsterdam 2015 - Ceph over IPv6
Ceph Day Amsterdam 2015 - Ceph over IPv6 Ceph Day Amsterdam 2015 - Ceph over IPv6
Ceph Day Amsterdam 2015 - Ceph over IPv6
 
Lightweight Virtualization with Linux Containers and Docker | YaC 2013
Lightweight Virtualization with Linux Containers and Docker | YaC 2013Lightweight Virtualization with Linux Containers and Docker | YaC 2013
Lightweight Virtualization with Linux Containers and Docker | YaC 2013
 
Lightweight Virtualization with Linux Containers and Docker I YaC 2013
Lightweight Virtualization with Linux Containers and Docker I YaC 2013Lightweight Virtualization with Linux Containers and Docker I YaC 2013
Lightweight Virtualization with Linux Containers and Docker I YaC 2013
 
Open Source Data Deduplication
Open Source Data DeduplicationOpen Source Data Deduplication
Open Source Data Deduplication
 
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes][BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]
[BarCamp2018][20180915][Tips for Virtual Hosting on Kubernetes]
 
Presentation for IGDCloud meetup: The clouds arena AWS ver. others
Presentation for IGDCloud meetup: The clouds arena AWS ver. othersPresentation for IGDCloud meetup: The clouds arena AWS ver. others
Presentation for IGDCloud meetup: The clouds arena AWS ver. others
 

Recently uploaded

Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Igalia
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 

Recently uploaded (20)

Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 

Ceph Day London 2014 - Deploying ceph in the wild

  • 1. Deploying Ceph in the wild
  • 2. Who am I? ● Wido den Hollander (1986) ● Co-owner and CTO of a PCextreme B.V., a dutch hosting company ● Ceph trainer and consultant at 42on B.V. ● Part of the Ceph community since late 2009 – Wrote the Apache CloudStack integration – libvirt RBD storage pool support – PHP and Java bindings for librados
  • 3. What is 42on? ● Consultancy company focused on Ceph and it's eco-system ● Founded in 2012 ● Based in the Netherlands ● I'm the only employee – My consultancy company
  • 4. Deploying Ceph ● As a consultant I see a lot of different organizations – From small companies to large governments – I see Ceph being used in all kinds of deployments ● It starts with gathering information about the use-case – Deployment application: RBD? Objects? – Storage requirements: TBs or PBs? – I/O requirements
  • 5. I/O is EXPENSIVE ● Everybody talks about storage capacity, almost nobody talks about Iops ● Think about IOps first and then about TerraBytes Storage type € per I/O Remark HDD € 1,60 Seagate 3TB drive for €150 with 90 IOps SSD € 0,01 Intel S3500 480GB with 25k iops for €410
  • 6. Design for I/O ● Use more, but smaller disks – More spindles means more I/O – Can go for consumer drives, cheaper ● Maybe deploy SSD-only – Intel S3500 or S3700 SSDs are reliable and fast ● You really want I/O during recovery operations – OSDs replay PGLogs and scan directories – Recovery operations require a lot of I/O
  • 7. Deployments ● I've done numerous Ceph deployments – From tiny to large ● Want to showcase two of the deployments – Use cases – Design principles
  • 8. Ceph with CloudStack ● Location: Belgium ● Organization: Government ● Use case: – RBD for CloudStack – S3 compatible storage ● Requirements: – Storage for ~1000 Virtual Machines ● Including PostgreSQL databases – TBs of S3 storage ● Actual data is unknown to me
  • 9. Ceph with CloudStack ● Cluster: – 16 nodes with 24 drives ● 19 1TB 7200RPM 2.5” ● 2 Intel S3700 200GB SSDs for journaling ● 2 Intel S3700 480GB SSDs for SSD-only storage ● 64GB of memory ● Xeon E5-2609 2.5Ghz CPU – 3x replication and 80% rounding provides: ● 81TB HDD storage ● 8TB SSD storage – 3 small nodes as monitors ● SSD for Operating System and monitor data
  • 11. Ceph with CloudStack ROTATIONAL=$(cat /sys/block/$DEV/queue/rotational) if [ $ROTATIONAL -eq 1 ]; then echo "root=hdd rack=${RACK}-hdd host=${HOST}-hdd" else echo "root=ssd rack=${RACK}-ssd host=${HOST}-ssd" fi ● If we detect the OSD is running on a SSD it goes into a different 'host' in the CRUSH Map – Rack is encoded in hostname (dc2-rk01) -48 2.88 rack rk01-ssd -33 0.72 host dc2-rk01-osd01-ssd 252 0.36 osd.252 up 1 253 0.36 osd.253 up 1 -41 69.16 rack rk01-hdd -10 17.29 host dc2-rk01-osd01-hdd 20 0.91 osd.20 up 1 19 0.91 osd.19 up 1 17 0.91 osd.17 up 1
  • 12. Ceph with CloudStack ● Download the script on my Github page: – Url: https://gist.github.com/wido – Place it in /usr/local/bin ● Configure it in your ceph.conf – Push the config to your nodes using Puppet, Chef, Ansible, ceph-deploy, etc [osd] osd_crush_location_hook = /usr/local/bin/crush-location-looukp
  • 13. Ceph with CloudStack ● Highlights: – Automatic assignment of OSDs to right type – Designed for IOps. More, smaller drives ● SSD for the real high I/O applications – RADOS Gateway for object storage ● Trying to push developers towards objects instead of shared filesystems. A challenge! ● Future: – Double cluster size within 6 months
  • 14. RBD with OCFS2 ● Location: Netherlands ● Organization: ISP ● Use case: – RBD for OCFS2 ● Requirements: – Shared filesystem between webservers ● Until CephFS is stable
  • 15. RBD with OCFS2 ● Cluster: – 9 nodes with 8 drives ● 1 SSD for Operating System ● 7 Samsung 840 Pro 512GB SSDs ● 10Gbit network (20Gbit LACP) – At 3x replication and 80% filling it provides 8.6TB of storage – 3 small nodes as monitors
  • 17. RBD with OCFS2 ● “OCFS2 is a general-purpose shared-disk cluster file system for Linux capable of providing both high performance and high availability.” – RBD disks are shared – ext4 or XFS can't be mounted on multiple locations at the same time
  • 18. RBD with OCFS2 ● All the challenges were in OCFS2, not in Ceph nor RBD – Running 3.14.17 kernel due to OCFS2 issues – Limited OCFS2 volumes to 200GB to minimize impact in case of volume corruption – Done multiple hardware upgrades without any service interruption ● Runs smoothly while waiting for CephFS to mature
  • 19. RBD with OCFS2 ● 10Gbit network for lower latency: – Lower network latency provides more performance – Lower latency means more IOps ● Design for I/O! ● 16k packet roundtrip times: – 1GbE: 0.8 ~ 1.1ms – 10GbE: 0.3 ~ 0.4ms ● It's not about the bandwidth, it's about latency!
  • 20. RBD with OCFS2 ● Highlights: – Full SSD cluster – 10GbE network for lower latency – Replaced all hardware since cluster was build ● From 8 to 16 bays machines ● Future: – Expand when required. No concrete planning
  • 21. DO and DON'T ● DO – Design for I/O, not raw TerraBytes – Think about network latency ● 1GbE vs 10GbE – Use small(er) machines – Test recovery situations ● Pull the plug out of those machines! – Reboot your machines regularly to verify it all works ● So do update those machines! – Use dedicated hardware for your monitors ● With a SSD for storage
  • 23. DO and DON'T ● DON'T – Create to many Placement Groups ● It might overload your CPUs during recovery situations – Fill your cluster over 80% – Try to be smarter then Ceph ● It's auto-healing. Give it some time. – Buy the most expensive machines ● Better to have two cheap(er) ones – Use RAID-1 for journaling SSDs ● Spread your OSDS over them
  • 25. REMEMBER ● Hardware failure is the rule, not the exception! ● Consistency goes over availability ● Ceph is designed to run on commodity hardware ● There is no more need for RAID – forget it ever existed
  • 26. Questions? ● Twitter: @widodh ● Skype: @widodh ● E-Mail: wido@42on.com ● Github: github.com/wido ● Blog: http://blog.widodh.nl/