SlideShare una empresa de Scribd logo
1 de 72
Descargar para leer sin conexión
Page 1 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Operational Best Practices Workshop
We Do Hadoop
Sean Roberts
Partner Solutions Engineer - EMEA
Page 2 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
$(whoami)
Sean Roberts
Partner Solutions Engineer
London, EMEA & everywhere
@seano
Page 3 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Operations
● Plan
● Provision & Deploy
● Secure
● Manage
● Workshop: Multi-Tenancy
Page 4 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Page 5 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
HDP - Runs Everywhere
On-Premises Cloud
Virtualized, Container, ...
Your deployment of Hadoop
• VMWare, HyperV, XEN, ...
• OpenStack
• Docker
HDP on Your Hardware
• Linux
• Windows
HDP on Appliance
Turnkey Hadoop Appliances
• Teradata
• Microsoft
• PSSC Labs
Infrastructure as a Service
(IAAS)
- Microsoft Azure - Rackspace
- Amazon EC2 - Google Compute
Hadoop as a Service (HAAS)
• Microsoft HDInsight
• Rackspace
Page 6 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Plan - Cluster Design
Design choices, component distribution, and an example
Page 7 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
More is better
● More is better than bigger when it comes to nodes
○ Faster healing
■ Nodes with larger storage take longer to recover
○ More nodes = more resilience, parallelism, power!
○ More racks = ...
Page 8 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Component Layout
All recommendations should be weighed against use cases
Master components
● Distribute across racks to spread risk
● As cluster grows, distribution will change
Worker components
● Typically identical across all worker nodes
Page 9 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Small Cluster without HA
Page 10 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Small Cluster with HA
Secondary
Page 11 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Component Layout
Multiple gateway nodes (load balanced)
● HA/distribute client services: SSH, Knox, Ambari View instances, data loaders,
● Host other services: ntp, package repositories, Ambari View servers
5 ZooKeepers for greater reliability (default is 3)
● Also eases taking them down for maintenance
● >5 will be slower due to more voters in elections
● Consider fast disks (SAS, SSD)
Some like Virtualized Masters
● Master services are not typically IO heavy / SAN compatible
● Eases hardware replacement (live vm migration)
Page 12 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Layout of External Databases
Several components require external databases:
● Ambari: PostgreSQL
● Hive: MySQL
● Oozie: Derby
● Ranger: MySQL
Supported: Oracle, MySQL, PostgreSQL
● Consider using same technology for ease of management
● Use the same servers where possible
● Let your DBAs manage so backups, HA, … are taken care of
Heavy usage of Falcon+Oozie or Ambari, may require dedicated
instances
Page 13 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Development cluster
Your development processes still apply
Ideally have a separate cluster for dev/test
● Many have dev/test combined
● Cloud: for scale up/down depending on need
● Often the pilot cluster becomes the dev cluster
Smaller but same configuration
● e.g. If prod has Kerberos, HA, etc., then dev should too
Page 14 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Plan - Hardware Selection
find your sweet spot
Page 15 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Sizing: Typical Worker
CPU
● Dual socket / 8-12 core each
RAM
● Typical: 128GB
● Not uncommon: 256GB
HDFS Storage
● 8-12 x 2-3TB (NL-SAS/SATA)
● 1TB for performance focus
● 4TB+ for storage archive focus
Page 16 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Hardware sizing
What is your workload?
● Balanced: most common
● Memory: such as Spark
● Compute: such as Storm
● Storage: such as archive
Mixed hardware in a cluster
● Ambari Configuration groups
● YARN Labels: Pin processing to subset of hardware
● Heterogeneous Storage
Page 17 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Sizing: Storm & Kafka
Storm & Kafka
● Deploy together
○ Storm is compute bound
○ Kafka is disk bound
● Same hardware as typical worker
● RAM: 128GB minimum
● Disks: 4-6 disks enough for most Kafka workloads
○ Assuming log retention of 2-3 days
● Nimbus should be on a master with HA hardware
Page 18 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Storage Configuration
Master nodes
● RAID-10
● O/S + Data disks
Data nodes
● Single O/S disk or RAID-1
○ Consider RAID-1 in small clusters
where losing a single node could
be a problem
Data node HDFS
● RAID-0 per disk or JBOD
○ e.g.12 disks = 12 x 1 disk arrays
○ Ensure they are mounted
separately!
● Data nodes survive disk loss:
○ `dfs.datanode.failed.volumes.tolerated`
Page 19 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Storage Types
Before
• DataNode is a single storage
• Storage is uniform - Only storage type Disk
• Storage types hidden from the file system
New Architecture
• DataNode is a collection of storages
All disks as a single storage
cloud storage
- gs,s3,swift,wasb
SAN
...
Collection of tiered storages
Page 20 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Storage Types & Policies
Archival Storage
Flash Storage
Single replica In-Memory Tier (Tech Preview)
SSD
DISK
DISK
SSD
DISK
DISK
SSD
DISK
DISK
SSD
DISK
DISK
SSD
DISK
DISK
HDP Cluster
A
SSD
DISK
DISK
A A
SSD
All replicas on SSDDataSet A
Page 21 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Storage Calculator
Page 22 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Storage Calculator
Page 23 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Network design
● Be prepared for overhead from node failure
○ Example: a single data node with 10 TB fails
■ The cluster will produce ~10 TB of network traffic to recover
● Typically
○ Data nodes: Bonded 1 GB or Single 10GB
■ Might want to go with 10GB to future proof
○ Switches dedicated to the cluster only!
○ Rack Interconnect: 2 x 10 GB
○ 2 TOR (top of rack) switches, or multiple spines so there is no SPOF
Page 24 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Provision & Deploy
Page 25 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Unix-y
Remember: Much of Hadoop follows “Unix-like” semantics
● Keep it simple. Often you simply do things the Unix way.
● That’s a good thing!
e.g. HDFS depends on system users & groups
Page 26 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Automate!
● You must automate!
○ Successful Hadoop clusters quickly reach to 100s or 1000s of nodes.
■ Data nodes are largely identical from an OS point of view
■ 10s of nodes are added at a time
■ Problem nodes are rarely fixed live, but instead decommissioned
● Use the OS automation you already have
puppet, chef, ansible, cfengine, kickstart, gold images, just a bunch of scripts,
...
Page 27 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Provisioning workflow
● Infrastructure
● Base node OS & software configuration
● Automate installation & registration of Ambari
● Deploy with Ambari
● Choose what else your cluster needs
● Validate the cluster
● Use the cluster
Page 28 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Prepare your infrastructure
● Packages available locally
○ See our docs for mirroring our repos
○ Red Hat Channel, Spacewalk, apt mirror, …
■ Automate the process of Red Hat registration!
● Host name resolution
○ or DNS that is reliably maintained for all hosts (forward & reverse!)
○ /etc/hosts distributed & updated through automation (don’t do it manually)
● Time server (never virtualize this)
○ On an edge node if you don’t already have one
Page 29 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Prepare your nodes
Burn in!
● Have your provider do this if possible.
● Many drives fail: http://goo.gl/lCvkej
● fio, dd (with direct & non-direct io)
Don’t skip this
Page 30 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Prepare your nodes
● CPUs & Drive cache at same time: hdparm -T /dev/sda
● RAM
● Network negotiation:
○ Negotiation: ethtool eth0 | grep Speed, mii-tool, dmesg
○ Errors: ifconfig |grep errors
○ Performance: iperf
● Disk formatting & mounts
○ Choose filesystem: http://hortonworks.com/kb/linux-file-systems-for-hdfs/
○ Format: By default the file system reserves 5% of space for root!
○ Disable: mkfs.ext4 -m 0 or tune2fs -m 0
○ Mount filesystem with ‘noatime’ to stop writing of access times
● Remove unneeded services (cups, postfix, ...)
● Names: /etc/hosts or DNS (hostname -f & hostname -i): http://goo.gl/vRlOAZ
Page 31 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Prepare your nodes - OS tuning
● File Handler Limits:
○ raise nofile & nproc in /etc/limits.conf or /etc/security/limits.conf
● Disable transparent huge pages
● Disable IPv6
● Name service caching (e.g. ncsd)
● Check BIOS power management settings
● Research IO scheduler (deadline, cfq, noop)
○ Choice varies on: SSDs, RAID card, virtualized, ...
● No swapping here: vm.swappiness=0 (debated in newer kernels)
● TCP Stack tuning: such as jumbo frame (MTU 9k)
Page 32 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Ambari
Page 33 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Bootstrap Ambari
https://raw.github.com/seanorama/ambari-bootstrap
## install the ambari-server
pdsh -w server_public_hostname "curl -sSL ${bootstrap_url} | install_ambari_server=true
sh"
## install to all other nodes.
pdsh -w cluster0[2-3].hortonworks.com "curl -sSL ${bootstrap_url} |
ambari_server=server_private_hostname"
Page 34 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Ambari Considerations
Security
● Change password
● Adding Ambari users/groups
● Configure HTTPS
● Run as non-root
Ambari Agent communication
● Configure agents to self register (during automated agent installation)
○ or distribute SSH keys
Consider the database
Page 35 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Deployment: Ambari Blueprints
Definition of a cluster which enables:
● Repeatable model for cluster provisioning
○ allows for consistency
● Method to automate cluster provisioning
○ Enables ad hoc cluster creation for both bare metal and cloud
● Portable and cohesive definition of a cluster
○ Enables sharing best practices on component layout and configuration.
Hands On Demo: Deploy a cluster using a blueprint
Page 36 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Post deployment considerations
● NameNode HA
○ Additional dfs.namenode.data.dirs, possibly on NFS
● ResourceManager HA
● AD/LDAP integration
○ will increase adoption and ease user management
● Security integrations
● Capacity scheduler, multi-tenancy, ...
● Document how users access & get access to the system
○ create a mail list for your users
Page 37 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Cluster Validation
● Ambari Smoke Tests are run automatically
○ Run them manually
● Validation in the docs
○ HDP Install Documentation: http://goo.gl/USjn1g
● terasort
● Other common tests
○ DFSIO
○ HiBench
○ If using Cloud Object Storage connectors, check them (gs, s3, swift, wasb)
● Run a representative workload/job/...
Page 38 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Manage
and ongoing Operations
Page 39 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tuning: Name node tuning
● Heap size needs to be tuned as the cluster grows
○ thumb rule: 200 bytes per object, i.e. file, block, dir
■ another rule: another 1.3GB of heap per 1PB of data
○ Young generation space ~⅛ of total heap (maxed of 5GB)
○ HDP Documentation: http://goo.gl/pA9wKC
● Use parallel GC (garbage collection)
Page 40 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Configuration Management
● Base OS (network configs, DNS, system level configuration)
○ Whatever you already use: Puppet, Chef, Ansible, cfengine, just a bunch of
scripts, enterprise CM solutions, ...
● Hadoop
○ Ambari management configuration and also includes
■ versioning, history, compare, revert, recommendations
Page 41 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Monitor: What to monitor?
Integrate with existing monitoring toolsets
● System Center, Teradata ViewPoint,
● HP BSM, BMC & CA
Review Ambari Dasbboard
● HDFS Disk Usage
● DataNodes Health
● ...
● Review Alert & Metrics configuration
Some monitoring is application/database specific
● HBase, Storm
Page 42 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Monitor: HBase
Top 10 things to monitor per HBase table:
● callQueueLength
● compaction queue size
● memstore size
● slowHLogAppendCount
● get, mutation ops/sec
● mean, 95pct, 99pct latency
● GCTime
● CPU load (proc.loadaverage.1min)
● CPU allocation (system, user, iowait)
● blockcache size vs page cache size
● IO request counts (IOPS)
Page 43 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Monitor: Storm
Storm: Cluster wide metrics
● Nimbus availability
● Total slots available
For each Storm application
● Capacity: Alarm at >80%
● Latency: Alarm at deviation from expected
● Failed event count: Alarm with increasing number
Page 44 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Monitor: Kafka
● Disk space available
● Lag between reads and writes
Page 45 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Logs
Most troubleshooting is of job/task failure
● Typically drilling down to the machine and then the daemon log
● Most troubleshooting is specific to your application
Most used
● Audit: HDFS audit log - often forgotten about
● Ops: Component logs: /var/log/hadoop*
● Apps: Application logs: Land in HDFS /app-logs/
Often forgotten
● If you use Hive CLI instead of beeline: /tmp/<userid>/hive.log
Page 46 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Backup & HA
● On masters: Use your current backup method/service
○ All configuration (/etc/, /usr/hdp, …)
○ NameNode: Additional metadata mounts, some can be on remote NFS
■ dfs.namenode.data.dir
○ Databases: Ambari, Hive Metastore, Oozie, Ranger
● HDFS
○ Don’t disable Trash!
■ Configure the default expunge period (can be set per user)
○ Use Snapshots
○ All of HDFS: It depends.
■ Replicate to other clusters (Falcon, distcp, …)
● High Availability for Hadoop: http://goo.gl/BLXykB
Page 47 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Adding new nodes
● Do the same node checks, burn-in and prep
● Add in phases/groups
○ e.g. 3+ nodes at a time, half rack, or full rack
● Don’t forget to re-balance HDFS
○ Balance means:
■ % difference between storage utilisation of node and cluster
■ lower % threshold the more balanced the cluster
○ Can tune bandwidth allocation for rebalancing
Page 48 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Secure
Page 49 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Secure
Page 50 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Kerberos in the field
Kerberos no longer “too complex”. Adoption growing.
● Ambari helps automate and manage kerberos integration with cluster
Use: Active directory or a combine Kerberos/Active Directory
● Active Directory is seen most commonly in the field
● Many start with separate MIT KDC and then later grow into the AD KDC
Knox should be considered for API/Perimeter security
● Removes need for Kerberos for end users
● Enables integration with different authentication standards
● Single location to manage security for REST APIs & HTTP based services
● Tip: In DMZ
Page 51 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Authorization/Audit/Protection
Ranger provides fined grain access control and auditing
● Provides policies for HDFS, Hive, HBase, Storm and Knox
● Integrates with LDAP/AD for users/groups
● Ranger installation is automated in Ambari 2.0
● Tip: Consider using HDFS as audit store if volume is high
Encryption/Data Protection
● Several options for encryption at REST, from disk to application level
● Fine grain encryption recommended, encrypt only sensitive data
● If enterprise-wide strategy is needed, use a Partner solution
○ (Voltage, Protegrity, Vormetric, DataGuise)
Page 52 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Security Workshop
Secure HDP 2.2/2.1 using FreeIPA LDAP
Recording & documented process for
● Authentication: Configure kerberos with LDAP on sandbox
● Authorization & Audit: access policies and audit around Hadoop from central
Ranger UI, integrated with LDAP
● Perimeter Security: Configure Knox for kerberized cluster to enable perimeter
security. Integrated with LDAP/Ranger
● Protection at rest: Setup Transport Data Encryption (TDE)
● http://hortonworks.com/partners/learn
http://hortonworks.com/partners/learn
http://hortonworks.com/partners/learn
Page 53 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tenant onboarding
Provision, onboard and secure new tenants (user groups, …)
Page 54 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Multi-Tenancy & Tenant Onboarding
The request
● “As an administrator, I want to quickly provision access to new
and separate tenants across HDFS, YARN, Hive, HBase,
Storm, Accumulo, Knox, ...”
For many IT organizations, provisioning access is the most time
consuming process (not just for Hadoop).
Page 55 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Multi-Tenancy & Tenant Onboarding
The request:
● “As head of the Awesome Department, I must ensure my
applications get appropriate resources and meet SLAs”
○ “The CogApp 4.0 is lower priority than SuperApp 2”
YARN to the rescue
Page 56 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Multi-tenancy
Multi-tenancy is one cluster with
● Multiple Business Units
● Multiple Application/Jobs
YARN enables multi-tenancy with
● Shared Processing Capacity
● Shared Storage Capacity
● Data Access Security
Page 57 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
YARN Capacity Scheduler
Queues with priorities
Job submission Access Control Lists
Capacity Sharing
FUNCTION
Max capacity per queue
User limits within queue
Capacity
Enforcement
FUNCTION
Management Admin accesses Access Control Lists
Capacity Scheduler Configuration File (via UI in 2.1+ Release)
Administration
FUNCTION
Page 58 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
The YARN Platform Journey
Page 59 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Multi-Tenancy with Capacity Scheduler
Queues
▪ Hierarchical
SLAs
▪ Preemption based on priority
Resource Isolation (Containers)
▪ Linux: cgroups
▪ MS Windows: Job Control
Administration
▪ Queue ACLs
▪ Run-time re-configuration for queues
ResourceManager
Scheduler
ro
ot
Ad
ho
c
10
%
D
W
70
%
Mr
kti
ng
20
%
D
ev
10
%
R
es
er
ve
d
20
%
Pr
od
70
%
Pr
od
80
%
D
ev
20
%
P0
70
%
P1
30
%
Capacity Scheduler
Hierarchical
Queues
Page 60 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Default Queue Mapping
What
● Admin can define a default queue per user or group
● If no queue is specified for an application, YARN Capacity Scheduler will use
the user/group’s default queue for application
Why
● Queues are required for enforcing SLAs, make it easy to utilize queues
● Users and Applications want to submit Yarn apps/jobs and not have to specify
queue
● Ease migration from Fair Scheduler usage
Page 61 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Default Queue Mapping
From Ambari, add to ‘custom yarn-site’:
● yarn.scheduler.capacity.queue-mappings
Format is: [u|g]:[name]:[queue_name][,next mapping]
For example: g:mktg:marketing,u:etl:dataLoad
Good use case is mapping users to a queue of their name/group:
u:%user:%user
Page 62 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tenant Details
Tenant Details
● Tenant
○ name
○ description
○ status
○ users/groups
● Associated components: HDFS, YARN, Hive, HBase, …
Page 63 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tenant Onboarding: HDFS
HDFS
● user validation: adding /user/${user} for each tenant user
● path for tenant in HDFS. e.g. /tenant/${tenant}
● default permissions and/or extended security (ACLs or Ranger)
● quotas
○ number of files
○ total allowed space (both raw & replicated)
● scheduled HDFS snapshots
● storage type(s)
Page 64 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tenant Onboarding: Hive
Hive
● Create database(s)
● Database privileges of users associated with tenant
● Database visibility (hidden or visible in catalog to other users)
● Replication (use of Falcon to replicate between clusters)
● Scheduled snapshots of HDFS & Hive Metadata
Page 65 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Tenant Onboarding: YARN
YARN
● Top-level queue and min/max allocations
● Sub-queues, their allocations, user limits, …
Page 66 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Onboard a tenant
Live
Page 67 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Wrapping up
Page 68 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Partner Integration
Integrate through
● YARN & Slider: Simplified on-boarding of existing apps to Hadoop YARN
● Ambari Services/Stacks: Plugin new services that can co-exist with Hadoop
● Ambari Views: New ways to interact with Hadoop and visualize operations
● Ambari Blueprints/APIs: automate cluster setup in repeatable way
● Ranger plugins: manage authorization/audit of 3rd party s/w via Ranger UI
More details in upcoming Slider (Feb-26) and Ambari (Mar-26) webinars
Get started today using demos and code samples on Partners Learn page
Page 69 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Upcoming Workshops
Upcoming workshops (see Partners Learn page)
● Build YARN Ready Application with Apache Slider – Feb 26
● In Memory Processing with Apache Spark – Mar 12
● Ambari Stacks, Views and Blueprints Workshop – Mar 26
Upcoming meetup (see hortonworks.com/events) :
● Long Running Services on YARN using Slider - March 4
Page 70 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Support
Hortonworks Support
● Engineers with talent across the entire Hadoop ecosystem
● Escalate to subject matter experts for depth in a particular area
● Challenging cases may escalate to Apache committers at Hortonworks if
additional expertise is required
Apache Community Support
● user questions and support: user@hadoop.apache.org
● reporting confirmed bugs: https://issues.apache.org/jira
● Users, contributors, committers & PMC members all participate actively in
these forums to help resolve issues
Page 71 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Support
What we see
● Core Hadoop components (HDFS, YARN and MapReduce) are
used across all deployments, and therefore receive
proportionally more support cases than other ecosystem
components.
● Misconfiguration is the dominant root cause.
● We are constantly improving the code to eliminate operational
issues, help with diagnosis and provide increased visibility.
Page 72 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
Thanks
http://hortonworks.com/partners/learn/
https://github.com/seanorama/workshop-hadoop-ops

Más contenido relacionado

La actualidad más candente

GENIVI Development Platform, la plataforma para desarrolladores de software l...
GENIVI Development Platform, la plataforma para desarrolladores de software l...GENIVI Development Platform, la plataforma para desarrolladores de software l...
GENIVI Development Platform, la plataforma para desarrolladores de software l...Agustin Benito Bethencourt
 
Embedded Android Workshop with Oreo
Embedded Android Workshop with OreoEmbedded Android Workshop with Oreo
Embedded Android Workshop with OreoOpersys inc.
 
Inside Android's UI
Inside Android's UIInside Android's UI
Inside Android's UIOpersys inc.
 
Study of Chromium OS
Study of Chromium OSStudy of Chromium OS
Study of Chromium OSWilliam Lee
 
Storage Technology Overview
Storage Technology OverviewStorage Technology Overview
Storage Technology Overviewnomathjobs
 
Front-end development for headless cms
Front-end development for headless cms Front-end development for headless cms
Front-end development for headless cms Frosmo
 
ExoPlayer for Application developers
ExoPlayer for Application developersExoPlayer for Application developers
ExoPlayer for Application developersHassan Abid
 
BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InSage Weil
 
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime Ripard
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime RipardKernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime Ripard
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime RipardAnne Nicolas
 
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018Amazon Web Services
 
EMC Data Domain Retention Lock Software: Detailed Review
EMC Data Domain Retention Lock Software: Detailed ReviewEMC Data Domain Retention Lock Software: Detailed Review
EMC Data Domain Retention Lock Software: Detailed ReviewEMC
 
Building Microservices with the 12 Factor App Pattern on AWS
Building Microservices with the 12 Factor App Pattern on AWSBuilding Microservices with the 12 Factor App Pattern on AWS
Building Microservices with the 12 Factor App Pattern on AWSAmazon Web Services
 
Android Boot Time Optimization
Android Boot Time OptimizationAndroid Boot Time Optimization
Android Boot Time OptimizationKan-Ru Chen
 
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...Andrei Kholodnyi
 
Sql vs NoSQL
Sql vs NoSQLSql vs NoSQL
Sql vs NoSQLRTigger
 
Alfresco Development Framework Basic
Alfresco Development Framework BasicAlfresco Development Framework Basic
Alfresco Development Framework BasicMario Romano
 
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL database
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL databaseHBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL database
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL databaseEdureka!
 

La actualidad más candente (20)

GENIVI Development Platform, la plataforma para desarrolladores de software l...
GENIVI Development Platform, la plataforma para desarrolladores de software l...GENIVI Development Platform, la plataforma para desarrolladores de software l...
GENIVI Development Platform, la plataforma para desarrolladores de software l...
 
Embedded Android Workshop with Oreo
Embedded Android Workshop with OreoEmbedded Android Workshop with Oreo
Embedded Android Workshop with Oreo
 
redis basics
redis basicsredis basics
redis basics
 
Huawei’s All-Flash Innovation
Huawei’s All-Flash InnovationHuawei’s All-Flash Innovation
Huawei’s All-Flash Innovation
 
Inside Android's UI
Inside Android's UIInside Android's UI
Inside Android's UI
 
Study of Chromium OS
Study of Chromium OSStudy of Chromium OS
Study of Chromium OS
 
Storage Technology Overview
Storage Technology OverviewStorage Technology Overview
Storage Technology Overview
 
Front-end development for headless cms
Front-end development for headless cms Front-end development for headless cms
Front-end development for headless cms
 
ExoPlayer for Application developers
ExoPlayer for Application developersExoPlayer for Application developers
ExoPlayer for Application developers
 
BlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year InBlueStore, A New Storage Backend for Ceph, One Year In
BlueStore, A New Storage Backend for Ceph, One Year In
 
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime Ripard
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime RipardKernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime Ripard
Kernel Recipes 2017 - An introduction to the Linux DRM subsystem - Maxime Ripard
 
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018
Migrating Your NoSQL Database to Amazon DynamoDB (DAT314) - AWS re:Invent 2018
 
EMC Data Domain Retention Lock Software: Detailed Review
EMC Data Domain Retention Lock Software: Detailed ReviewEMC Data Domain Retention Lock Software: Detailed Review
EMC Data Domain Retention Lock Software: Detailed Review
 
Building Microservices with the 12 Factor App Pattern on AWS
Building Microservices with the 12 Factor App Pattern on AWSBuilding Microservices with the 12 Factor App Pattern on AWS
Building Microservices with the 12 Factor App Pattern on AWS
 
Android Boot Time Optimization
Android Boot Time OptimizationAndroid Boot Time Optimization
Android Boot Time Optimization
 
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...
Mixed-critical adaptive AUTOSAR stack based on VxWorks, Linux, and virtualiza...
 
Sql vs NoSQL
Sql vs NoSQLSql vs NoSQL
Sql vs NoSQL
 
Alfresco Development Framework Basic
Alfresco Development Framework BasicAlfresco Development Framework Basic
Alfresco Development Framework Basic
 
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL database
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL databaseHBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL database
HBase Vs Cassandra Vs MongoDB - Choosing the right NoSQL database
 
Storage Basics
Storage BasicsStorage Basics
Storage Basics
 

Similar a Hortonworks Technical Workshop - Operational Best Practices Workshop

We4IT lcty 2013 - infra-man - domino run faster
We4IT lcty 2013 - infra-man - domino run faster We4IT lcty 2013 - infra-man - domino run faster
We4IT lcty 2013 - infra-man - domino run faster We4IT Group
 
Red Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed_Hat_Storage
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Community
 
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraBackup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraCeph Community
 
Ozone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityOzone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityDinesh Chitlangia
 
Effectively deploying hadoop to the cloud
Effectively  deploying hadoop to the cloudEffectively  deploying hadoop to the cloud
Effectively deploying hadoop to the cloudAvinash Ramineni
 
Developing a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure EnvironmentsDeveloping a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure EnvironmentsCeph Community
 
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowOpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowEd Balduf
 
Hadoop operations-2015-hadoop-summit-san-jose-v5
Hadoop operations-2015-hadoop-summit-san-jose-v5Hadoop operations-2015-hadoop-summit-san-jose-v5
Hadoop operations-2015-hadoop-summit-san-jose-v5Chris Nauroth
 
Hadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldHadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldDataWorks Summit
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackOpenStack_Online
 
Read-only rootfs: theory and practice
Read-only rootfs: theory and practiceRead-only rootfs: theory and practice
Read-only rootfs: theory and practiceChris Simmonds
 
Boosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringBoosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringShapeBlue
 
Hadoop Architecture_Cluster_Cap_Plan
Hadoop Architecture_Cluster_Cap_PlanHadoop Architecture_Cluster_Cap_Plan
Hadoop Architecture_Cluster_Cap_PlanNarayana B
 
Best Practices for Virtualizing Apache Hadoop
Best Practices for Virtualizing Apache HadoopBest Practices for Virtualizing Apache Hadoop
Best Practices for Virtualizing Apache HadoopHortonworks
 
OpenStack Cinder Best Practices - Meet Up
OpenStack Cinder Best Practices - Meet UpOpenStack Cinder Best Practices - Meet Up
OpenStack Cinder Best Practices - Meet UpAaron Delp
 
Benchmarking for postgresql workloads in kubernetes
Benchmarking for postgresql workloads in kubernetesBenchmarking for postgresql workloads in kubernetes
Benchmarking for postgresql workloads in kubernetesDoKC
 
Ceph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongCeph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongPatrick McGarry
 
Hadoop Meetup Jan 2019 - Overview of Ozone
Hadoop Meetup Jan 2019 - Overview of OzoneHadoop Meetup Jan 2019 - Overview of Ozone
Hadoop Meetup Jan 2019 - Overview of OzoneErik Krogen
 

Similar a Hortonworks Technical Workshop - Operational Best Practices Workshop (20)

We4IT lcty 2013 - infra-man - domino run faster
We4IT lcty 2013 - infra-man - domino run faster We4IT lcty 2013 - infra-man - domino run faster
We4IT lcty 2013 - infra-man - domino run faster
 
Red Hat Storage for Mere Mortals
Red Hat Storage for Mere MortalsRed Hat Storage for Mere Mortals
Red Hat Storage for Mere Mortals
 
Ceph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade inCeph Day New York: Ceph: one decade in
Ceph Day New York: Ceph: one decade in
 
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix BarbeiraBackup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
Backup management with Ceph Storage - Camilo Echevarne, Félix Barbeira
 
Ozone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalabilityOzone - Evolution of hdfs scalability
Ozone - Evolution of hdfs scalability
 
Effectively deploying hadoop to the cloud
Effectively  deploying hadoop to the cloudEffectively  deploying hadoop to the cloud
Effectively deploying hadoop to the cloud
 
Developing a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure EnvironmentsDeveloping a Ceph Appliance for Secure Environments
Developing a Ceph Appliance for Secure Environments
 
VM-aware Adaptive Storage Cache Prefetching
VM-aware Adaptive Storage Cache PrefetchingVM-aware Adaptive Storage Cache Prefetching
VM-aware Adaptive Storage Cache Prefetching
 
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowOpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
 
Hadoop operations-2015-hadoop-summit-san-jose-v5
Hadoop operations-2015-hadoop-summit-san-jose-v5Hadoop operations-2015-hadoop-summit-san-jose-v5
Hadoop operations-2015-hadoop-summit-san-jose-v5
 
Hadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the FieldHadoop Operations - Best Practices from the Field
Hadoop Operations - Best Practices from the Field
 
Introduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStackIntroduction into Ceph storage for OpenStack
Introduction into Ceph storage for OpenStack
 
Read-only rootfs: theory and practice
Read-only rootfs: theory and practiceRead-only rootfs: theory and practice
Read-only rootfs: theory and practice
 
Boosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uringBoosting I/O Performance with KVM io_uring
Boosting I/O Performance with KVM io_uring
 
Hadoop Architecture_Cluster_Cap_Plan
Hadoop Architecture_Cluster_Cap_PlanHadoop Architecture_Cluster_Cap_Plan
Hadoop Architecture_Cluster_Cap_Plan
 
Best Practices for Virtualizing Apache Hadoop
Best Practices for Virtualizing Apache HadoopBest Practices for Virtualizing Apache Hadoop
Best Practices for Virtualizing Apache Hadoop
 
OpenStack Cinder Best Practices - Meet Up
OpenStack Cinder Best Practices - Meet UpOpenStack Cinder Best Practices - Meet Up
OpenStack Cinder Best Practices - Meet Up
 
Benchmarking for postgresql workloads in kubernetes
Benchmarking for postgresql workloads in kubernetesBenchmarking for postgresql workloads in kubernetes
Benchmarking for postgresql workloads in kubernetes
 
Ceph: A decade in the making and still going strong
Ceph: A decade in the making and still going strongCeph: A decade in the making and still going strong
Ceph: A decade in the making and still going strong
 
Hadoop Meetup Jan 2019 - Overview of Ozone
Hadoop Meetup Jan 2019 - Overview of OzoneHadoop Meetup Jan 2019 - Overview of Ozone
Hadoop Meetup Jan 2019 - Overview of Ozone
 

Más de Hortonworks

Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next Level
Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next LevelHortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next Level
Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next LevelHortonworks
 
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT Strategy
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT StrategyIoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT Strategy
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT StrategyHortonworks
 
Getting the Most Out of Your Data in the Cloud with Cloudbreak
Getting the Most Out of Your Data in the Cloud with CloudbreakGetting the Most Out of Your Data in the Cloud with Cloudbreak
Getting the Most Out of Your Data in the Cloud with CloudbreakHortonworks
 
Johns Hopkins - Using Hadoop to Secure Access Log Events
Johns Hopkins - Using Hadoop to Secure Access Log EventsJohns Hopkins - Using Hadoop to Secure Access Log Events
Johns Hopkins - Using Hadoop to Secure Access Log EventsHortonworks
 
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad Guys
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad GuysCatch a Hacker in Real-Time: Live Visuals of Bots and Bad Guys
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad GuysHortonworks
 
HDF 3.2 - What's New
HDF 3.2 - What's NewHDF 3.2 - What's New
HDF 3.2 - What's NewHortonworks
 
Curing Kafka Blindness with Hortonworks Streams Messaging Manager
Curing Kafka Blindness with Hortonworks Streams Messaging ManagerCuring Kafka Blindness with Hortonworks Streams Messaging Manager
Curing Kafka Blindness with Hortonworks Streams Messaging ManagerHortonworks
 
Interpretation Tool for Genomic Sequencing Data in Clinical Environments
Interpretation Tool for Genomic Sequencing Data in Clinical EnvironmentsInterpretation Tool for Genomic Sequencing Data in Clinical Environments
Interpretation Tool for Genomic Sequencing Data in Clinical EnvironmentsHortonworks
 
IBM+Hortonworks = Transformation of the Big Data Landscape
IBM+Hortonworks = Transformation of the Big Data LandscapeIBM+Hortonworks = Transformation of the Big Data Landscape
IBM+Hortonworks = Transformation of the Big Data LandscapeHortonworks
 
Premier Inside-Out: Apache Druid
Premier Inside-Out: Apache DruidPremier Inside-Out: Apache Druid
Premier Inside-Out: Apache DruidHortonworks
 
Accelerating Data Science and Real Time Analytics at Scale
Accelerating Data Science and Real Time Analytics at ScaleAccelerating Data Science and Real Time Analytics at Scale
Accelerating Data Science and Real Time Analytics at ScaleHortonworks
 
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATA
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATATIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATA
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATAHortonworks
 
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...Hortonworks
 
Delivering Real-Time Streaming Data for Healthcare Customers: Clearsense
Delivering Real-Time Streaming Data for Healthcare Customers: ClearsenseDelivering Real-Time Streaming Data for Healthcare Customers: Clearsense
Delivering Real-Time Streaming Data for Healthcare Customers: ClearsenseHortonworks
 
Making Enterprise Big Data Small with Ease
Making Enterprise Big Data Small with EaseMaking Enterprise Big Data Small with Ease
Making Enterprise Big Data Small with EaseHortonworks
 
Webinewbie to Webinerd in 30 Days - Webinar World Presentation
Webinewbie to Webinerd in 30 Days - Webinar World PresentationWebinewbie to Webinerd in 30 Days - Webinar World Presentation
Webinewbie to Webinerd in 30 Days - Webinar World PresentationHortonworks
 
Driving Digital Transformation Through Global Data Management
Driving Digital Transformation Through Global Data ManagementDriving Digital Transformation Through Global Data Management
Driving Digital Transformation Through Global Data ManagementHortonworks
 
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming Features
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming FeaturesHDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming Features
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming FeaturesHortonworks
 
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...Hortonworks
 
Unlock Value from Big Data with Apache NiFi and Streaming CDC
Unlock Value from Big Data with Apache NiFi and Streaming CDCUnlock Value from Big Data with Apache NiFi and Streaming CDC
Unlock Value from Big Data with Apache NiFi and Streaming CDCHortonworks
 

Más de Hortonworks (20)

Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next Level
Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next LevelHortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next Level
Hortonworks DataFlow (HDF) 3.3 - Taking Stream Processing to the Next Level
 
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT Strategy
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT StrategyIoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT Strategy
IoT Predictions for 2019 and Beyond: Data at the Heart of Your IoT Strategy
 
Getting the Most Out of Your Data in the Cloud with Cloudbreak
Getting the Most Out of Your Data in the Cloud with CloudbreakGetting the Most Out of Your Data in the Cloud with Cloudbreak
Getting the Most Out of Your Data in the Cloud with Cloudbreak
 
Johns Hopkins - Using Hadoop to Secure Access Log Events
Johns Hopkins - Using Hadoop to Secure Access Log EventsJohns Hopkins - Using Hadoop to Secure Access Log Events
Johns Hopkins - Using Hadoop to Secure Access Log Events
 
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad Guys
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad GuysCatch a Hacker in Real-Time: Live Visuals of Bots and Bad Guys
Catch a Hacker in Real-Time: Live Visuals of Bots and Bad Guys
 
HDF 3.2 - What's New
HDF 3.2 - What's NewHDF 3.2 - What's New
HDF 3.2 - What's New
 
Curing Kafka Blindness with Hortonworks Streams Messaging Manager
Curing Kafka Blindness with Hortonworks Streams Messaging ManagerCuring Kafka Blindness with Hortonworks Streams Messaging Manager
Curing Kafka Blindness with Hortonworks Streams Messaging Manager
 
Interpretation Tool for Genomic Sequencing Data in Clinical Environments
Interpretation Tool for Genomic Sequencing Data in Clinical EnvironmentsInterpretation Tool for Genomic Sequencing Data in Clinical Environments
Interpretation Tool for Genomic Sequencing Data in Clinical Environments
 
IBM+Hortonworks = Transformation of the Big Data Landscape
IBM+Hortonworks = Transformation of the Big Data LandscapeIBM+Hortonworks = Transformation of the Big Data Landscape
IBM+Hortonworks = Transformation of the Big Data Landscape
 
Premier Inside-Out: Apache Druid
Premier Inside-Out: Apache DruidPremier Inside-Out: Apache Druid
Premier Inside-Out: Apache Druid
 
Accelerating Data Science and Real Time Analytics at Scale
Accelerating Data Science and Real Time Analytics at ScaleAccelerating Data Science and Real Time Analytics at Scale
Accelerating Data Science and Real Time Analytics at Scale
 
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATA
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATATIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATA
TIME SERIES: APPLYING ADVANCED ANALYTICS TO INDUSTRIAL PROCESS DATA
 
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...
 
Delivering Real-Time Streaming Data for Healthcare Customers: Clearsense
Delivering Real-Time Streaming Data for Healthcare Customers: ClearsenseDelivering Real-Time Streaming Data for Healthcare Customers: Clearsense
Delivering Real-Time Streaming Data for Healthcare Customers: Clearsense
 
Making Enterprise Big Data Small with Ease
Making Enterprise Big Data Small with EaseMaking Enterprise Big Data Small with Ease
Making Enterprise Big Data Small with Ease
 
Webinewbie to Webinerd in 30 Days - Webinar World Presentation
Webinewbie to Webinerd in 30 Days - Webinar World PresentationWebinewbie to Webinerd in 30 Days - Webinar World Presentation
Webinewbie to Webinerd in 30 Days - Webinar World Presentation
 
Driving Digital Transformation Through Global Data Management
Driving Digital Transformation Through Global Data ManagementDriving Digital Transformation Through Global Data Management
Driving Digital Transformation Through Global Data Management
 
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming Features
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming FeaturesHDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming Features
HDF 3.1 pt. 2: A Technical Deep-Dive on New Streaming Features
 
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...
Hortonworks DataFlow (HDF) 3.1 - Redefining Data-In-Motion with Modern Data A...
 
Unlock Value from Big Data with Apache NiFi and Streaming CDC
Unlock Value from Big Data with Apache NiFi and Streaming CDCUnlock Value from Big Data with Apache NiFi and Streaming CDC
Unlock Value from Big Data with Apache NiFi and Streaming CDC
 

Último

Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESmohitsingh558521
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 

Último (20)

Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 

Hortonworks Technical Workshop - Operational Best Practices Workshop

  • 1. Page 1 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Operational Best Practices Workshop We Do Hadoop Sean Roberts Partner Solutions Engineer - EMEA
  • 2. Page 2 © Hortonworks Inc. 2011 – 2015. All Rights Reserved $(whoami) Sean Roberts Partner Solutions Engineer London, EMEA & everywhere @seano
  • 3. Page 3 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Operations ● Plan ● Provision & Deploy ● Secure ● Manage ● Workshop: Multi-Tenancy
  • 4. Page 4 © Hortonworks Inc. 2011 – 2015. All Rights Reserved
  • 5. Page 5 © Hortonworks Inc. 2011 – 2015. All Rights Reserved HDP - Runs Everywhere On-Premises Cloud Virtualized, Container, ... Your deployment of Hadoop • VMWare, HyperV, XEN, ... • OpenStack • Docker HDP on Your Hardware • Linux • Windows HDP on Appliance Turnkey Hadoop Appliances • Teradata • Microsoft • PSSC Labs Infrastructure as a Service (IAAS) - Microsoft Azure - Rackspace - Amazon EC2 - Google Compute Hadoop as a Service (HAAS) • Microsoft HDInsight • Rackspace
  • 6. Page 6 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Plan - Cluster Design Design choices, component distribution, and an example
  • 7. Page 7 © Hortonworks Inc. 2011 – 2015. All Rights Reserved More is better ● More is better than bigger when it comes to nodes ○ Faster healing ■ Nodes with larger storage take longer to recover ○ More nodes = more resilience, parallelism, power! ○ More racks = ...
  • 8. Page 8 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Component Layout All recommendations should be weighed against use cases Master components ● Distribute across racks to spread risk ● As cluster grows, distribution will change Worker components ● Typically identical across all worker nodes
  • 9. Page 9 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Small Cluster without HA
  • 10. Page 10 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Small Cluster with HA Secondary
  • 11. Page 11 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Component Layout Multiple gateway nodes (load balanced) ● HA/distribute client services: SSH, Knox, Ambari View instances, data loaders, ● Host other services: ntp, package repositories, Ambari View servers 5 ZooKeepers for greater reliability (default is 3) ● Also eases taking them down for maintenance ● >5 will be slower due to more voters in elections ● Consider fast disks (SAS, SSD) Some like Virtualized Masters ● Master services are not typically IO heavy / SAN compatible ● Eases hardware replacement (live vm migration)
  • 12. Page 12 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Layout of External Databases Several components require external databases: ● Ambari: PostgreSQL ● Hive: MySQL ● Oozie: Derby ● Ranger: MySQL Supported: Oracle, MySQL, PostgreSQL ● Consider using same technology for ease of management ● Use the same servers where possible ● Let your DBAs manage so backups, HA, … are taken care of Heavy usage of Falcon+Oozie or Ambari, may require dedicated instances
  • 13. Page 13 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Development cluster Your development processes still apply Ideally have a separate cluster for dev/test ● Many have dev/test combined ● Cloud: for scale up/down depending on need ● Often the pilot cluster becomes the dev cluster Smaller but same configuration ● e.g. If prod has Kerberos, HA, etc., then dev should too
  • 14. Page 14 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Plan - Hardware Selection find your sweet spot
  • 15. Page 15 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Sizing: Typical Worker CPU ● Dual socket / 8-12 core each RAM ● Typical: 128GB ● Not uncommon: 256GB HDFS Storage ● 8-12 x 2-3TB (NL-SAS/SATA) ● 1TB for performance focus ● 4TB+ for storage archive focus
  • 16. Page 16 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Hardware sizing What is your workload? ● Balanced: most common ● Memory: such as Spark ● Compute: such as Storm ● Storage: such as archive Mixed hardware in a cluster ● Ambari Configuration groups ● YARN Labels: Pin processing to subset of hardware ● Heterogeneous Storage
  • 17. Page 17 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Sizing: Storm & Kafka Storm & Kafka ● Deploy together ○ Storm is compute bound ○ Kafka is disk bound ● Same hardware as typical worker ● RAM: 128GB minimum ● Disks: 4-6 disks enough for most Kafka workloads ○ Assuming log retention of 2-3 days ● Nimbus should be on a master with HA hardware
  • 18. Page 18 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Storage Configuration Master nodes ● RAID-10 ● O/S + Data disks Data nodes ● Single O/S disk or RAID-1 ○ Consider RAID-1 in small clusters where losing a single node could be a problem Data node HDFS ● RAID-0 per disk or JBOD ○ e.g.12 disks = 12 x 1 disk arrays ○ Ensure they are mounted separately! ● Data nodes survive disk loss: ○ `dfs.datanode.failed.volumes.tolerated`
  • 19. Page 19 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Storage Types Before • DataNode is a single storage • Storage is uniform - Only storage type Disk • Storage types hidden from the file system New Architecture • DataNode is a collection of storages All disks as a single storage cloud storage - gs,s3,swift,wasb SAN ... Collection of tiered storages
  • 20. Page 20 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Storage Types & Policies Archival Storage Flash Storage Single replica In-Memory Tier (Tech Preview) SSD DISK DISK SSD DISK DISK SSD DISK DISK SSD DISK DISK SSD DISK DISK HDP Cluster A SSD DISK DISK A A SSD All replicas on SSDDataSet A
  • 21. Page 21 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Storage Calculator
  • 22. Page 22 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Storage Calculator
  • 23. Page 23 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Network design ● Be prepared for overhead from node failure ○ Example: a single data node with 10 TB fails ■ The cluster will produce ~10 TB of network traffic to recover ● Typically ○ Data nodes: Bonded 1 GB or Single 10GB ■ Might want to go with 10GB to future proof ○ Switches dedicated to the cluster only! ○ Rack Interconnect: 2 x 10 GB ○ 2 TOR (top of rack) switches, or multiple spines so there is no SPOF
  • 24. Page 24 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Provision & Deploy
  • 25. Page 25 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Unix-y Remember: Much of Hadoop follows “Unix-like” semantics ● Keep it simple. Often you simply do things the Unix way. ● That’s a good thing! e.g. HDFS depends on system users & groups
  • 26. Page 26 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Automate! ● You must automate! ○ Successful Hadoop clusters quickly reach to 100s or 1000s of nodes. ■ Data nodes are largely identical from an OS point of view ■ 10s of nodes are added at a time ■ Problem nodes are rarely fixed live, but instead decommissioned ● Use the OS automation you already have puppet, chef, ansible, cfengine, kickstart, gold images, just a bunch of scripts, ...
  • 27. Page 27 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Provisioning workflow ● Infrastructure ● Base node OS & software configuration ● Automate installation & registration of Ambari ● Deploy with Ambari ● Choose what else your cluster needs ● Validate the cluster ● Use the cluster
  • 28. Page 28 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Prepare your infrastructure ● Packages available locally ○ See our docs for mirroring our repos ○ Red Hat Channel, Spacewalk, apt mirror, … ■ Automate the process of Red Hat registration! ● Host name resolution ○ or DNS that is reliably maintained for all hosts (forward & reverse!) ○ /etc/hosts distributed & updated through automation (don’t do it manually) ● Time server (never virtualize this) ○ On an edge node if you don’t already have one
  • 29. Page 29 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Prepare your nodes Burn in! ● Have your provider do this if possible. ● Many drives fail: http://goo.gl/lCvkej ● fio, dd (with direct & non-direct io) Don’t skip this
  • 30. Page 30 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Prepare your nodes ● CPUs & Drive cache at same time: hdparm -T /dev/sda ● RAM ● Network negotiation: ○ Negotiation: ethtool eth0 | grep Speed, mii-tool, dmesg ○ Errors: ifconfig |grep errors ○ Performance: iperf ● Disk formatting & mounts ○ Choose filesystem: http://hortonworks.com/kb/linux-file-systems-for-hdfs/ ○ Format: By default the file system reserves 5% of space for root! ○ Disable: mkfs.ext4 -m 0 or tune2fs -m 0 ○ Mount filesystem with ‘noatime’ to stop writing of access times ● Remove unneeded services (cups, postfix, ...) ● Names: /etc/hosts or DNS (hostname -f & hostname -i): http://goo.gl/vRlOAZ
  • 31. Page 31 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Prepare your nodes - OS tuning ● File Handler Limits: ○ raise nofile & nproc in /etc/limits.conf or /etc/security/limits.conf ● Disable transparent huge pages ● Disable IPv6 ● Name service caching (e.g. ncsd) ● Check BIOS power management settings ● Research IO scheduler (deadline, cfq, noop) ○ Choice varies on: SSDs, RAID card, virtualized, ... ● No swapping here: vm.swappiness=0 (debated in newer kernels) ● TCP Stack tuning: such as jumbo frame (MTU 9k)
  • 32. Page 32 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Ambari
  • 33. Page 33 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Bootstrap Ambari https://raw.github.com/seanorama/ambari-bootstrap ## install the ambari-server pdsh -w server_public_hostname "curl -sSL ${bootstrap_url} | install_ambari_server=true sh" ## install to all other nodes. pdsh -w cluster0[2-3].hortonworks.com "curl -sSL ${bootstrap_url} | ambari_server=server_private_hostname"
  • 34. Page 34 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Ambari Considerations Security ● Change password ● Adding Ambari users/groups ● Configure HTTPS ● Run as non-root Ambari Agent communication ● Configure agents to self register (during automated agent installation) ○ or distribute SSH keys Consider the database
  • 35. Page 35 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Deployment: Ambari Blueprints Definition of a cluster which enables: ● Repeatable model for cluster provisioning ○ allows for consistency ● Method to automate cluster provisioning ○ Enables ad hoc cluster creation for both bare metal and cloud ● Portable and cohesive definition of a cluster ○ Enables sharing best practices on component layout and configuration. Hands On Demo: Deploy a cluster using a blueprint
  • 36. Page 36 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Post deployment considerations ● NameNode HA ○ Additional dfs.namenode.data.dirs, possibly on NFS ● ResourceManager HA ● AD/LDAP integration ○ will increase adoption and ease user management ● Security integrations ● Capacity scheduler, multi-tenancy, ... ● Document how users access & get access to the system ○ create a mail list for your users
  • 37. Page 37 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Cluster Validation ● Ambari Smoke Tests are run automatically ○ Run them manually ● Validation in the docs ○ HDP Install Documentation: http://goo.gl/USjn1g ● terasort ● Other common tests ○ DFSIO ○ HiBench ○ If using Cloud Object Storage connectors, check them (gs, s3, swift, wasb) ● Run a representative workload/job/...
  • 38. Page 38 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Manage and ongoing Operations
  • 39. Page 39 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tuning: Name node tuning ● Heap size needs to be tuned as the cluster grows ○ thumb rule: 200 bytes per object, i.e. file, block, dir ■ another rule: another 1.3GB of heap per 1PB of data ○ Young generation space ~⅛ of total heap (maxed of 5GB) ○ HDP Documentation: http://goo.gl/pA9wKC ● Use parallel GC (garbage collection)
  • 40. Page 40 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Configuration Management ● Base OS (network configs, DNS, system level configuration) ○ Whatever you already use: Puppet, Chef, Ansible, cfengine, just a bunch of scripts, enterprise CM solutions, ... ● Hadoop ○ Ambari management configuration and also includes ■ versioning, history, compare, revert, recommendations
  • 41. Page 41 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Monitor: What to monitor? Integrate with existing monitoring toolsets ● System Center, Teradata ViewPoint, ● HP BSM, BMC & CA Review Ambari Dasbboard ● HDFS Disk Usage ● DataNodes Health ● ... ● Review Alert & Metrics configuration Some monitoring is application/database specific ● HBase, Storm
  • 42. Page 42 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Monitor: HBase Top 10 things to monitor per HBase table: ● callQueueLength ● compaction queue size ● memstore size ● slowHLogAppendCount ● get, mutation ops/sec ● mean, 95pct, 99pct latency ● GCTime ● CPU load (proc.loadaverage.1min) ● CPU allocation (system, user, iowait) ● blockcache size vs page cache size ● IO request counts (IOPS)
  • 43. Page 43 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Monitor: Storm Storm: Cluster wide metrics ● Nimbus availability ● Total slots available For each Storm application ● Capacity: Alarm at >80% ● Latency: Alarm at deviation from expected ● Failed event count: Alarm with increasing number
  • 44. Page 44 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Monitor: Kafka ● Disk space available ● Lag between reads and writes
  • 45. Page 45 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Logs Most troubleshooting is of job/task failure ● Typically drilling down to the machine and then the daemon log ● Most troubleshooting is specific to your application Most used ● Audit: HDFS audit log - often forgotten about ● Ops: Component logs: /var/log/hadoop* ● Apps: Application logs: Land in HDFS /app-logs/ Often forgotten ● If you use Hive CLI instead of beeline: /tmp/<userid>/hive.log
  • 46. Page 46 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Backup & HA ● On masters: Use your current backup method/service ○ All configuration (/etc/, /usr/hdp, …) ○ NameNode: Additional metadata mounts, some can be on remote NFS ■ dfs.namenode.data.dir ○ Databases: Ambari, Hive Metastore, Oozie, Ranger ● HDFS ○ Don’t disable Trash! ■ Configure the default expunge period (can be set per user) ○ Use Snapshots ○ All of HDFS: It depends. ■ Replicate to other clusters (Falcon, distcp, …) ● High Availability for Hadoop: http://goo.gl/BLXykB
  • 47. Page 47 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Adding new nodes ● Do the same node checks, burn-in and prep ● Add in phases/groups ○ e.g. 3+ nodes at a time, half rack, or full rack ● Don’t forget to re-balance HDFS ○ Balance means: ■ % difference between storage utilisation of node and cluster ■ lower % threshold the more balanced the cluster ○ Can tune bandwidth allocation for rebalancing
  • 48. Page 48 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Secure
  • 49. Page 49 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Secure
  • 50. Page 50 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Kerberos in the field Kerberos no longer “too complex”. Adoption growing. ● Ambari helps automate and manage kerberos integration with cluster Use: Active directory or a combine Kerberos/Active Directory ● Active Directory is seen most commonly in the field ● Many start with separate MIT KDC and then later grow into the AD KDC Knox should be considered for API/Perimeter security ● Removes need for Kerberos for end users ● Enables integration with different authentication standards ● Single location to manage security for REST APIs & HTTP based services ● Tip: In DMZ
  • 51. Page 51 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Authorization/Audit/Protection Ranger provides fined grain access control and auditing ● Provides policies for HDFS, Hive, HBase, Storm and Knox ● Integrates with LDAP/AD for users/groups ● Ranger installation is automated in Ambari 2.0 ● Tip: Consider using HDFS as audit store if volume is high Encryption/Data Protection ● Several options for encryption at REST, from disk to application level ● Fine grain encryption recommended, encrypt only sensitive data ● If enterprise-wide strategy is needed, use a Partner solution ○ (Voltage, Protegrity, Vormetric, DataGuise)
  • 52. Page 52 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Security Workshop Secure HDP 2.2/2.1 using FreeIPA LDAP Recording & documented process for ● Authentication: Configure kerberos with LDAP on sandbox ● Authorization & Audit: access policies and audit around Hadoop from central Ranger UI, integrated with LDAP ● Perimeter Security: Configure Knox for kerberized cluster to enable perimeter security. Integrated with LDAP/Ranger ● Protection at rest: Setup Transport Data Encryption (TDE) ● http://hortonworks.com/partners/learn http://hortonworks.com/partners/learn http://hortonworks.com/partners/learn
  • 53. Page 53 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tenant onboarding Provision, onboard and secure new tenants (user groups, …)
  • 54. Page 54 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Multi-Tenancy & Tenant Onboarding The request ● “As an administrator, I want to quickly provision access to new and separate tenants across HDFS, YARN, Hive, HBase, Storm, Accumulo, Knox, ...” For many IT organizations, provisioning access is the most time consuming process (not just for Hadoop).
  • 55. Page 55 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Multi-Tenancy & Tenant Onboarding The request: ● “As head of the Awesome Department, I must ensure my applications get appropriate resources and meet SLAs” ○ “The CogApp 4.0 is lower priority than SuperApp 2” YARN to the rescue
  • 56. Page 56 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Multi-tenancy Multi-tenancy is one cluster with ● Multiple Business Units ● Multiple Application/Jobs YARN enables multi-tenancy with ● Shared Processing Capacity ● Shared Storage Capacity ● Data Access Security
  • 57. Page 57 © Hortonworks Inc. 2011 – 2015. All Rights Reserved YARN Capacity Scheduler Queues with priorities Job submission Access Control Lists Capacity Sharing FUNCTION Max capacity per queue User limits within queue Capacity Enforcement FUNCTION Management Admin accesses Access Control Lists Capacity Scheduler Configuration File (via UI in 2.1+ Release) Administration FUNCTION
  • 58. Page 58 © Hortonworks Inc. 2011 – 2015. All Rights Reserved The YARN Platform Journey
  • 59. Page 59 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Multi-Tenancy with Capacity Scheduler Queues ▪ Hierarchical SLAs ▪ Preemption based on priority Resource Isolation (Containers) ▪ Linux: cgroups ▪ MS Windows: Job Control Administration ▪ Queue ACLs ▪ Run-time re-configuration for queues ResourceManager Scheduler ro ot Ad ho c 10 % D W 70 % Mr kti ng 20 % D ev 10 % R es er ve d 20 % Pr od 70 % Pr od 80 % D ev 20 % P0 70 % P1 30 % Capacity Scheduler Hierarchical Queues
  • 60. Page 60 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Default Queue Mapping What ● Admin can define a default queue per user or group ● If no queue is specified for an application, YARN Capacity Scheduler will use the user/group’s default queue for application Why ● Queues are required for enforcing SLAs, make it easy to utilize queues ● Users and Applications want to submit Yarn apps/jobs and not have to specify queue ● Ease migration from Fair Scheduler usage
  • 61. Page 61 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Default Queue Mapping From Ambari, add to ‘custom yarn-site’: ● yarn.scheduler.capacity.queue-mappings Format is: [u|g]:[name]:[queue_name][,next mapping] For example: g:mktg:marketing,u:etl:dataLoad Good use case is mapping users to a queue of their name/group: u:%user:%user
  • 62. Page 62 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tenant Details Tenant Details ● Tenant ○ name ○ description ○ status ○ users/groups ● Associated components: HDFS, YARN, Hive, HBase, …
  • 63. Page 63 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tenant Onboarding: HDFS HDFS ● user validation: adding /user/${user} for each tenant user ● path for tenant in HDFS. e.g. /tenant/${tenant} ● default permissions and/or extended security (ACLs or Ranger) ● quotas ○ number of files ○ total allowed space (both raw & replicated) ● scheduled HDFS snapshots ● storage type(s)
  • 64. Page 64 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tenant Onboarding: Hive Hive ● Create database(s) ● Database privileges of users associated with tenant ● Database visibility (hidden or visible in catalog to other users) ● Replication (use of Falcon to replicate between clusters) ● Scheduled snapshots of HDFS & Hive Metadata
  • 65. Page 65 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Tenant Onboarding: YARN YARN ● Top-level queue and min/max allocations ● Sub-queues, their allocations, user limits, …
  • 66. Page 66 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Onboard a tenant Live
  • 67. Page 67 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Wrapping up
  • 68. Page 68 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Partner Integration Integrate through ● YARN & Slider: Simplified on-boarding of existing apps to Hadoop YARN ● Ambari Services/Stacks: Plugin new services that can co-exist with Hadoop ● Ambari Views: New ways to interact with Hadoop and visualize operations ● Ambari Blueprints/APIs: automate cluster setup in repeatable way ● Ranger plugins: manage authorization/audit of 3rd party s/w via Ranger UI More details in upcoming Slider (Feb-26) and Ambari (Mar-26) webinars Get started today using demos and code samples on Partners Learn page
  • 69. Page 69 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Upcoming Workshops Upcoming workshops (see Partners Learn page) ● Build YARN Ready Application with Apache Slider – Feb 26 ● In Memory Processing with Apache Spark – Mar 12 ● Ambari Stacks, Views and Blueprints Workshop – Mar 26 Upcoming meetup (see hortonworks.com/events) : ● Long Running Services on YARN using Slider - March 4
  • 70. Page 70 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Support Hortonworks Support ● Engineers with talent across the entire Hadoop ecosystem ● Escalate to subject matter experts for depth in a particular area ● Challenging cases may escalate to Apache committers at Hortonworks if additional expertise is required Apache Community Support ● user questions and support: user@hadoop.apache.org ● reporting confirmed bugs: https://issues.apache.org/jira ● Users, contributors, committers & PMC members all participate actively in these forums to help resolve issues
  • 71. Page 71 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Support What we see ● Core Hadoop components (HDFS, YARN and MapReduce) are used across all deployments, and therefore receive proportionally more support cases than other ecosystem components. ● Misconfiguration is the dominant root cause. ● We are constantly improving the code to eliminate operational issues, help with diagnosis and provide increased visibility.
  • 72. Page 72 © Hortonworks Inc. 2011 – 2015. All Rights Reserved Thanks http://hortonworks.com/partners/learn/ https://github.com/seanorama/workshop-hadoop-ops