2. 7 Key Recipes For Data Eng
Introduction
We will explore 7 key recipes
on Data Engineering.
If you could only pick one, the 5th
on joins/cogroups is essential.
2
文字数制限あり。折りたたみやエンコーディングは無し。
データ・エンジニアリングの 7大レシピ
3. 7 Key Recipes For Data Eng
About Me
Jonathan WINANDY
Scala user (6 years)
Lead Data Engineer:
- Data Lake building,
- Audit/Coaching,
- Spark/Scala/Kafka Trainings.
Founder of Univalence (BI / Big Data)
Co-Founder of CYM (Predictive Maintenance),
and Valwin (Health Care Data).
3
データエンジニアとしてデータ基盤構築やトレーニング等を実施
Univalence、CYM、Valwin などのデータ分析ビジネスを創業
4. 7 Key Recipes For Data Eng
Bachir AIT MBAREK
4
Thank you
5. 7 Key Recipes For Data Eng
Outline
1. Organisations
2. Work Optimization
3. Staging
4. RDD/Dataframe
5. Join/Cogroup
6. Data quality
7. Real Programs
5
6. 1. It’s always about our organizations!
(in Europe)
6
一に組織 (ヨーロッパはこればっかり)
7. 7 Key Recipes For Data Eng 7
1. Organisations
In Data Engineering, we tend to think our problems
come from or are solved by those tools :
データエンジニアリングではツールが問題の原因であるとか
あるいはツールによって問題を解くのだと思われがち
8. 7 Key Recipes For Data Eng
1. Organisations
However our most difficult problems or durable
solutions come from organisational contexts.
It’s true for IT at large, but it’s much more
dominant in Data Engineering.
8
IT において、最も困難な課題や持続的な解決策は組織の文脈からやってくる
この点、データエンジニアリングではさらに支配的
9. 7 Key Recipes For Data Eng
1. Organisations
9
Because Data Engineering
enables access to Data!
理由はデータ・エンジニアリングはデータへのアクセスを活性化させるから
10. 7 Key Recipes For Data Eng 10
It enables access to Data in
very complex organisations.
1. Organisations
Product BI
Your TeamMarketing
data
new data
複雑な組織においてデータアクセスを活性化させると…
11. 7 Key Recipes For Data Eng
data
11
Your Team
Global
Marketing
1. Organisations
It enables access to Data in
very complex organisations. Global
IT
Marketing IT
BI
Holding
Subsidies
Marketing IT
BI
Marketing IT
BI
「超」複雑な組織においてデータアクセスを活性化させると…
12. 7 Key Recipes For Data Eng
It happens to be very frustrating!
12
1. Organisations
By being a Data Eng, you take part in some of the most
technically diverse teams that are:
● Running cutting edge technologies,
● Solving some of the hardest problems,
while being constantly dependent on other teams that
often don’t share your vision.
先端技術を駆使して難題に取り組みつつ、ビジョンを共有しない他のチームに依存して仕
事を進めざるをえない。とてもフラストレーションが溜まる状況だ
13. 7 Key Recipes For Data Eng
1. Organisations
Small tips:
● One hadoop cluster (no Test or QA clusters).
● Document your vision, so it can be shared.
● What happens between teams matters a lot.
13
コツ: Hadoopクラスタは1つに、ビジョンは文書化して事前に根回し
チーム間の関係は大切
15. 7 Key Recipes For Data Eng
2. Work Optimization
To optimize our work, there are 3 key concerns
governing our decisions :
● Lead time,
● Impact,
● Failure management.
15
業務最適化における意思決定で大切なこと:
リードタイム、インパクト、失敗の管理
16. 7 Key Recipes For Data Eng
2. Work Optimization
Lead time:
The period of time between the
initial phase and the completion.
Impact:
Positive effects beyond the
current context.
Failure management:
Failure is the nominal case.
Unprepared failures will pile up.
16
リードタイム→企画から完成までの期間、インパクト→今の文脈を超えた良い効果失敗の
管理→想定外の失敗は積み上がる
17. 7 Key Recipes For Data Eng
2. Work Optimization
Being Proactive!
To avoid the “MapReduce then Wait”,
two methods :
● Proactive Task Simulation,
● “What will fail?”
17
先を見越して動こう!
「MapReduce を動かして待機」を回避するには?
18. 7 Key Recipes For Data Eng
2. Work Optimization
Proactive Task Simulation.
The idea is to solve a task:
● map all the possible ways,
● on each way estimate:
○ Lead time and cost,
○ Decidability,
○ Success rate,
○ Generated opportunities,
○ and other By-Products.
● then choose which way to start with.
18
解決したいタスクについて、ありうる可能性を全て挙げてリードタイムやコストなどを見積
もった上で、どの方法から始めるかを選ぶ
19. 7 Key Recipes For Data Eng
2. Work Optimization
What will fail ?
The idea is to guess what may fail on a
given component.
Then you can engage in a discussion on:
● Knowing how likely it will fail,
● Preventing that failure,
● Planning the recovery ahead.
19
あるコンポーネントで何が失敗しそうか考え、
その頻度や予防策、復旧プランを議論する
21. 7 Key Recipes For Data Eng
3. Staging
Data is moving around, freeze it!
Staging changed with Big Data. We moved from
transient staging (FTP, NFS, etc.) to persistent
staging thank to distributed solutions:
● in Kafka, we can retain logs for months,
● in HDFS, we can retain sources for years.
21
まずは、動いているデータを凍結する
Kafka や HDFS のおかげでビッグデータを長期間ステージングできるように
22. 7 Key Recipes For Data Eng
3. Staging
But there are a lot of staging
anti-patterns out there:
● Updating directories,
● Incomplete datasets,
● Short retention.
Staging should be seen as a
persistent data structure.
If you liked immutability in Scala, go for it with your Data!
22
ステージングは永続データ構造として見えるようにすべき
データは Scala のイミュータブルと同じように扱おう
23. 7 Key Recipes For Data Eng
3. Staging
Example, with HDFS:
Writing in unique directories:
/staging
|-- $tablename
|-- dtint=$dtint
|-- dsparam.name=$dsparam.value
|-- ...
|-- ...
|-- uuid=$uuid
23
UUID を使ったディレクトリに書き込む
25. 7 Key Recipes For Data Eng
4. RDD/Dataframe
Dataframes have great performance,
but are “untyped” and foreign.
RDDs have a robust Scala API,
but are a difficult to map from data sources.
SQL is the current lingua franca of Data.
25
データ操作にはなんだかんだ言っても SQL
26. 7 Key Recipes For Data Eng
4. RDD/Dataframe
Dataframe RDD
Predicate push down Types!!
Bare metal / unboxed Nested structures
Connectors Better unit tests
Pluggable Optimizer Less stages
SQL + Meta Scala * Scala
26
Comparative Advantages
27. 7 Key Recipes For Data Eng
RDD based jobs are like
marine mammals, fit for their
environnement starting from a
certain size.
RDDs are building blocks for
large jobs.
27
RDD は海獣みたいなもので、その環境に特化している
RDD は大きい仕事のビルディング・ブロック
4. RDD/Dataframe
28. 7 Key Recipes For Data Eng
4. RDD/Dataframe
RDDs are very good for ETL workloads:
● Control over shuffles,
● Unit tests are easier to write.
They can leverage Dataframe API for job
boundaries:
● Loading, storing data with Dataframe APIs,
● Map Dataframe in case classes,
● Perform type safe transformations.
28
RDD は ETL に向いている
データ順の制御や単体テストの書き易さ
29. 7 Key Recipes For Data Eng
4. RDD/Dataframe
Dataframes are perfect for:
● Data Exploration (notebook),
● Light Jobs (SQL + DF) ,
● Dynamic jobs (xlsx specs =>
spark job).
User Defined Functions improve
code reuse,
User Defined Aggregate Functions
improve performance over
Standard SQL. 29
Dataframe は Notebook を使ったデータ探索や SQL と組み合わせた軽量なジョブ、
動的なジョブに向いている
31. 7 Key Recipes For Data Eng
5. Cogroup
The cogroup is the best operation
to link data together.
31
データの連結に使える
32. 7 Key Recipes For Data Eng
Cogroup API
from (left:RDD[(K,A)],right:RDD[(K,B)])
○ join : RDD[(K,( A , B ))]
○ outerJoin : RDD[(K,(Option[A],Option[B]))]
○ cogroup : RDD[(K,( Seq[A], Seq[B]))]
from (rdd:RDD[(K,A)])
○ groupBy : RDD[(K,Seq[A])]
On cogroup and groupBy, for a given key:K, there is only
one unique row with that key in the output dataset.
5. Cogroup
32
cogroup と groupBy は任意のキーに対して単一の行を返す
33. 7 Key Recipes For Data Eng
5. Cogroup
33
rddL
.filter(pL)
.map(mL)
.keyBy(kL)
.cogroup(
rddR
.filter(pR)
.map(mR)
.keyBy(kL))
.map(mC)
34. 7 Key Recipes For Data Eng
5. Cogroup
CHECKPOINT on DISK (save)
34
rddL.keyBy(mL.andThen(kL))
.cogroup(
rddR.keyBy(mR.andThen(kR)))
.map({case (k,(ls,rs)) =>
(k,(ls.filter(pL).map(mL),
rs.filter(pR).map(mR)))})
.map(mC)
REWRITE
35. 7 Key Recipes For Data Eng
5. Cogroup
Lines of Code : 3000
Duration : 30min
(non-blocking)
Lines of Code : 15
Duration : 11h
(blocking)
35
CHECKPOINT on DISK
Moving the code after
the checkpoint allows
fast feedback loops.
ディスク書き出しの後にコードを置くことで素早くフィードバックループを回せる
36. 7 Key Recipes For Data Eng
5. Cogroup
Cogroups allow writing tests on a
minimised case.
Test workflow:
● Isolate potential cases,
● Get the smallest cogrouped row
○ output the row in test resources,
● Reproduce the bug,
● Write tests and fix code.
36
cogroup を使うと問題を最小化してテストを書けるのでバグを再現しやすい
38. 7 Key Recipes For Data Eng
6. Inline data quality
Data quality improves resilience to bad data.
However, data quality concerns often come second.
38
データ品質を高めることでバッドデータへのレジリエンスが向上するが
データ品質は二の次にされがち
39. 7 Key Recipes For Data Eng
6. Inline data quality
Our solution: Integrate Data Quality deep inside jobs, by
unifying Data quality with Data transformation.
We defined a structure Result similar to ValidationNeL
(Applicatives).
39
データ品質はジョブの奥まで統合させる
ValidationNeL的な Result というものを定義した
40. 7 Key Recipes For Data Eng
case class Result[T](value:Option[T],
annotations:Seq[Annotation])
case class Annotation(path:String,
typeName:String,
msg:String,
discardedData:Seq[String],
entityIdType:Option[String],
entityId:Option[String],
level:Int,
stage:String)
6. Inline data quality
40
41. 7 Key Recipes For Data Eng
case class Result[T](value:Option[T],
annotations:Seq[Annotation])
Result is either:
● Containing a value, with a list of warnings,
● Empty, with a list containing the error and
warnings.
(Serialization and Big Data don’t like Sum types, so it’s pre-projected
onto a product type)
6. Inline data quality
41
値を持つか、Empty の二値
それぞれ警告やエラーを表す注釈も持つ
42. 7 Key Recipes For Data Eng
case class Result[T](value:Option[T],
annotations:Seq[Annotation])
Then we can use applicatives to combine results.
case class Person(name:String,age:Int)
def build(name:Result[String],
age:Result[Int]):Result[Person] =
...
6. Inline data quality
42
アプリカティブを使って結果を組み合わせる
43. 7 Key Recipes For Data Eng
case class Result[T](value:Option[T],
annotations:Seq[Annotation])
The annotations are accumulated at the top of
the hierarchy, and saved with the data:
6. Inline data quality
43
注釈は蓄積されて、データと一緒に保存される
44. 7 Key Recipes For Data Eng
Annotations can be aggregated on dimensions:
6. Inline data quality
Message:
● EMPTY_STRING
● MULTIPLE_VALUES
● NOT_IN_ENUM
● PARSE_ERROR
● ______________
Levels:
● WARNING
● ERROR
● CRITICAL
44
注釈は次元ごとに集約できる
45. 7 Key Recipes For Data Eng
6. Inline data quality
If you are interested by the approach, you can take a look at
this repository:
Macros based on Shapeless to build Result[T] from case classes.
https://github.com/ahoy-jon/autoBuild (~october 2015)
45
気になった人はレポジトリをみてください
47. 7 Key Recipes For Data Eng
7. Real programs
Most pipelines parts are designed as
Stateless computations.
They either require no external state (great)
or infer their state based on filesystem state (meh).
47
ステートレスな計算が基本
48. 7 Key Recipes For Data Eng
7. Real programs
Spark allows us to program inside the Driver.
We can create actual programs.
In Scala, we can use:
● Scopt to parse common args and feature flips,
● TypesafeConfig to load/overload program settings,
● EventSourcing to read/write app events,
● Sbt-Pack Coursier to package and create launchers.
48
Spark なら Scala を使ってちゃんとしたプログラムが書ける
49. 7 Key Recipes For Data Eng
Deterministic effects
We then make sure that our program are as
deterministic as possible, and are
idempotent (if possible).
Example: Storing past execution so as to not recompute
something already computed, unless forced.
49
7. Real programs
できるかぎり決定論的なプログラムを目指す
50. 7 Key Recipes For Data Eng
Level 0 Event Sourcing
Level 1 Name resolving
Level 2 Triggered exec (schema capture, deltaQA, …)
Level 3 Scheduling (replay, coherence, ...)
Level 4 “code as data” (=> continuous delivery)
7. Real programs
In progress: project Kerguelen, API for data jobs.
Enable the creation of coherent jobs, integrating different
abstraction levels:
50
プロジェクト Kerguelen というものを作っている
51. 7 Key Recipes For Data Eng
8. More
More recipes:
● Automatic QA,
● Structural Sharing for Datasets,
● Jsonoids mapping generation,
● Advanced UDAF,
● ...
But that’s it for today!
51
他にもあるけど、今日はここまで
54. 7 Key Recipes For Data Eng
PSUG Note
54
If you happen to visit Paris, don’t
hesitate to submit a talk at our Paris
Scala User Group.
パリに来たら是非 Paris 勉強会でトークしてください