Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.
SPARK SHUFFLE
INTRODUCTION
天火@蘑菇街
About Me
Spark / Hadoop / Hbase / Phoenix
contributor
For spark mainly contributes in:
•  Yarn
•  Shuffle
•  BlockManager
...
Why Spark is fast(er)
•  Whom do we compare to?
•  What do we mean by fast?
•  fast to write
•  fast to run
Why Spark is fast(er) cont.
•  But the figure in previous page is some how misleading.
•  The key is the flexible programm...
What is shuffle
7
Shuffle overview
Aggregator
Aggregator
AggregatorAggregator
AggregatorAggregator
How does shuffle come into the picture
•  Spark run job stage by stage.
•  Stages are build up by DAGScheduler according t...
So everyone should have seen this before
join	
  
union	
  
groupBy	
  
map	
  
Stage	
  3	
  
Stage	
  1	
  
Stage	
  2	
...
why shuffle is expensive
•  When doing shuffle, data no longer stay in memory only
•  For spark, shuffle process might inv...
History
•  Spark 0.6-0.7, same code path with RDD’s persistent
method, can choose MEMORY_ONLY and DISK_ONLY
(default).
•  ...
LOOK
INSIDE
13
Pluggable Shuffle Framework
•  ShuffleManager
•  Manage shuffle related components, registered in SparkEnv,
configured ...
High level data flow
BlockManager
HashShuffleManager
DiskBlockManager
FileShuffleBlockManager
Local File System
SortShuffl...
15
Hash Based Shuffle - Shuffle Writer
•  Basic shuffle writer
Map	
  Task Map	
  Task Map	
  Task Map	
  Task
File
File
F...
16
Hash Based Shuffle - Shuffle Writer
•  Consolidate Shuffle Writer
Each bucket is mapping to a segment of file
Aggregato...
17
Hash Based Shuffle - Shuffle Writer
•  Basic Shuffle Writer
•  M * R shuffle spill files
•  Concurrent C * R opened shu...
18
Sort Based Shuffle - Shuffle Writer
•  Sort Shuffle Writer
Map	
  Task Map	
  Task Map	
  Task Map	
  Task
FileSegment
...
19
Sort Based Shuffle - Shuffle Writer
•  Each map task generates 1 shuffle data file + 1 index file
•  Utilize ExternalSo...
20
Hash Based Shuffle - Shuffle Reader
•  Actually, at present, Sort Based Shuffle also go with
HashShuffleReader
Bucket
B...
BLOCK
TRANSFER
SERVICE
Related conceptions
•  BlockTransferService
•  Provide a general interface for ShuffleFetcher and working with
BlockDataMa...
ShuffleManager
Data Flow
23
BlockManager
NioBlockTransferService
GetBlockData
BlockDataManager
ConnectionManager
NioBlockT...
ShuffleManager
Data Flow
24
BlockManager
NettyBlockTransferService
GetBlockData
BlockDataManager
TransportClient
NettyBloc...
External Shuffle Service
•  Design goal
•  allow for the service to be long-running
•  possibly much longer-running than S...
External Shuffle Service
•  Current Status
•  Basic framework seems ready.
•  A Network module extracted from the core mod...
28
Sort Merge Shuffle Reader
•  Background:
•  Current HashShuffleReader does not utilize the sort result within
partition...
Some shuffle related configs
•  spark.shuffle.spill (true)
•  spark.shuffle.memoryFraction (0.2)
•  spark.shuffle.manager ...
What’s next?
•  Other custom shuffle logic?
•  Alternative way to save shuffle data blocks
•  E.g. in memory (again)
•  Ot...
Thanks to Jerry Shao
•  Some of this ppt’s material came from Jerry Shao@Intel
weibo: @saisai_shao
•  Jerry also contribut...
Join Us !
• 加盟 / 合作 / 讨论 统统欢迎
• 数据平台开发,大数据相关技术,只要够
Cool,我们都玩
• tianhuo@mogujie.com
Spark shuffle introduction
Spark shuffle introduction
Spark shuffle introduction
Upcoming SlideShare
Loading in …5
×

Spark shuffle introduction

35,703 views

Published on

spark shuffle introduction, background, current status, problems, futures etc.

Published in: Data & Analytics
  • Hi there! Essay Help For Students | Discount 10% for your first order! - Check our website! https://vk.cc/80SakO
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here

Spark shuffle introduction

  1. 1. SPARK SHUFFLE INTRODUCTION 天火@蘑菇街
  2. 2. About Me Spark / Hadoop / Hbase / Phoenix contributor For spark mainly contributes in: •  Yarn •  Shuffle •  BlockManager •  Scala2.10 update •  Standalone HA •  Various other fixes. tianhuo@mogujie.com Weibo @冷冻蚂蚁 blog.csdn.net/colorant
  3. 3. Why Spark is fast(er) •  Whom do we compare to? •  What do we mean by fast? •  fast to write •  fast to run
  4. 4. Why Spark is fast(er) cont. •  But the figure in previous page is some how misleading. •  The key is the flexible programming mode. •  Which lead to more reasonable data flow. •  Which lead to less IO operation. •  Especially for iterative heavy workloads like ML. •  Which potentially cut off a lot of shuffle operations needed. •  But, you won’t always be lucky. •  Many app logic did need to exchange a lot of data. •  In the end, you will still need to deal with shuffle •  And which usually impact performance a lot.
  5. 5. What is shuffle
  6. 6. 7 Shuffle overview Aggregator Aggregator AggregatorAggregator AggregatorAggregator
  7. 7. How does shuffle come into the picture •  Spark run job stage by stage. •  Stages are build up by DAGScheduler according to RDD’s ShuffleDependency •  e.g. ShuffleRDD / CoGroupedRDD will have a ShuffleDependency •  Many operator will create ShuffleRDD / CoGroupedRDD under the hook. •  Repartition/CombineByKey/GroupBy/ReduceByKey/cogroup •  many other operator will further call into the above operators •  e.g. various join operator will call cogroup. •  Each ShuffleDependency maps to one stage in Spark Job and then will lead to a shuffle.
  8. 8. So everyone should have seen this before join   union   groupBy   map   Stage  3   Stage  1   Stage  2   A:   B:   C:   D:   E:   F:   G:  
  9. 9. why shuffle is expensive •  When doing shuffle, data no longer stay in memory only •  For spark, shuffle process might involve •  data partition: which might involve very expensive data sorting works etc. •  data ser/deser: to enable data been transfer through network or across processes. •  data compression: to reduce IO bandwidth etc. •  DISK IO: probably multiple times on one single data block •  E.g. Shuffle Spill, Merge combine
  10. 10. History •  Spark 0.6-0.7, same code path with RDD’s persistent method, can choose MEMORY_ONLY and DISK_ONLY (default). •  Spark 0.8-0.9: •  separate shuffle code path from BM and create ShuffleBlockManager and BlockObjectWriter only for shuffle, now shuffle data can only be written to disk. •  Shuffle optimization: Consolidate shuffle write. •  Spark 1.0, pluggable shuffle framework. •  Spark 1.1, sort-based shuffle implementation. •  Spark 1.2 netty transfer service reimplementation. sort- based shuffle by default •  Spark 1.2+ on the go: external shuffle service etc.
  11. 11. LOOK INSIDE
  12. 12. 13 Pluggable Shuffle Framework •  ShuffleManager •  Manage shuffle related components, registered in SparkEnv, configured through SparkConf, default is sort (pre 1.2 is hash), •  ShuffleWriter •  Handle shuffle data output logics. Will return MapStatus to be tracked by MapOutputTracker. •  ShuffleReader •  Fetch shuffle data to be used by e.g. ShuffleRDD •  ShuffleBlockManager •  Manage the mapping relation between abstract bucket and materialized data block.
  13. 13. High level data flow BlockManager HashShuffleManager DiskBlockManager FileShuffleBlockManager Local File System SortShuffleManager IndexShuffleBlockManager GetBlockData BlockTransferService GetBlockData Direct mapping or mapping by File Groups Map to One Data File and One Index File per mapId Just do one-one File mapping
  14. 14. 15 Hash Based Shuffle - Shuffle Writer •  Basic shuffle writer Map  Task Map  Task Map  Task Map  Task File File File File File File File File File File File File File File File File Aggregator AggregatorAggregatorAggregator Each bucket is mapping to a single file
  15. 15. 16 Hash Based Shuffle - Shuffle Writer •  Consolidate Shuffle Writer Each bucket is mapping to a segment of file Aggregator Aggregator Aggregator Aggregator
  16. 16. 17 Hash Based Shuffle - Shuffle Writer •  Basic Shuffle Writer •  M * R shuffle spill files •  Concurrent C * R opened shuffle files. •  If shuffle spill enabled, could generate more tmp spill files say N. •  Consolidate Shuffle Writer •  Reduce the total spilled files into C * R if (M >> C) •  Concurrent opened is the same as the basic shuffle writer. •  Memory consumption •  Thus Concurrent C * R + N file handlers. •  Each file handler could take up to 32~100KB+ Memory for various buffers across the writer stream chain.
  17. 17. 18 Sort Based Shuffle - Shuffle Writer •  Sort Shuffle Writer Map  Task Map  Task Map  Task Map  Task FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileSegment FileFile FileFile FileFile FileFile ExternalSorter ExternalSorterExternalSorterExternalSorter
  18. 18. 19 Sort Based Shuffle - Shuffle Writer •  Each map task generates 1 shuffle data file + 1 index file •  Utilize ExternalSorter to do the sort works. •  If map-side combine is required, data will be sorted by key and partition for aggregation. Otherwise data will only be sorted by partition. •  If reducer number <= 200 and no need to do aggregation or ordering, data will not be sorted at all. •  Will go with hash way and spill to separate files for each reduce partition, then merge them into one per map for final output.
  19. 19. 20 Hash Based Shuffle - Shuffle Reader •  Actually, at present, Sort Based Shuffle also go with HashShuffleReader Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Bucket Reduce  TaskReduce  Task Reduce  Task Reduce  Task Reduce  Task Aggregator AggregatorAggregatorAggregator
  20. 20. BLOCK TRANSFER SERVICE
  21. 21. Related conceptions •  BlockTransferService •  Provide a general interface for ShuffleFetcher and working with BlockDataManager to get local data. •  ShuffleClient •  Wrap up the fetching data process for the client side, say setup TransportContext, new TransportClient etc. •  TransportContext •  Context to setup the transport layer •  TransportServer •  low-level streaming service server •  TransportClient •  Client for fetching consecutive chunks TransportServer
  22. 22. ShuffleManager Data Flow 23 BlockManager NioBlockTransferService GetBlockData BlockDataManager ConnectionManager NioBlockTransferService ConnectionManager GetBlock GotBlock BlockStoreShuffleFetcher ShuffleBlockFetcherIterator Block Manager Local Blocks Remote Blocks Local Remote HashShuffleReader fetch ShuffleManager ShuffleBlockManager GetBlockData Can Switch to different BlockTransferService
  23. 23. ShuffleManager Data Flow 24 BlockManager NettyBlockTransferService GetBlockData BlockDataManager TransportClient NettyBlockTransferService TransportServer BlockStoreShuffleFetcher ShuffleBlockFetcherIterator Block Manager Local Blocks Remote Blocks Local Remote HashShuffleReader fetch ShuffleManager ShuffleBlockManager GetBlockData clientHandler TransportChannel HandlerclientHandler TransportChannel Handler Fetch Request Fetch Results
  24. 24. External Shuffle Service •  Design goal •  allow for the service to be long-running •  possibly much longer-running than Spark •  support multiple version of Spark simultaneously etc. •  can be integrated into YARN NodeManager, Standalone Worker, or on its own •  The entire service been ported to Java •  do not include Spark's dependencies •  full control over the binary compatibility of the components •  not depend on the Scala runtime or version.
  25. 25. External Shuffle Service •  Current Status •  Basic framework seems ready. •  A Network module extracted from the core module •  BlockManager could be configured with executor built-in shuffle service or external standalone shuffle service •  A standaloneWorkerShuffleService could be launched by worker •  Disabled by default. •  How it works •  Shuffle data is still written by the shuffleWriter to local disks. •  The external shuffle service knows how to read these files on disks (executor will registered related info to it, e.g. shuffle manager type, file dir layout etc.), it follow the same rules applied for written these file, so it could serve the data correctly.
  26. 26. 28 Sort Merge Shuffle Reader •  Background: •  Current HashShuffleReader does not utilize the sort result within partition in map-side. •  The actual by key sort work is always done at reduce side. •  While the map side will do by-partition sort anyway ( sort shuffle ) •  Change it to a by-key-and-partition sort does not bring many extra overhead. •  Current Status •  [WIP] https://github.com/apache/spark/pull/3438
  27. 27. Some shuffle related configs •  spark.shuffle.spill (true) •  spark.shuffle.memoryFraction (0.2) •  spark.shuffle.manager [sort]/hash •  spark.shuffle.sort.bypassMergeThreshold (200) •  spark.shuffle.blockTransferService [netty]/nio •  spark.shuffle.consolidateFiles (false) •  spark.shuffle.service.enabled (false)
  28. 28. What’s next? •  Other custom shuffle logic? •  Alternative way to save shuffle data blocks •  E.g. in memory (again) •  Other transport mechanism? •  Break stage barrier? •  To fetch shuffle data when part of the map tasks are done. •  Push mode instead of pull mode?
  29. 29. Thanks to Jerry Shao •  Some of this ppt’s material came from Jerry Shao@Intel weibo: @saisai_shao •  Jerry also contributes a lot of essential patches for spark core / spark streaming etc.
  30. 30. Join Us ! • 加盟 / 合作 / 讨论 统统欢迎 • 数据平台开发,大数据相关技术,只要够 Cool,我们都玩 • tianhuo@mogujie.com

×