CS 744: MESOS Shivaram Venkataraman Fall 2020 ADMINISTRIVIA lie - - PowerPoint PPT Presentation

cs 744 mesos
SMART_READER_LITE
LIVE PREVIEW

CS 744: MESOS Shivaram Venkataraman Fall 2020 ADMINISTRIVIA lie - - PowerPoint PPT Presentation

! morning good CS 744: MESOS Shivaram Venkataraman Fall 2020 ADMINISTRIVIA lie poll ! fill out - Assignment 1: How did it go? distributed - Assignment 2 out tonight ML - Project details - N 3 students - Create project


slide-1
SLIDE 1

CS 744: MESOS

Shivaram Venkataraman Fall 2020

good

morning

!

slide-2
SLIDE 2

ADMINISTRIVIA

  • Assignment 1: How did it go?
  • Assignment 2 out tonight
  • Project details
  • Create project groups
  • Bid for projects/Propose your own
  • Work on Introduction

→ fill out

lie poll

!

ML

distributed

  • N 3 students

next

week

1

  • 2

page

  • check
  • in
  • Final

report

and

poster

session

slide-3
SLIDE 3

COURSE FORMAT

Paper reviews “Compare, contrast and evaluate research papers” Discussion

slide-4
SLIDE 4

Scalable Storage Systems Datacenter Architecture Resource Management Computational Engines Machine Learning SQL Streaming Graph Applications

Assignment

T

MR ,

d park

→ GFS

slide-5
SLIDE 5

MapReduce GFS Spark

=

slide-6
SLIDE 6

BACKGROUND: OS SCHEDULING

code, static data heap stack code, static data heap stack code, static data heap stack CPU

How do we share CPU between processes ?

pt

P2

B

Evin

= , gee = . chrome

time sharing

rim

for

  • - 10ms

go

for

;

lo
  • as

an¥777

. . . .
  • time
slide-7
SLIDE 7

CLUSTER SCHEDULING

Scale

large

number of machines

  • ne

scheduler

?

Fairness

  • '

naff:h

space

searing

multi

  • WT

fault

tolerance

"""1M€

, C

time

  • r

/

preferences (placement

, constraint

)

pump

. aware

scheduling

slide-8
SLIDE 8

TARGET ENVIRONMENT

Multiple MapReduce versions Mix of frameworks: MPI, Spark, MR Data sharing across frameworks Avoid per-framework clusters

utilization

↳ Not all resources are

used

g→

{

Different

kinds

  • f

applications

  • n
same

cluster

  • Faris
word count 100 martinet MR

hankering .

. .

L

t !

  • in

:

F

¥

slide-9
SLIDE 9

DESIGN

ars

. . fifteenth

Two

  • level scheduling

awww

stoked

  • ¥÷m5onYs

I

↳ scheduling

across

framework

Single

per -framework

master-

scheduler

fi¥

ME

wide

scheduler

fret

^

↳ Add

new frameworks

www.oibi

"

in ke fibre

Scalability

, Flexibility
slide-10
SLIDE 10

RESOURCE OFFERS

' ""

¥m!:#

ant

Dared

7

reply

  • ffering
.

===

zcpuisgb

' ' policy "

c-

ri:

he:*

====

slide-11
SLIDE 11

CONSTRAINTS

Examples of constraints Constraints in Mesos:

  • Dita

locality

soft

Gpu

machines

hard

frameworks

can

reject

  • ffer

"

fitters

"

Boolean

functions

slide-12
SLIDE 12

DESIGN DETAILS

Allocation: Guaranteed allocation, revocation Isolation Containers (Docker)

Dai

T

  • ,kfd f To Hers
1000 an ④

short . lived

tasks ! L

, gong

running task

can
  • '

be

pre

  • empted ,
when

podcefya.me

" 4¥

Other

frameworks

express

interest

slide-13
SLIDE 13

FAULT TOLERANCE

¥:&

.

son

.

l

ft

meso,

master failure

+adf.qt.ws ¥

doesnt affect

jobs

# heartbeat

slide-14
SLIDE 14

PLACEMENT PREFERENCES

What is the problem? How do we do allocations?

↳ If

you

more

frameworks

with prep

than machines

available

in the

cluster

weighted

lottery

scheme

make

  • ffers

µportioned

in size to

the
  • verall
resources that a framework

needs

slide-15
SLIDE 15

CENTRALIZED VS DECENTRALIZED

Centralized

Decentralised

→ Scalability ~ loos
  • f frameworks

each

rloos

  • f apps
  • ptimal

solution

handle

new frameworks

1

Complexity

for framework developer

slide-16
SLIDE 16

CENTRALIZED VS DECENTRALIZED

Framework complexity Fragmentation, Starvation Inter-dependent framework ✓

→ If

resource

  • ffers
are

too

small

slide-17
SLIDE 17

COMPARISON: YARN

Per-job scheduler AM asks for resource RM replies

→ Apache Hadoop

ng

" Meroe

matter

"

¥g

per

framework

⇐ -

  • Fer - job

scheduler

slide-18
SLIDE 18

COMPARISON: BORG

Single centralized scheduler Requests mem, cpu in cfg Priority per user / service Support for quotas / reservations

→ Google

  • I Better

packing

slide-19
SLIDE 19

SUMMARY

  • Mesos: Scheduler to share cluster between Spark, MR, etc.
  • Two-level scheduling with app-specific schedulers
  • Provides scalable, decentralized scheduling
  • Pluggable Policy ? Next class!

framework

Go

slide-20
SLIDE 20

DISCUSSION

https://forms.gle/urHSeukfyipCKjue6

slide-21
SLIDE 21

What are some problems that could come up if we scale from 10 frameworks to 1000 frameworks in Mesos?

Fragmentation / starvation

  • dds

go

up

Master

bottleneck

?

it

takes

time

to

wait

for

frameworks to

reply

Mems master

pre - emption ?

Yes

. T

has

soft state

failure

recovery

takes longer ?

why ? /

n unclear ?
slide-22
SLIDE 22 ~ 2x

penni

:

":

pongee

.

O

O O

Rigid

framework

terror

:

y

Ihle!fain

MPI 's

share
slide-23
SLIDE 23

List any one difference between an OS scheduler and Mesos

Motivation

part

  • f
the

lecture

  • spark
  • n
  • versubscribed

clusters

Data locality

↳ a

:÷÷÷÷:*

. . .. ..

÷÷÷÷÷

.

→ felt pre
  • empted

cache

is

blown

away

→ shuffle files

long

lived

. coarse Grained

Layard

"

gamanteed

share

" Executor Backend
slide-24
SLIDE 24 resource
  • ffers
how does it "

perform

" "

better

" ?

I

e

" ramp
  • up
"

C-

Ci ) Time to schedule
  • ptimal

c-dis

Time to completion policy :

bonparisons

with YARN ,

Borg

slide-25
SLIDE 25

NEXT STEPS

Next class: Scheduling Policy Further reading

  • https://www.umbrant.com/2015/05/27/mesos-omega-borg-a-survey/
  • https://queue.acm.org/detail.cfm?id=3173558
thrashing

µ

"in:*

"

Athgnmentz

will

be

released

  • Delay

scheduling

  • ffer
m2 task part

wait for

Ss
  • r
so

D

' ' : Ss

after

m2
  • ffer
is made

Holik

← offer me

m2 , m3,m4