44
Embracing Redundancy in Mobile Computing Jason Flinn MobiHeld 2011 Jason Flinn 1 MobiHeld 2011

Embracing Redundancy in Mobile Computing

Embed Size (px)

DESCRIPTION

Embracing Redundancy in Mobile Computing. Jason Flinn MobiHeld 2011. Game plan. A bit of a rant about mobile system design Motivated by difficulties deploying successful(?) research Motivated by common mistakes I (we?) have made. Sadly, I’m not this guy - PowerPoint PPT Presentation

Citation preview

MobiHeld 2011 1

Embracing Redundancy in Mobile Computing

Jason Flinn

MobiHeld 2011

Jason Flinn

MobiHeld 2011 2

Game plan

• A bit of a rant about mobile system design– Motivated by difficulties deploying successful(?) research– Motivated by common mistakes I (we?) have made

• Sadly, I’m not this guy– But, I will try to be a little provocative – Hopefully, we can start a good discussion

Jason Flinn

MobiHeld 2011 3

Resource scarcity

• What distinguishes mobile systems?

• One clear distinction: resource scarcity– Less powerful CPU– Less memory– Smaller storage– Poorer network quality– Limited battery power…

• What’s the correct lesson for system design?

Jason Flinn

MobiHeld 2011 4

The real lesson?

• Don’t waste CPU cycles!

• Every Joule is precious!

• Minimize network usage!

• Is this how we should build mobile systems?

Jason Flinn

Networks

Phones

Stop!

Danger!

Caution!

Warning!

MobiHeld 2011 5

Pitfall #1

• Don’t focus on computational resources– Network, battery, CPU, etc.– Easy to measure (esp. with micro-benchmarks)

• Instead, focus on human resources– E.g., interactive response time– Application latency, not network bandwidth– Remember the person waiting for results

Jason Flinn

MobiHeld 2011 6

Diversity of networks

How to exploit this?

One systems challenge

Jason Flinn

1. Use option with strongest signal?

2. Use option with lowest RTT?

3. Use option with highest bandwidth?

4. Stripe data across all networks?

All optimize a computational resource• Focus on using network efficiently• May or may not save human resources

MobiHeld 2011 7

Diversity of networks

Diversity of behavior

Email• Fetch messages

Challenge (cont.)

Media•Download video

How to minimize response time?• Low latency network for e-mail• Stripe across networks for video

Jason Flinn

MobiHeld 2011 8

Intentional Networking

• Can’t consider just network usage– Must consider impact on user response time– Unfortunately, no one-size-fits-all strategy

• Must consider intentions in using the network– Ask the user? No, the user’s time is precious– Instead, ask the applications

• work with Brett Higgins, Brian Noble, TJ Giuli, David Watson

Jason Flinn

MobiHeld 2011 9

Parallels to Parallel Programming

• We’ve seen this before!– Transition from uniprocessing to multiprocessing– Now: uninetworking to multinetworking

• What lessons can we learn from history?– Automatic parallelization is hard!– More effective to provide abstractions to applications• Applications express policies for parallelization• System provides mechanism to realize policies

Jason Flinn

MobiHeld 2011 10

Abstractions

Jason Flinn

Multiple Processors Multiple Networks

Multithreaded programs Multi-Sockets

Locks IROBS

Condition Variables Ordering Constraints

Priorities Labels

Async. Events/Signals Thunks

• Key insight: find parallel abstractions for applications

11

Abstraction: Socket

• Socket: logical connection between endpoints

Client Server

MobiHeld 2011Jason Flinn

MobiHeld 2011 12

Abstraction: Multi-socket

• Multi-socket: virtual connection– Measures performance of each alternative– Encapsulates transient network failure

Client Server

Jason Flinn

MobiHeld 2011 13

Abstraction: Label

• Qualitative description of network traffic• Size: small (latency) or large (bandwidth)• Interactivity: foreground vs. background

Client Server

Jason Flinn

MobiHeld 2011 14

Abstraction: IROB

• IROB: Isolated Reliable Ordered Bytestream• Guarantees atomic delivery of data chunk• Application specifies data, atomicity boundary

4 3 2 1

IROB 1

Jason Flinn

MobiHeld 2011 15

Abstraction: Ordering Constraints

• App specifies partial ordering on IROBs• Receiving end enforces delivery order

ServerIROB 1

IROB 3

IROB 2

IROB 2

IROB 3Dep: 2

IROB 1

use_data()

Jason Flinn

MobiHeld 2011 16

Abstraction: Thunk

• What happens when traffic should be deferred?– Application passes an optional callback + state– Borrows from PL domain

• If no suitable network is available:– Operation will fail with a special code– Callback will be fired when a suitable network appears

• Use case: periodic background messages– Send once, at the right time

Jason Flinn

MobiHeld 2011 17

Evaluation: Methodology

• Gathered network traces in a moving vehicle– Sprint 3G & open WiFi– BW up/down, RTT

• Replayed in lab (trace map

here)

Jason Flinn

MobiHeld 2011 18

Evaluation: Comparison Strategies

• Generated from the network traces• Idealized migration– Always use best-bandwidth network– Always use best-latency network

• Idealized aggregation– Aggregate bandwidth, minimum latency

• Upper bounds on app-oblivious performance

Jason Flinn

MobiHeld 2011 19

Evaluation Results: Email

• Trace #2: Ypsilanti, MI

0

10

20

30

40

50

60

70

On-demand fetch time

Best band-widthBest latency

Aggregate

Intentional Networking

Tim

e (s

econ

ds)

0

50

100

150

200

250

300

350

400BG fetch time

Tim

e (s

econ

ds)

3%

7x

Jason Flinn

MobiHeld 2011 20

Evaluation Results: Vehicular Sensing

• Trace #2: Ypsilanti, MI

0

1

2

3

4

5

6

7

Urgent update time

Best band-width

Best latency

Aggregate

Intentional Networking

Tim

e (s

econ

ds)

0

1

2

3

4

5

6

7

8BG throughput

Thro

ughp

ut (K

B/se

c)

48%

6%

Jason Flinn

MobiHeld 2011 21

But, some issues in practice…

• Followed lab evaluation with field trial• Deployed on Android phones• Lessons:– Networks failed in unpredictable ways• Hard to differentiate failures and transient delays• Timeouts either too conservative/aggressive

– Predictions not as effective as during lab evaluation• May be caused by walking vs. driving• May be caused by more urban environment

Jason Flinn

MobiHeld 2011 22

Pitfall #2

• It’s harder than we think to predict the future• Many predictors that work well in the lab– Mobile environments more unpredictable– Traces often capture only limited types of variance

• Modularity exacerbates the problem– Higher levels often assume predictions 100% correct– Pick optimal strategy given predictions

Jason Flinn

MobiHeld 2011 23

An exercise

• Number generator picks 0 or 1:– Picking the correct number wins $10– Costs $4 to pick 0– Costs $4 to pick 1

• Which one should you pick if 0 and 1 equally likely?– Both! (trick question - sorry)

• What if I told you that 0 is more likely than 1?– “How much more likely?”• “A lot” -> pick 0, “A little”, pick both• (most mobile systems just pick 0)

Jason Flinn

MobiHeld 2011 24

Strawman adaptation strategy

• Estimate future conditions (e.g., latency, bandwidth)• Calculate benefit of each option– E.g., Response time

• Calculate cost of each option– Energy, wireless minutes

• Equate diverse metrics– Utility function, constraints, etc.

• Choose the best option• Optimal, but only if predictions 100% accurate

Jason Flinn

MobiHeld 2011 25

A better strategy

• Estimates should be probability distributions– Need to understand independence of estimates

• Calculate best single option as before• But, also consider redundant strategies– E.g., send request over both networks– Does decrease in exp. response time outweigh costs?

• Redundant strategies employed when uncertain• Single strategies used when confident

Jason Flinn

MobiHeld 2011 26

Back to intentional networking

• Consider latency-sensitive traffic– Predict distribution of network latencies– Send over lowest latency network– Send over additional networks when benefit exceeds cost

• But, not quite there yet…– Predict a 10 ms. RTT for network x with high confidence– How confident are we after 20 ms. with no response?

Jason Flinn

MobiHeld 2011 27

Pitfall #3

• Re-evaluate predictions due to new information• Sometimes, lack of response is new information• Consider conditional distributions– Expected RTT given no response after n ms.– Eventually, starting redundant option makes sense

• Intentional networking: trouble mode– If no response after x ms., send request over 2nd network– Cost: bandwidth, energy – Benefit: less fragile system, less variance in response time

Jason Flinn

MobiHeld 2011 28

Generalizing: File systems

• BlueFS can read data from multiple locations– Predict latency, energy for each, fetch from best option

Jason Flinn

>/bluefsBlueFS Server

Fallacy: assumes perfect prediction• N/w degradation -> high latency• Detect n/w failures w/long timeout

Instead, embrace redundancy!• Uncertainty in device predictions• Devices largely independent• Consider fetching from >1 location

MobiHeld 2011 29

Generalizing: Cyber-foraging

• By another name: remote execution, code offload• Idea: move computation from mobile to server• Not always a good idea– Shipping inputs and outputs over n/w takes time, energy– Need to recoup costs through faster, remote execution

• Example systems:– Spectra (I’ll pick on this one)– RPF, CloneCloud, MAUI, etc.

Jason Flinn

MobiHeld 2011 30

Example: Language Translation

OutputText

DictionaryEngine

LanguageModeler

EBMTEngine

GlossaryEngine

InputText

hypotheses

4 components that could be executed remotelyExecution of each engine is optional.

Input parameters: translation type and text sizeJason Flinn

MobiHeld 2011 31

Example: Execution Plan

• Spectra chooses and execution plan:– which components to execute– where to execute them

OutputText

DictionaryEngine

LanguageModeler

EBMTEngine

GlossaryEngine

InputText

hypotheses

Remote

Local

Jason Flinn

MobiHeld 2011 32

Choosing an Execution PlanPredict ResourceSupply

Choose anExecution Plan

PredictResourceDemand

Calculate Performance,Energy, & Quality

CalculateUtility Heuristic Solver

Choose Plan w/Maximum Utility

Jason Flinn

MobiHeld 2011 33

Cyber-foraging: summary

• Fallacy: assumes many perfect predictions– Network latency, bandwidth, CPU load, file system state– Also computation, communication per input

• Reality: predicting supply and demand both hard

• A better approach:– Consider redundant execution (local & remote)– Reevaluate execution plan based on (lack of) progress

Jason Flinn

MobiHeld 2011 34

Embracing redundancy

• When does it make sense to consider redundancy?– Imperfect predictions– Options that are (largely) non-interfering– Trading computational resources for human resources

• Why is this painful?– With 20/20 hindsight, redundancy always wrong!– But, right when the future is unknown.

Jason Flinn

MobiHeld 2011 35

Pushing the needle too far?

• Sometimes embracing redundancy opens new doors• Thought experiment for cyber-foraging:– What if redundant execution is the common case?– Work with Mark Gordon, Morley Mao

Jason Flinn

MobiHeld 2011 36

• What really distinguishes mobile systems?

• One clear distinction: resource scarcity – Less powerful CPU– Less memory– Smaller storage– Poorer network quality– Limited battery power…

• What’s the correct lesson for system design?

• What really distinguishes mobile systems?

• One clear distinction: resource variability– Variable CPU load – Different platform capacities– Variable network bandwidth– Variable network latency– Varying importance of energy…

• What’s the correct lesson for system design?

Reflections

Jason Flinn

MobiHeld 2011 37

Conclusions

• Build systems robust to variance– Trade computational resources for human resources– Accept that our predictions are imperfect– Embrace redundancy

• Questions?

Jason Flinn

MobiHeld 2011 38

Deterministic replay

• Record an execution, reproduce it later– Academic and industry implementations

• Uses include:– Debugging: reproducing a software bug– Forensics: trace actions taken by an attacker– Fault tolerance: run multiple copies of execution– Many others…

Jason Flinn

MobiHeld 2011 39

How deterministic replay works

• Most parts of an execution are deterministic• Only a few source of non-determinism– E.g., system calls, context switches, and signals

Log

RecordedExecution

Initial state

Non-DeterministicEvents

ReplayedExecution

Jason Flinn

MobiHeld 2011 40

Prior approaches to improving latencyCode offload Remote execution

Compute

State

Compute(faster)

Results

Disadvantages:• Need large compute chunks• Need accurate CPU, N/W predictions

I/O

I/O

Compute

Compute

Compute

Disadvantages:• Poor interactive performance• App state disappears with N/W

Jason Flinn

MobiHeld 2011 41

Idea: Replay-enabled redundancy

• Execute 2 copies of the application– One on the mobile phone– One on the server– Use deterministic replay to make them the same

• Potential advantages:– App state still on phone if server/network dies– Less waiting: most communication asynchronous– No prediction: fastest execution generates output

Jason Flinn

MobiHeld 2011 42

Redundant execution example

Initial app state

Get user input “A”“A”

Supply “A” from log

Get user input “B”“B”

Supply “B” from log

Get “Best of both worlds” by overlapping I/O w/slower execution

Jason Flinn

MobiHeld 2011 43

How much extra communication?

• Common sources of non-determinism:– User input (infrequent, not much data)– Network I/O (server can act as n/w proxy)– File I/O (potentially can use distributed storage)– Scheduling (make scheduler more deterministic)– Sensors (usually polled infrequently)• High-data rate sensors (e.g., video) may be hard

• Hypothesis: will not send much extra data for most mobile applications

Jason Flinn

MobiHeld 2011 44

What about energy usage?

• Compare to local execution– Extra cost: sending initial state, logged events

• Several opportunities to save energy:– Faster app execution -> less energy used– No need for mobile to send network output• Server proxy will send output on mobile’s behalf

– Can skip some code execution on the mobile• Need to transmit state delta at end of skipped section• A big win if application terminates after section

Jason Flinn