Upload
curtis-daniel
View
214
Download
0
Embed Size (px)
Citation preview
Time Synchronization
Vinod Kulathumani
West Virginia University
2
Need for TimeSync
• Typical use
Collect sensor data, correlate with location, time
• Coordinated actuation
• Reaction to sensed data in real time
3
Needed clock properties
• Different applications have different needs
Seconds to micro-seconds
• Other parameters
Logical time or real time?
Synchronized with GPS?
Monotonic or backward correction allowed?
delta-synchronization only with neighbors?
4
Special requirements
• Efficiency
Processing
Memory
Energy
• Scalability
• Robustness
Failures
Additions
Mobility
5
Debate
• Claim: GPS is solution to all problems of keeping time, synchronizing clocks
doubtful for many wireless sensor networks, for several reasons
• Claim: Synchronizing clocks of nodes in sensor networks is not needed for applications that only collect data
actually true for some specific cases
Nodes can track the delays incurred at each hop
Base-station can punch a timestamp for received message
6
GPS based
• Relatively high-power (GPS)
• Need special GPS / antenna hardware
• Need “clear view” to transmissions
• Precision of transmitted message is in seconds (not millisecond, microsecond, etc)
• “Pulse-per-Second” (PPS) can be highly precise (1/4 microsecond), but not easy to use
• Cannot use indoors [e.g. control applications]
7
Clock hardware in sensors
• Typical sensor CPU has counters that increment by each cycle, generating interrupt upon overflow
we can keep track of time, but managing interrupts is error-prone
• External oscillator (with hardware counter) can increment, generate interrupt
even when CPU is “off” to save power
8
Uncertainties in clock hardware
• cheap, off-the-shelf oscillators
can deviate from ideal oscillator rate by one unit per 10-5 (for a microsecond counter, accuracy could diverge by up to 40 microseconds each second)
oscillator rates vary depending on power supply, temperature
9
Uncertainties in radio
• Send time: Nondeterministic, ~100ms
Delay for assembling message & passing it to MAC layer
• Access time: Nondeterministic, ~1ms-~1sec
Delay for accessing a clear transmission channel
• Transmission time: ~25ms
Time for transmitting (depends on message size, radio clock speed)
• Propagation time: <1microsecond
Time bit spends on the air (depends on distance between nodes)
• Reception time: overlaps with transmission time
• Receive time: Nondeterministic, ~100ms
Delay for processing incoming message and notifying application
10
Close up
• Interrupt handling time ~1microsec-~??microsec
Delay between radio chip raising and CPU responding
Abuses of interrupt disabling may cause problems
• Encoding time ~100microsec
Delay for putting the wave on air
• Decoding time ~100microsec
Delay for decoding wave to binary data
• Byte alignment time
Delay due to different byte alignment of sender and receiver
11
Close up…
12
Delays in message transmission
13
Protocols
• RBS (Receiver-receiver based)
• TPSN (sender-receiver based)
• Multi-hop strategies
• RBS uses zones
• TPSN uses hierarchies
• Uniform convergence time sync
14
RBS idea
• Use broadcast as a relative time reference
Broadcast packet does NOT include timestamp
Any broadcast, e.g., RTS/CTS, can be used
• A set of nodes synchronize with each other (not with the sender)
Significantly reduces non-determinism
15
Reference broadcast
• when operating system cannot record instant of message transmission (access delay unknown), but can record instant of reception
m1m1 is received simultaneously by multiple receivers: each records a timestamp value contained in m1
16
RBS: Minimizing the critical path
• RBS removes send and access time errors
Broadcast is used as a relative time reference Each receiver synchronizing to a reference packet
Ref. packet was injected into the channel at the same instant for all receivers
Message doesn’t contain timestampAlmost any broadcast packet can be used, e.g ARP, RTS/CTS,
route discovery packets, etc
All figures from Elson et. al.
17
Reference broadcast…
• after getting m1, all receivers share their local timestamps at instant of reception
now, receivers come to consensus on a value for synchronized time: for example, each adjusts local clock/counter to agree with average of local timestamps
18
RBS: Phase Offset Estimation
19
RBS: Phase Offset Estimation
20
RBS: Phase Offset Estimation
Analysis of expected group dispersion (i.e., maximum pair-wise error) after reference-broadcast synchronization. Each point represents the average of 1,000 simulated trials. The underlying receiver is assumed to report packet arrivals with error conforming to last figure. Mean group dispersion, and standard deviation from the mean, for a 2-receiver (bottom) and 20-receiver (top) group.
21
RBS: Phase Offset with Skew
Each point represents the phase offset between two nodes as implied by the value of their clocks after receiving a reference broadcast. A node can compute a least-squared error fit to these observations (diagonal line), and convert time values between its own clock and that of its peer. Synchronization of the Mote’s internal clock
2222
RBS: Multi-hop Time Synchronization
23
RBS: Multi-hop Time Synchronization
Example, we can compare the time of E1(R1) with E10(R10) by transforming E1(R1) ► E1(R4) ► E1(R8) ► E1(R10).Conversions can be made automatically by using the shortest path algorithm
24
Multi-Hop RBS Performance
Average path error is approximately n for an n hop path
The key point is the growth is not linear
25
Basic idea – sender/receiver
• Sender sends a message with local timestamp
• Receiver timestamps message upon arrival
• Forms basis for synchronization
• Could be accurate if
• Sender could generate timestamp at the instant bit was generated
• Receiver could time stamp instant when bit arrived
• Concurrent view obtained – propagation delay insignificant
26
TSPN: Synchronization phase Synchronization using handshake between a pair of nodes
(sender-initiated)
Level #Value of T1 Level #
T1, T2, T3
Assuming no clock drift and propagation delay do not change
Clock drift
Delay
A can now synchronize with B
2727
Error Analysis: TPSN & RBS
• Analyze sources of error for the algorithms
• Compare TPSN and RBS
• Trade-Offs
Level #Value of T1 Level #
T1, T2, T3
Clock drift
2828
Error Analysis: TPSN & RBS
• T2 = T1 + SA + PA->B + RB + DA->B
• T4 = T3 + SB + PB->A + RA + DB->A
= T3 + SB + PB->A + RA - DA->B
• DA->B = [(T2-T1) – (T4-T3)]/2 + [SUC + PUC + RUC]/2
• If (SA=SB) and (Delays both ways are same) and (RA = RB), then our clock drift expression is correct
• If there are uncertainities, error in synchronization is
[SUC + PUC + RUC]/2
• For RBS, the error was
[PUC + RUC]
29
Performance (TPSN vs. RBS)
30
TPSN (Multi-hop)
Clock Sync Algorithm involves 2 steps
Level Discovery phase
Synchronization phase
TSPN makes the following assumptions
Sensor nodes have unique identifiers
Node is aware of its neighbors
Bi-directional links
Creating the hierarchical tree is not exclusive to TSPN
31
TSPN: Level Discovery
Algorithm
Root node is assigned level i = 0
broadcasts “level discovery pkt.”
Neighbors get packet and assign level i+1 to themselves
Ignore discovery pkts. once level has been assigned
Broadcast “level discovery pkt.” with own level
STOP when all nodes have been assigned a level
Optimization
Use minimum spanning trees instead of flooding
3232
TSPN: Synchronization
Algorithm
Root node initiates the protocol
broadcasts “time sync pkt.”
Nodes at level = 1
Wait for a random time, initiate time-sync with root
On receiving ACK .. Synchronize with root
Nodes at level = 2 overhear sync from nodes at level 1
Do a random back-off, initiate time-sync with level 1 node
Node sends ACK only if it has been synchronized
If a node does not receive an ACK, resend time-sync pulse after a timeout.
33
Techniques for multi-hop - 1
• Regional time zones and conversion
e.g. RBS
we can compare the time of E1(R1) with E10(R10) by transforming E1(R1) ► E1(R4) ► E1(R8) ► E1(R10)
Average path error is approximately n for an n hop path
34
Techniques for multi-hop 2
• Leader based E.g. TSPN, FTSP
Leader clock at root
Others follow the leader through tree structure
• Robustness Leader failure: tolerated by new leader election
Node or link failure: tolerate by finding alternate path
Like routing table recovery
• Excellent synchronization Claim is error is constant over multi-hop [+ve, -ve neutralize]
But theoretically error grows lineraly
• Rapid setup for on-demand synchronization ?
• Suitable for low link failure, stable nodes
• Unsuitable in mobile settings / dynamic settings
35
Techniques for multi-hop 3
• Uniform convergence
• Basic idea: instead of a leader node, have all nodes follow a “leader value”
leader clock could be one with largest value
leader clock could be one with smallest value
leader value could be mean, median, etc
• local convergence -> global convergence
• send periodic timesync messages, use easy algorithm to adjust offset
if (received_time> local_clock)
local_clock= received_time
36
Uniform convergence - advantages
• Fault tolerance is automatic
• Each node takes input from all neighbors
• Mobility of sensor nodes is no problem
• Extremely simple implementation
• Self-stabilizing from all possible states and system configurations, partitions & rejoins
• Was implemented for “Line in the Sand” demonstration
37
Uniform convergence - challenges
• Even one failure can contaminate entire network (when failure introduces new, larger clock value)
• More difficult to correct skew than for tree
• How to integrate GPS or other time source?
• What does “largest clock” mean when clock reaches maximum value and rolls over?
rare occurrence, but happens someday
transient failures could cause rollover sooner
38
Preventing contamination
• Build picture of neighborhood
• Node collects timesync messages from all neighbors
• Are they all reasonably close?
yes : adjust local clock to maximum value
No: cases to consider:
more than one outlier : no consensus, adjust to maximum value
only one outlier from “consensus clock range”
if pis outlier, then p “reboots” its clock if other neighbor is outlier, ignore that neighbor
handles single-fault cases only
39
Preventing contamination
• Build picture of neighborhood
• Node collects timesync messages from all neighbors
• Are they all reasonably close?
yes : adjust local clock to maximum value
No: cases to consider:
more than one outlier : no consensus, adjust to maximum value
only one outlier from “consensus clock range”
if pis outlier, then p “reboots” its clock if other neighbor is outlier, ignore that neighbor
handles single-fault cases only
40
Special case: restarting / new node
• Again, build picture of neighborhood
• Node joining network or rebooting clock
• look for “normal” neighbors to trust
normal neighbors : copy maximum of normal neighbors
no normal neighbors : adjust local clock to maximum value from any neighbor (including restarting ones)
after adjusting to maximum, node becomes “normal”
41
Clock rollover
• Node p’s clock advances from 232-1 back to zero
• q (neighbor of p) has clock value 232-35
• question: what should q think of p’sclock?
proposal: use (<,max) cyclic ordering around domain of values [0,232-1]
42
Bad case for cyclic ordering
• Network is in “ring” topology
• values (w,x,y,z) are about ¼ of 232 apart in domain of clock values -> in ordering cycle
• Maybe, each node follows larger value of neighbor in parallel
never synchronizing!
a solution to this problem
reset to zero when neighbor clocks are too far apart, use special rule after reset
43
Open questions
• Energy conservation
• Special needs
Coordinated actuation
Long term sleeping
Low duty cycles
• Tuning time-sync to application requirements
44
References
• Slides on time synchronization by Prof Ted Herman, University of Iowa
• Timing-sync Protocol for Sensor Networks (TPSN)
Ganeriwal S, Kumar R, Srivastava M [Sensys 2003]
• Fine-Grained Network Time Synchronization using Reference Broadcasts
Elson J, Girod L, Estrin D [OSDI 2002]
• The Flooding Time Synchronization Protocol
Maroti M, Kusy B [Sensys 2004]