|
|
|
@ -117,16 +117,9 @@ type AURPPeer struct {
|
|
|
|
|
// Event tuples yet to be sent to this peer in an RI-Upd.
|
|
|
|
|
pendingEvents chan aurp.EventTuple
|
|
|
|
|
|
|
|
|
|
// The internal states below are only set within the Handle loop, but can
|
|
|
|
|
// be read concurrently from outside.
|
|
|
|
|
mu sync.RWMutex
|
|
|
|
|
rstate ReceiverState
|
|
|
|
|
sstate SenderState
|
|
|
|
|
lastReconnect time.Time
|
|
|
|
|
lastHeardFrom time.Time
|
|
|
|
|
lastSend time.Time // TODO: clarify use of lastSend / sendRetries
|
|
|
|
|
lastUpdate time.Time
|
|
|
|
|
sendRetries int
|
|
|
|
|
mu sync.RWMutex
|
|
|
|
|
rstate ReceiverState
|
|
|
|
|
sstate SenderState
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func NewAURPPeer(routes *RouteTable, udpConn *net.UDPConn, peerAddr string, raddr *net.UDPAddr, localDI, remoteDI aurp.DomainIdentifier, connID uint16) *AURPPeer {
|
|
|
|
@ -188,7 +181,7 @@ func (p *AURPPeer) Forward(ddpkt *ddp.ExtPacket) error {
|
|
|
|
|
if err != nil {
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
_, err = p.send(p.Transport.NewAppleTalkPacket(outPkt))
|
|
|
|
|
_, err = p.Send(p.Transport.NewAppleTalkPacket(outPkt))
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
@ -204,36 +197,6 @@ func (p *AURPPeer) SenderState() SenderState {
|
|
|
|
|
return p.sstate
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) LastReconnectAgo() string {
|
|
|
|
|
p.mu.RLock()
|
|
|
|
|
defer p.mu.RUnlock()
|
|
|
|
|
return ago(p.lastReconnect)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) LastHeardFromAgo() string {
|
|
|
|
|
p.mu.RLock()
|
|
|
|
|
defer p.mu.RUnlock()
|
|
|
|
|
return ago(p.lastHeardFrom)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) LastSendAgo() string {
|
|
|
|
|
p.mu.RLock()
|
|
|
|
|
defer p.mu.RUnlock()
|
|
|
|
|
return ago(p.lastSend)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) LastUpdateAgo() string {
|
|
|
|
|
p.mu.RLock()
|
|
|
|
|
defer p.mu.RUnlock()
|
|
|
|
|
return ago(p.lastUpdate)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) SendRetries() int {
|
|
|
|
|
p.mu.RLock()
|
|
|
|
|
defer p.mu.RUnlock()
|
|
|
|
|
return p.sendRetries
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) setRState(rstate ReceiverState) {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
@ -246,42 +209,6 @@ func (p *AURPPeer) setSState(sstate SenderState) {
|
|
|
|
|
p.sstate = sstate
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) incSendRetries() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.sendRetries++
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) resetSendRetries() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.sendRetries = 0
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) bumpLastHeardFrom() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.lastHeardFrom = time.Now()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) bumpLastReconnect() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.lastReconnect = time.Now()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) bumpLastSend() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.lastSend = time.Now()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) bumpLastUpdate() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
|
p.lastUpdate = time.Now()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (p *AURPPeer) disconnect() {
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
defer p.mu.Unlock()
|
|
|
|
@ -289,8 +216,8 @@ func (p *AURPPeer) disconnect() {
|
|
|
|
|
p.sstate = SenderUnconnected
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// send encodes and sends pkt to the remote host.
|
|
|
|
|
func (p *AURPPeer) send(pkt aurp.Packet) (int, error) {
|
|
|
|
|
// Send encodes and sends pkt to the remote host.
|
|
|
|
|
func (p *AURPPeer) Send(pkt aurp.Packet) (int, error) {
|
|
|
|
|
var b bytes.Buffer
|
|
|
|
|
if _, err := pkt.WriteTo(&b); err != nil {
|
|
|
|
|
return 0, err
|
|
|
|
@ -305,20 +232,18 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
sticker := time.NewTicker(1 * time.Second)
|
|
|
|
|
defer sticker.Stop()
|
|
|
|
|
|
|
|
|
|
p.mu.Lock()
|
|
|
|
|
p.lastReconnect = time.Now()
|
|
|
|
|
p.lastHeardFrom = time.Now()
|
|
|
|
|
p.lastSend = time.Now() // TODO: clarify use of lastSend / sendRetries
|
|
|
|
|
p.lastUpdate = time.Now()
|
|
|
|
|
p.sendRetries = 0
|
|
|
|
|
p.mu.Unlock()
|
|
|
|
|
lastReconnect := time.Now()
|
|
|
|
|
lastHeardFrom := time.Now()
|
|
|
|
|
lastSend := time.Now() // TODO: clarify use of lastSend / sendRetries
|
|
|
|
|
lastUpdate := time.Now()
|
|
|
|
|
sendRetries := 0
|
|
|
|
|
|
|
|
|
|
var lastRISent aurp.Packet
|
|
|
|
|
|
|
|
|
|
p.disconnect()
|
|
|
|
|
|
|
|
|
|
// Write an Open-Req packet
|
|
|
|
|
if _, err := p.send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -334,7 +259,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
// Send a best-effort Router Down before returning
|
|
|
|
|
lastRISent = p.Transport.NewRDPacket(aurp.ErrCodeNormalClose)
|
|
|
|
|
if _, err := p.send(lastRISent); err != nil {
|
|
|
|
|
if _, err := p.Send(lastRISent); err != nil {
|
|
|
|
|
log.Printf("Couldn't send RD packet: %v", err)
|
|
|
|
|
}
|
|
|
|
|
return ctx.Err()
|
|
|
|
@ -342,60 +267,60 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
case <-rticker.C:
|
|
|
|
|
switch p.rstate {
|
|
|
|
|
case ReceiverWaitForOpenRsp:
|
|
|
|
|
if time.Since(p.lastSend) <= sendRetryTimer {
|
|
|
|
|
if time.Since(lastSend) <= sendRetryTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if p.sendRetries >= sendRetryLimit {
|
|
|
|
|
if sendRetries >= sendRetryLimit {
|
|
|
|
|
log.Printf("AURP Peer: Send retry limit reached while waiting for Open-Rsp, closing connection")
|
|
|
|
|
p.setRState(ReceiverUnconnected)
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Send another Open-Req
|
|
|
|
|
p.incSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
sendRetries++
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
if _, err := p.Send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case ReceiverConnected:
|
|
|
|
|
// Check LHFT, send tickle?
|
|
|
|
|
if time.Since(p.lastHeardFrom) <= lastHeardFromTimer {
|
|
|
|
|
if time.Since(lastHeardFrom) <= lastHeardFromTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if _, err := p.send(p.Transport.NewTicklePacket()); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewTicklePacket()); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Tickle: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
p.setRState(ReceiverWaitForTickleAck)
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
|
|
|
|
|
case ReceiverWaitForTickleAck:
|
|
|
|
|
if time.Since(p.lastSend) <= sendRetryTimer {
|
|
|
|
|
if time.Since(lastSend) <= sendRetryTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if p.sendRetries >= tickleRetryLimit {
|
|
|
|
|
if sendRetries >= tickleRetryLimit {
|
|
|
|
|
log.Printf("AURP Peer: Send retry limit reached while waiting for Tickle-Ack, closing connection")
|
|
|
|
|
p.setRState(ReceiverUnconnected)
|
|
|
|
|
p.RouteTable.DeleteAURPPeer(p)
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
p.incSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewTicklePacket()); err != nil {
|
|
|
|
|
sendRetries++
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
if _, err := p.Send(p.Transport.NewTicklePacket()); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Tickle: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
// still in Wait For Tickle-Ack
|
|
|
|
|
|
|
|
|
|
case ReceiverWaitForRIRsp:
|
|
|
|
|
if time.Since(p.lastSend) <= sendRetryTimer {
|
|
|
|
|
if time.Since(lastSend) <= sendRetryTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if p.sendRetries >= sendRetryLimit {
|
|
|
|
|
if sendRetries >= sendRetryLimit {
|
|
|
|
|
log.Printf("AURP Peer: Send retry limit reached while waiting for RI-Rsp, closing connection")
|
|
|
|
|
p.setRState(ReceiverUnconnected)
|
|
|
|
|
p.RouteTable.DeleteAURPPeer(p)
|
|
|
|
@ -404,9 +329,8 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
|
|
|
|
|
// RI-Req is stateless, so we don't need to cache the one we
|
|
|
|
|
// sent earlier just to send it again
|
|
|
|
|
p.incSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewRIReqPacket()); err != nil {
|
|
|
|
|
sendRetries++
|
|
|
|
|
if _, err := p.Send(p.Transport.NewRIReqPacket()); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -415,18 +339,18 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
case ReceiverUnconnected:
|
|
|
|
|
// Data receiver is unconnected. If data sender is connected,
|
|
|
|
|
// send a null RI-Upd to check if the sender is also unconnected
|
|
|
|
|
if p.sstate == SenderConnected && time.Since(p.lastSend) > sendRetryTimer {
|
|
|
|
|
if p.sendRetries >= sendRetryLimit {
|
|
|
|
|
log.Printf("AURP Peer: Send retry limit reached while probing sender connect, closing connection")
|
|
|
|
|
if p.sstate == SenderConnected && time.Since(lastSend) > sendRetryTimer {
|
|
|
|
|
if sendRetries >= sendRetryLimit {
|
|
|
|
|
log.Print("AURP Peer: Send retry limit reached while probing sender connect, closing connection")
|
|
|
|
|
}
|
|
|
|
|
p.incSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
sendRetries++
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
aurp.Inc(&p.Transport.LocalSeq)
|
|
|
|
|
events := aurp.EventTuples{{
|
|
|
|
|
EventCode: aurp.EventCodeNull,
|
|
|
|
|
}}
|
|
|
|
|
lastRISent = p.Transport.NewRIUpdPacket(events)
|
|
|
|
|
if _, err := p.send(lastRISent); err != nil {
|
|
|
|
|
if _, err := p.Send(lastRISent); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Upd packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -435,7 +359,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
|
|
|
|
|
if p.ConfiguredAddr != "" {
|
|
|
|
|
// Periodically try to reconnect, if this peer is in the config file
|
|
|
|
|
if time.Since(p.lastReconnect) <= reconnectTimer {
|
|
|
|
|
if time.Since(lastReconnect) <= reconnectTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
@ -448,10 +372,10 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
// log.Printf("AURP Peer: resolved %q to %v", p.ConfiguredAddr, raddr)
|
|
|
|
|
p.RemoteAddr = raddr
|
|
|
|
|
|
|
|
|
|
p.bumpLastReconnect()
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
lastReconnect = time.Now()
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
if _, err := p.Send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -465,41 +389,40 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
// Do nothing
|
|
|
|
|
|
|
|
|
|
case SenderConnected:
|
|
|
|
|
if time.Since(p.lastUpdate) <= updateTimer {
|
|
|
|
|
if time.Since(lastUpdate) <= updateTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
// TODO: is there a routing update to send?
|
|
|
|
|
p.bumpLastUpdate()
|
|
|
|
|
|
|
|
|
|
case SenderWaitForRIRspAck, SenderWaitForRIUpdAck:
|
|
|
|
|
if time.Since(p.lastSend) <= sendRetryTimer {
|
|
|
|
|
if time.Since(lastSend) <= sendRetryTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if lastRISent == nil {
|
|
|
|
|
log.Print("AURP Peer: sender retry: lastRISent = nil?")
|
|
|
|
|
continue
|
|
|
|
|
}
|
|
|
|
|
if p.sendRetries >= sendRetryLimit {
|
|
|
|
|
if sendRetries >= sendRetryLimit {
|
|
|
|
|
log.Printf("AURP Peer: Send retry limit reached, closing connection")
|
|
|
|
|
p.setSState(SenderUnconnected)
|
|
|
|
|
continue
|
|
|
|
|
}
|
|
|
|
|
p.incSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(lastRISent); err != nil {
|
|
|
|
|
sendRetries++
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
if _, err := p.Send(lastRISent); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't re-send %T: %v", lastRISent, err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case SenderWaitForRDAck:
|
|
|
|
|
if time.Since(p.lastSend) <= sendRetryTimer {
|
|
|
|
|
if time.Since(lastSend) <= sendRetryTimer {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
p.setSState(SenderUnconnected)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case pkt := <-p.ReceiveCh:
|
|
|
|
|
p.bumpLastHeardFrom()
|
|
|
|
|
lastHeardFrom = time.Now()
|
|
|
|
|
|
|
|
|
|
switch pkt := pkt.(type) {
|
|
|
|
|
case *aurp.OpenReqPacket:
|
|
|
|
@ -526,7 +449,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
orsp = p.Transport.NewOpenRspPacket(0, 1, nil)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if _, err := p.send(orsp); err != nil {
|
|
|
|
|
if _, err := p.Send(orsp); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Rsp: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -536,9 +459,9 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
|
|
|
|
|
// If receiver is unconnected, commence connecting
|
|
|
|
|
if p.rstate == ReceiverUnconnected {
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
if _, err := p.Send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -559,8 +482,8 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
p.setRState(ReceiverConnected)
|
|
|
|
|
|
|
|
|
|
// Send an RI-Req
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
if _, err := p.send(p.Transport.NewRIReqPacket()); err != nil {
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
if _, err := p.Send(p.Transport.NewRIReqPacket()); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -583,7 +506,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
p.Transport.LocalSeq = 1
|
|
|
|
|
// TODO: Split tuples across multiple packets as required
|
|
|
|
|
lastRISent = p.Transport.NewRIRspPacket(aurp.RoutingFlagLast, nets)
|
|
|
|
|
if _, err := p.send(lastRISent); err != nil {
|
|
|
|
|
if _, err := p.Send(lastRISent); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Rsp packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -608,7 +531,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
|
|
|
|
|
// TODO: track which networks we don't have zone info for, and
|
|
|
|
|
// only set SZI for those ?
|
|
|
|
|
if _, err := p.send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, aurp.RoutingFlagSendZoneInfo)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, aurp.RoutingFlagSendZoneInfo)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Ack packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -634,7 +557,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
p.setSState(SenderConnected)
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
|
|
|
|
|
// If SZI flag is set, send ZI-Rsp (transaction)
|
|
|
|
|
if pkt.Flags&aurp.RoutingFlagSendZoneInfo != 0 {
|
|
|
|
@ -659,7 +582,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
zones := p.RouteTable.ZonesForNetworks(nets)
|
|
|
|
|
// TODO: split ZI-Rsp packets similarly to ZIP Replies
|
|
|
|
|
if _, err := p.send(p.Transport.NewZIRspPacket(zones)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewZIRspPacket(zones)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send ZI-Rsp packet: %v", err)
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
@ -670,9 +593,9 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
// Receiver is unconnected, but their receiver sent us an
|
|
|
|
|
// RI-Ack for something
|
|
|
|
|
// Try to reconnect?
|
|
|
|
|
p.resetSendRetries()
|
|
|
|
|
p.bumpLastSend()
|
|
|
|
|
if _, err := p.send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
lastSend = time.Now()
|
|
|
|
|
sendRetries = 0
|
|
|
|
|
if _, err := p.Send(p.Transport.NewOpenReqPacket(nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Open-Req packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -680,6 +603,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case *aurp.RIUpdPacket:
|
|
|
|
|
|
|
|
|
|
var ackFlag aurp.RoutingFlag
|
|
|
|
|
|
|
|
|
|
for _, et := range pkt.Events {
|
|
|
|
@ -719,7 +643,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if _, err := p.send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, ackFlag)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, ackFlag)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Ack: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -733,7 +657,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
p.RouteTable.DeleteAURPPeer(p)
|
|
|
|
|
|
|
|
|
|
// Respond with RI-Ack
|
|
|
|
|
if _, err := p.send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, 0)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewRIAckPacket(pkt.ConnectionID, pkt.Sequence, 0)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send RI-Ack: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -743,7 +667,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
case *aurp.ZIReqPacket:
|
|
|
|
|
// TODO: split ZI-Rsp packets similarly to ZIP Replies
|
|
|
|
|
zones := p.RouteTable.ZonesForNetworks(pkt.Networks)
|
|
|
|
|
if _, err := p.send(p.Transport.NewZIRspPacket(zones)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewZIRspPacket(zones)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send ZI-Rsp packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -755,7 +679,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
case *aurp.GDZLReqPacket:
|
|
|
|
|
if _, err := p.send(p.Transport.NewGDZLRspPacket(-1, nil)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewGDZLRspPacket(-1, nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send GDZL-Rsp packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -764,7 +688,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
log.Printf("AURP Peer: Received a GDZL-Rsp, but I wouldn't have sent a GDZL-Req - that's weird")
|
|
|
|
|
|
|
|
|
|
case *aurp.GZNReqPacket:
|
|
|
|
|
if _, err := p.send(p.Transport.NewGZNRspPacket(pkt.ZoneName, false, nil)); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewGZNRspPacket(pkt.ZoneName, false, nil)); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send GZN-Rsp packet: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
@ -774,7 +698,7 @@ func (p *AURPPeer) Handle(ctx context.Context) error {
|
|
|
|
|
|
|
|
|
|
case *aurp.TicklePacket:
|
|
|
|
|
// Immediately respond with Tickle-Ack
|
|
|
|
|
if _, err := p.send(p.Transport.NewTickleAckPacket()); err != nil {
|
|
|
|
|
if _, err := p.Send(p.Transport.NewTickleAckPacket()); err != nil {
|
|
|
|
|
log.Printf("AURP Peer: Couldn't send Tickle-Ack: %v", err)
|
|
|
|
|
return err
|
|
|
|
|
}
|
|
|
|
|