partial switch to new stopgroup. need to refactor to take advantage of child cancelation
This commit is contained in:
parent
66ca77b690
commit
4e78c08818
|
@ -64,7 +64,7 @@ func (b *BootstrapNode) Connect(conn UDPConn) error {
|
||||||
select {
|
select {
|
||||||
case <-t.C:
|
case <-t.C:
|
||||||
b.check()
|
b.check()
|
||||||
case <-b.stop.Ch():
|
case <-b.grp.Ch():
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -129,8 +129,8 @@ func (b *BootstrapNode) get(limit int) []Contact {
|
||||||
// ping pings a node. if the node responds, it is added to the list. otherwise, it is removed
|
// ping pings a node. if the node responds, it is added to the list. otherwise, it is removed
|
||||||
func (b *BootstrapNode) ping(c Contact) {
|
func (b *BootstrapNode) ping(c Contact) {
|
||||||
log.Debugf("[%s] bootstrap: pinging %s", b.id.HexShort(), c.ID.HexShort())
|
log.Debugf("[%s] bootstrap: pinging %s", b.id.HexShort(), c.ID.HexShort())
|
||||||
b.stop.Add(1)
|
b.grp.Add(1)
|
||||||
defer b.stop.Done()
|
defer b.grp.Done()
|
||||||
|
|
||||||
resCh := b.SendAsync(c, Request{Method: pingMethod})
|
resCh := b.SendAsync(c, Request{Method: pingMethod})
|
||||||
|
|
||||||
|
@ -138,7 +138,7 @@ func (b *BootstrapNode) ping(c Contact) {
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case res = <-resCh:
|
case res = <-resCh:
|
||||||
case <-b.stop.Ch():
|
case <-b.grp.Ch():
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
39
dht/dht.go
39
dht/dht.go
|
@ -12,13 +12,20 @@ import (
|
||||||
peerproto "github.com/lbryio/reflector.go/peer"
|
peerproto "github.com/lbryio/reflector.go/peer"
|
||||||
|
|
||||||
"github.com/lbryio/lbry.go/errors"
|
"github.com/lbryio/lbry.go/errors"
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
|
|
||||||
log "github.com/sirupsen/logrus"
|
"github.com/sirupsen/logrus"
|
||||||
"github.com/spf13/cast"
|
"github.com/spf13/cast"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var log *logrus.Logger
|
||||||
|
|
||||||
|
func UseLogger(l *logrus.Logger) {
|
||||||
|
log = l
|
||||||
|
}
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
log = logrus.StandardLogger()
|
||||||
//log.SetFormatter(&log.TextFormatter{ForceColors: true})
|
//log.SetFormatter(&log.TextFormatter{ForceColors: true})
|
||||||
//log.SetLevel(log.DebugLevel)
|
//log.SetLevel(log.DebugLevel)
|
||||||
}
|
}
|
||||||
|
@ -87,8 +94,8 @@ type DHT struct {
|
||||||
contact Contact
|
contact Contact
|
||||||
// node
|
// node
|
||||||
node *Node
|
node *Node
|
||||||
// stopper to shut down DHT
|
// stopGroup to shut down DHT
|
||||||
stop *stopOnce.Stopper
|
grp *stop.Group
|
||||||
// channel is closed when DHT joins network
|
// channel is closed when DHT joins network
|
||||||
joined chan struct{}
|
joined chan struct{}
|
||||||
// lock for announced list
|
// lock for announced list
|
||||||
|
@ -107,7 +114,7 @@ func New(config *Config) *DHT {
|
||||||
|
|
||||||
d := &DHT{
|
d := &DHT{
|
||||||
conf: config,
|
conf: config,
|
||||||
stop: stopOnce.New(),
|
grp: stop.New(),
|
||||||
joined: make(chan struct{}),
|
joined: make(chan struct{}),
|
||||||
lock: &sync.RWMutex{},
|
lock: &sync.RWMutex{},
|
||||||
announced: make(map[bits.Bitmap]bool),
|
announced: make(map[bits.Bitmap]bool),
|
||||||
|
@ -177,7 +184,7 @@ func (dht *DHT) join() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// now call iterativeFind on yourself
|
// now call iterativeFind on yourself
|
||||||
_, _, err := FindContacts(dht.node, dht.node.id, false, dht.stop.Ch())
|
_, _, err := FindContacts(dht.node, dht.node.id, false, dht.grp.Child())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("[%s] join: %s", dht.node.id.HexShort(), err.Error())
|
log.Errorf("[%s] join: %s", dht.node.id.HexShort(), err.Error())
|
||||||
}
|
}
|
||||||
|
@ -197,7 +204,7 @@ func (dht *DHT) WaitUntilJoined() {
|
||||||
// Shutdown shuts down the dht
|
// Shutdown shuts down the dht
|
||||||
func (dht *DHT) Shutdown() {
|
func (dht *DHT) Shutdown() {
|
||||||
log.Debugf("[%s] DHT shutting down", dht.node.id.HexShort())
|
log.Debugf("[%s] DHT shutting down", dht.node.id.HexShort())
|
||||||
dht.stop.StopAndWait()
|
dht.grp.StopAndWait()
|
||||||
dht.node.Shutdown()
|
dht.node.Shutdown()
|
||||||
log.Debugf("[%s] DHT stopped", dht.node.id.HexShort())
|
log.Debugf("[%s] DHT stopped", dht.node.id.HexShort())
|
||||||
}
|
}
|
||||||
|
@ -221,7 +228,7 @@ func (dht *DHT) Ping(addr string) error {
|
||||||
|
|
||||||
// Get returns the list of nodes that have the blob for the given hash
|
// Get returns the list of nodes that have the blob for the given hash
|
||||||
func (dht *DHT) Get(hash bits.Bitmap) ([]Contact, error) {
|
func (dht *DHT) Get(hash bits.Bitmap) ([]Contact, error) {
|
||||||
contacts, found, err := FindContacts(dht.node, hash, true, dht.stop.Ch())
|
contacts, found, err := FindContacts(dht.node, hash, true, dht.grp.Child())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
@ -242,9 +249,9 @@ func (dht *DHT) Add(hash bits.Bitmap) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
dht.stop.Add(1)
|
dht.grp.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
defer dht.stop.Done()
|
defer dht.grp.Done()
|
||||||
err := dht.announce(hash)
|
err := dht.announce(hash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error(errors.Prefix("error announcing bitmap", err))
|
log.Error(errors.Prefix("error announcing bitmap", err))
|
||||||
|
@ -254,7 +261,7 @@ func (dht *DHT) Add(hash bits.Bitmap) {
|
||||||
|
|
||||||
// Announce announces to the DHT that this node has the blob for the given hash
|
// Announce announces to the DHT that this node has the blob for the given hash
|
||||||
func (dht *DHT) announce(hash bits.Bitmap) error {
|
func (dht *DHT) announce(hash bits.Bitmap) error {
|
||||||
contacts, _, err := FindContacts(dht.node, hash, false, dht.stop.Ch())
|
contacts, _, err := FindContacts(dht.node, hash, false, dht.grp.Child())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -290,14 +297,14 @@ func (dht *DHT) startReannouncer() {
|
||||||
tick := time.NewTicker(tReannounce)
|
tick := time.NewTicker(tReannounce)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-dht.stop.Ch():
|
case <-dht.grp.Ch():
|
||||||
return
|
return
|
||||||
case <-tick.C:
|
case <-tick.C:
|
||||||
dht.lock.RLock()
|
dht.lock.RLock()
|
||||||
for h := range dht.announced {
|
for h := range dht.announced {
|
||||||
dht.stop.Add(1)
|
dht.grp.Add(1)
|
||||||
go func(bm bits.Bitmap) {
|
go func(bm bits.Bitmap) {
|
||||||
defer dht.stop.Done()
|
defer dht.grp.Done()
|
||||||
err := dht.announce(bm)
|
err := dht.announce(bm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("error re-announcing bitmap - ", err)
|
log.Error("error re-announcing bitmap - ", err)
|
||||||
|
@ -316,7 +323,7 @@ func (dht *DHT) storeOnNode(hash bits.Bitmap, c Contact) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
token := dht.tokenCache.Get(c, hash, dht.stop.Ch())
|
token := dht.tokenCache.Get(c, hash, dht.grp.Ch())
|
||||||
|
|
||||||
resCh := dht.node.SendAsync(c, Request{
|
resCh := dht.node.SendAsync(c, Request{
|
||||||
Method: storeMethod,
|
Method: storeMethod,
|
||||||
|
@ -333,7 +340,7 @@ func (dht *DHT) storeOnNode(hash bits.Bitmap, c Contact) {
|
||||||
go func() {
|
go func() {
|
||||||
select {
|
select {
|
||||||
case <-resCh:
|
case <-resCh:
|
||||||
case <-dht.stop.Ch():
|
case <-dht.grp.Ch():
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
|
|
28
dht/node.go
28
dht/node.go
|
@ -8,7 +8,7 @@ import (
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/lbryio/errors.go"
|
"github.com/lbryio/errors.go"
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
"github.com/lbryio/lbry.go/util"
|
"github.com/lbryio/lbry.go/util"
|
||||||
"github.com/lbryio/reflector.go/dht/bits"
|
"github.com/lbryio/reflector.go/dht/bits"
|
||||||
|
|
||||||
|
@ -60,7 +60,7 @@ type Node struct {
|
||||||
requestHandler RequestHandlerFunc
|
requestHandler RequestHandlerFunc
|
||||||
|
|
||||||
// stop the node neatly and clean up after itself
|
// stop the node neatly and clean up after itself
|
||||||
stop *stopOnce.Stopper
|
grp *stop.Group
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewNode returns an initialized Node's pointer.
|
// NewNode returns an initialized Node's pointer.
|
||||||
|
@ -73,7 +73,7 @@ func NewNode(id bits.Bitmap) *Node {
|
||||||
txLock: &sync.RWMutex{},
|
txLock: &sync.RWMutex{},
|
||||||
transactions: make(map[messageID]*transaction),
|
transactions: make(map[messageID]*transaction),
|
||||||
|
|
||||||
stop: stopOnce.New(),
|
grp: stop.New(),
|
||||||
tokens: &tokenManager{},
|
tokens: &tokenManager{},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -86,7 +86,7 @@ func (n *Node) Connect(conn UDPConn) error {
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
// stop tokens and close the connection when we're shutting down
|
// stop tokens and close the connection when we're shutting down
|
||||||
<-n.stop.Ch()
|
<-n.grp.Ch()
|
||||||
n.tokens.Stop()
|
n.tokens.Stop()
|
||||||
n.connClosed = true
|
n.connClosed = true
|
||||||
err := n.conn.Close()
|
err := n.conn.Close()
|
||||||
|
@ -97,9 +97,9 @@ func (n *Node) Connect(conn UDPConn) error {
|
||||||
|
|
||||||
packets := make(chan packet)
|
packets := make(chan packet)
|
||||||
|
|
||||||
n.stop.Add(1)
|
n.grp.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
defer n.stop.Done()
|
defer n.grp.Done()
|
||||||
|
|
||||||
buf := make([]byte, udpMaxMessageLength)
|
buf := make([]byte, udpMaxMessageLength)
|
||||||
|
|
||||||
|
@ -121,15 +121,15 @@ func (n *Node) Connect(conn UDPConn) error {
|
||||||
|
|
||||||
select { // needs select here because packet consumer can quit and the packets channel gets filled up and blocks
|
select { // needs select here because packet consumer can quit and the packets channel gets filled up and blocks
|
||||||
case packets <- packet{data: data, raddr: raddr}:
|
case packets <- packet{data: data, raddr: raddr}:
|
||||||
case <-n.stop.Ch():
|
case <-n.grp.Ch():
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
n.stop.Add(1)
|
n.grp.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
defer n.stop.Done()
|
defer n.grp.Done()
|
||||||
|
|
||||||
var pkt packet
|
var pkt packet
|
||||||
|
|
||||||
|
@ -137,7 +137,7 @@ func (n *Node) Connect(conn UDPConn) error {
|
||||||
select {
|
select {
|
||||||
case pkt = <-packets:
|
case pkt = <-packets:
|
||||||
n.handlePacket(pkt)
|
n.handlePacket(pkt)
|
||||||
case <-n.stop.Ch():
|
case <-n.grp.Ch():
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -156,7 +156,7 @@ func (n *Node) Connect(conn UDPConn) error {
|
||||||
// Shutdown shuts down the node
|
// Shutdown shuts down the node
|
||||||
func (n *Node) Shutdown() {
|
func (n *Node) Shutdown() {
|
||||||
log.Debugf("[%s] node shutting down", n.id.HexShort())
|
log.Debugf("[%s] node shutting down", n.id.HexShort())
|
||||||
n.stop.StopAndWait()
|
n.grp.StopAndWait()
|
||||||
log.Debugf("[%s] node stopped", n.id.HexShort())
|
log.Debugf("[%s] node stopped", n.id.HexShort())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -426,7 +426,7 @@ func (n *Node) SendAsync(contact Contact, req Request, options ...SendOptions) <
|
||||||
case res := <-tx.res:
|
case res := <-tx.res:
|
||||||
ch <- &res
|
ch <- &res
|
||||||
return
|
return
|
||||||
case <-n.stop.Ch():
|
case <-n.grp.Ch():
|
||||||
return
|
return
|
||||||
case <-time.After(udpTimeout):
|
case <-time.After(udpTimeout):
|
||||||
}
|
}
|
||||||
|
@ -457,8 +457,8 @@ func (n *Node) startRoutingTableGrooming() {
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-refreshTicker.C:
|
case <-refreshTicker.C:
|
||||||
RoutingTableRefresh(n, tRefresh, n.stop.Ch())
|
RoutingTableRefresh(n, tRefresh, n.grp.Child())
|
||||||
case <-n.stop.Ch():
|
case <-n.grp.Ch():
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -5,9 +5,9 @@ import (
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/lbryio/internal-apis/app/crypto"
|
"github.com/lbryio/lbry.go/crypto"
|
||||||
"github.com/lbryio/lbry.go/errors"
|
"github.com/lbryio/lbry.go/errors"
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
"github.com/lbryio/reflector.go/dht/bits"
|
"github.com/lbryio/reflector.go/dht/bits"
|
||||||
|
|
||||||
"github.com/sirupsen/logrus"
|
"github.com/sirupsen/logrus"
|
||||||
|
@ -23,7 +23,7 @@ func init() {
|
||||||
cfLog = logrus.StandardLogger()
|
cfLog = logrus.StandardLogger()
|
||||||
}
|
}
|
||||||
|
|
||||||
func NodeFinderUserLogger(l *logrus.Logger) {
|
func NodeFinderUseLogger(l *logrus.Logger) {
|
||||||
cfLog = l
|
cfLog = l
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -32,7 +32,7 @@ type contactFinder struct {
|
||||||
target bits.Bitmap
|
target bits.Bitmap
|
||||||
node *Node
|
node *Node
|
||||||
|
|
||||||
stop *stopOnce.Stopper
|
grp *stop.Group
|
||||||
|
|
||||||
findValueMutex *sync.Mutex
|
findValueMutex *sync.Mutex
|
||||||
findValueResult []Contact
|
findValueResult []Contact
|
||||||
|
@ -49,7 +49,7 @@ type contactFinder struct {
|
||||||
notGettingCloser *atomic.Bool
|
notGettingCloser *atomic.Bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func FindContacts(node *Node, target bits.Bitmap, findValue bool, upstreamStop stopOnce.Chan) ([]Contact, bool, error) {
|
func FindContacts(node *Node, target bits.Bitmap, findValue bool, parentGrp *stop.Group) ([]Contact, bool, error) {
|
||||||
cf := &contactFinder{
|
cf := &contactFinder{
|
||||||
node: node,
|
node: node,
|
||||||
target: target,
|
target: target,
|
||||||
|
@ -58,27 +58,16 @@ func FindContacts(node *Node, target bits.Bitmap, findValue bool, upstreamStop s
|
||||||
activeContactsMutex: &sync.Mutex{},
|
activeContactsMutex: &sync.Mutex{},
|
||||||
shortlistMutex: &sync.Mutex{},
|
shortlistMutex: &sync.Mutex{},
|
||||||
shortlistAdded: make(map[bits.Bitmap]bool),
|
shortlistAdded: make(map[bits.Bitmap]bool),
|
||||||
stop: stopOnce.New(),
|
grp: stop.New(parentGrp),
|
||||||
closestContactMutex: &sync.RWMutex{},
|
closestContactMutex: &sync.RWMutex{},
|
||||||
notGettingCloser: atomic.NewBool(false),
|
notGettingCloser: atomic.NewBool(false),
|
||||||
}
|
}
|
||||||
|
|
||||||
if upstreamStop != nil {
|
|
||||||
go func() {
|
|
||||||
select {
|
|
||||||
case <-upstreamStop:
|
|
||||||
cf.Stop()
|
|
||||||
case <-cf.stop.Ch():
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
|
|
||||||
return cf.Find()
|
return cf.Find()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (cf *contactFinder) Stop() {
|
func (cf *contactFinder) Stop() {
|
||||||
cf.stop.Stop()
|
cf.grp.StopAndWait()
|
||||||
cf.stop.Wait()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (cf *contactFinder) Find() ([]Contact, bool, error) {
|
func (cf *contactFinder) Find() ([]Contact, bool, error) {
|
||||||
|
@ -100,7 +89,7 @@ CycleLoop:
|
||||||
select {
|
select {
|
||||||
case <-time.After(timeout):
|
case <-time.After(timeout):
|
||||||
go cf.cycle(false)
|
go cf.cycle(false)
|
||||||
case <-cf.stop.Ch():
|
case <-cf.grp.Ch():
|
||||||
break CycleLoop
|
break CycleLoop
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -176,7 +165,7 @@ func (cf *contactFinder) cycle(bigCycle bool) {
|
||||||
}
|
}
|
||||||
|
|
||||||
if cf.isSearchFinished() {
|
if cf.isSearchFinished() {
|
||||||
cf.stop.Stop()
|
cf.grp.Stop()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -225,7 +214,7 @@ func (cf *contactFinder) probe(cycleID string) *Contact {
|
||||||
resCh := cf.node.SendAsync(c, req)
|
resCh := cf.node.SendAsync(c, req)
|
||||||
select {
|
select {
|
||||||
case res = <-resCh:
|
case res = <-resCh:
|
||||||
case <-cf.stop.Ch():
|
case <-cf.grp.Ch():
|
||||||
cf.debug("|%s| probe %s: canceled", cycleID, c.ID.HexShort())
|
cf.debug("|%s| probe %s: canceled", cycleID, c.ID.HexShort())
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -240,7 +229,7 @@ func (cf *contactFinder) probe(cycleID string) *Contact {
|
||||||
cf.findValueMutex.Lock()
|
cf.findValueMutex.Lock()
|
||||||
cf.findValueResult = res.Contacts
|
cf.findValueResult = res.Contacts
|
||||||
cf.findValueMutex.Unlock()
|
cf.findValueMutex.Unlock()
|
||||||
cf.stop.Stop()
|
cf.grp.Stop()
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -321,7 +310,7 @@ func (cf *contactFinder) isSearchFinished() bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case <-cf.stop.Ch():
|
case <-cf.grp.Ch():
|
||||||
return true
|
return true
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,7 +11,7 @@ import (
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/lbryio/lbry.go/errors"
|
"github.com/lbryio/lbry.go/errors"
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
"github.com/lbryio/reflector.go/dht/bits"
|
"github.com/lbryio/reflector.go/dht/bits"
|
||||||
|
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
|
@ -336,14 +336,14 @@ func (rt *routingTable) UnmarshalJSON(b []byte) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// RoutingTableRefresh refreshes any buckets that need to be refreshed
|
// RoutingTableRefresh refreshes any buckets that need to be refreshed
|
||||||
func RoutingTableRefresh(n *Node, refreshInterval time.Duration, upstreamStop stopOnce.Chan) {
|
func RoutingTableRefresh(n *Node, refreshInterval time.Duration, parentGrp *stop.Group) {
|
||||||
done := stopOnce.New()
|
done := stop.New()
|
||||||
|
|
||||||
for _, id := range n.rt.GetIDsForRefresh(refreshInterval) {
|
for _, id := range n.rt.GetIDsForRefresh(refreshInterval) {
|
||||||
done.Add(1)
|
done.Add(1)
|
||||||
go func(id bits.Bitmap) {
|
go func(id bits.Bitmap) {
|
||||||
defer done.Done()
|
defer done.Done()
|
||||||
_, _, err := FindContacts(n, id, false, upstreamStop)
|
_, _, err := FindContacts(n, id, false, parentGrp)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("error finding contact during routing table refresh - ", err)
|
log.Error("error finding contact during routing table refresh - ", err)
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,7 +6,7 @@ import (
|
||||||
|
|
||||||
"github.com/lbryio/reflector.go/dht/bits"
|
"github.com/lbryio/reflector.go/dht/bits"
|
||||||
|
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
)
|
)
|
||||||
|
|
||||||
// TODO: this should be moved out of dht and into node, and it should be completely hidden inside node. dht should not need to know about tokens
|
// TODO: this should be moved out of dht and into node, and it should be completely hidden inside node. dht should not need to know about tokens
|
||||||
|
@ -32,7 +32,7 @@ func newTokenCache(node *Node, expiration time.Duration) *tokenCache {
|
||||||
return tc
|
return tc
|
||||||
}
|
}
|
||||||
|
|
||||||
func (tc *tokenCache) Get(c Contact, hash bits.Bitmap, cancelCh stopOnce.Chan) string {
|
func (tc *tokenCache) Get(c Contact, hash bits.Bitmap, cancelCh stop.Chan) string {
|
||||||
tc.lock.RLock()
|
tc.lock.RLock()
|
||||||
token, exists := tc.tokens[c.String()]
|
token, exists := tc.tokens[c.String()]
|
||||||
tc.lock.RUnlock()
|
tc.lock.RUnlock()
|
||||||
|
|
|
@ -9,7 +9,7 @@ import (
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/lbryio/lbry.go/stopOnce"
|
"github.com/lbryio/lbry.go/stop"
|
||||||
"github.com/lbryio/reflector.go/dht/bits"
|
"github.com/lbryio/reflector.go/dht/bits"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -17,14 +17,14 @@ type tokenManager struct {
|
||||||
secret []byte
|
secret []byte
|
||||||
prevSecret []byte
|
prevSecret []byte
|
||||||
lock *sync.RWMutex
|
lock *sync.RWMutex
|
||||||
stop *stopOnce.Stopper
|
stop *stop.Group
|
||||||
}
|
}
|
||||||
|
|
||||||
func (tm *tokenManager) Start(interval time.Duration) {
|
func (tm *tokenManager) Start(interval time.Duration) {
|
||||||
tm.secret = make([]byte, 64)
|
tm.secret = make([]byte, 64)
|
||||||
tm.prevSecret = make([]byte, 64)
|
tm.prevSecret = make([]byte, 64)
|
||||||
tm.lock = &sync.RWMutex{}
|
tm.lock = &sync.RWMutex{}
|
||||||
tm.stop = stopOnce.New()
|
tm.stop = stop.New()
|
||||||
|
|
||||||
tm.rotateSecret()
|
tm.rotateSecret()
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue