Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

NPeersForCpl and collapse empty buckets #77

Merged
merged 7 commits into from
May 11, 2020
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
51 changes: 49 additions & 2 deletions table.go
Original file line number Diff line number Diff line change
Expand Up @@ -85,11 +85,37 @@ func (rt *RoutingTable) Close() error {
return nil
}

// TryAddPeer tries to add a peer to the Routing table. If the peer ALREADY exists in the Routing Table, this call is a no-op.
// NPeersForCPL returns the number of peers we have for a given Cpl
func (rt *RoutingTable) NPeersForCpl(cpl uint) int {

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can you get away without this function, and just call
len(GetPeersForCpl) instead?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Have removed GetPeersForCpl as we don't need it anymore.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We've removed GetPeersForCpl as we don't need it anymore.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I guess this is a tiny bit more efficient then just len(GetPeersForCpl) and we can always add GetPeersForCpl back in if we need it. Your call 😄

aschmahmann marked this conversation as resolved.
Show resolved Hide resolved
rt.tabLock.RLock()
defer rt.tabLock.RUnlock()

// it's in the last bucket
if int(cpl) >= len(rt.buckets)-1 {
count := 0
b := rt.buckets[len(rt.buckets)-1]
for _, p := range b.peers() {
if CommonPrefixLen(rt.local, p.dhtId) == int(cpl) {
count++
}
}
return count
} else {
return rt.buckets[cpl].len()
}
}

// TryAddPeer tries to add a peer to the Routing table.
// If the peer ALREADY exists in the Routing Table and has been queried before, this call is a no-op.
// If the peer ALREADY exists in the Routing Table but hasn't been queried before, we set it's LastUsefulAt value to
// the current time. This needs to done because we don't mark peers as "Useful"(by setting the LastUsefulAt value)
// when we first connect to them.
//
// If the peer is a queryPeer i.e. we queried it or it queried us, we set the LastSuccessfulOutboundQuery to the current time.
// If the peer is just a peer that we connect to/it connected to us without any DHT query, we consider it as having
// no LastSuccessfulOutboundQuery.
//
//
// If the logical bucket to which the peer belongs is full and it's not the last bucket, we try to replace an existing peer
// whose LastSuccessfulOutboundQuery is above the maximum allowed threshold in that bucket with the new peer.
// If no such peer exists in that bucket, we do NOT add the peer to the Routing Table and return error "ErrPeerRejectedNoCapacity".
Expand Down Expand Up @@ -117,6 +143,11 @@ func (rt *RoutingTable) addPeer(p peer.ID, queryPeer bool) (bool, error) {

// peer already exists in the Routing Table.
if peer := bucket.getPeer(p); peer != nil {
// if we're querying the peer first time after adding it, let's give it a
// usefulness bump. This will ONLY happen once.
if peer.LastUsefulAt.IsZero() && queryPeer {
peer.LastUsefulAt = lastUsefulAt
}
Comment on lines +146 to +150
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is this fix related to the rest of the PR, or just a "while I'm already here" kind of thing?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey,

This change is not related to the rest of the PR.
I think I wasn't being mindful when I put this change in as I didn't realize I hadn't changed the branch. Apologies.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

No problem, while you're here and changing this do you mind add more information to the TryAddPeer comment about when we set LastUsefulAt? There is currently only information about when we set LastSuccessfulOutboundQuery

return false, nil
}

Expand Down Expand Up @@ -228,9 +259,25 @@ func (rt *RoutingTable) removePeer(p peer.ID) {
bucketID := rt.bucketIdForPeer(p)
bucket := rt.buckets[bucketID]
if bucket.remove(p) {
for {
lastBucketIndex := len(rt.buckets) - 1

// remove the last bucket if it's empty and it isn't the only bucket we have
if len(rt.buckets) > 1 && rt.buckets[lastBucketIndex].len() == 0 {
rt.buckets[lastBucketIndex] = nil
rt.buckets = rt.buckets[:lastBucketIndex]
} else if len(rt.buckets) >= 2 && rt.buckets[lastBucketIndex-1].len() == 0 {
// if the second last bucket just became empty, remove and replace it with the last bucket.
rt.buckets[lastBucketIndex-1] = rt.buckets[lastBucketIndex]
rt.buckets[lastBucketIndex] = nil
rt.buckets = rt.buckets[:lastBucketIndex]
} else {
break
}
}

// peer removed callback
rt.PeerRemoved(p)
return
}
}

Expand Down
131 changes: 131 additions & 0 deletions table_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -80,6 +80,137 @@ func TestBucket(t *testing.T) {
}
}

func TestNPeersForCpl(t *testing.T) {
t.Parallel()
local := test.RandPeerIDFatal(t)
m := pstore.NewMetrics()
rt, err := NewRoutingTable(2, ConvertPeerID(local), time.Hour, m, NoOpThreshold)
require.NoError(t, err)

require.Equal(t, 0, rt.NPeersForCpl(0))
require.Equal(t, 0, rt.NPeersForCpl(1))

// one peer with cpl 1
p, _ := rt.GenRandPeerID(1)
rt.TryAddPeer(p, true)
require.Equal(t, 0, rt.NPeersForCpl(0))
require.Equal(t, 1, rt.NPeersForCpl(1))
require.Equal(t, 0, rt.NPeersForCpl(2))

// one peer with cpl 0
p, _ = rt.GenRandPeerID(0)
rt.TryAddPeer(p, true)
require.Equal(t, 1, rt.NPeersForCpl(0))
require.Equal(t, 1, rt.NPeersForCpl(1))
require.Equal(t, 0, rt.NPeersForCpl(2))

// split the bucket with a peer with cpl 1
p, _ = rt.GenRandPeerID(1)
rt.TryAddPeer(p, true)
require.Equal(t, 1, rt.NPeersForCpl(0))
require.Equal(t, 2, rt.NPeersForCpl(1))
require.Equal(t, 0, rt.NPeersForCpl(2))

p, _ = rt.GenRandPeerID(0)
rt.TryAddPeer(p, true)
require.Equal(t, 2, rt.NPeersForCpl(0))
}

func TestEmptyBucketCollapse(t *testing.T) {
t.Parallel()
local := test.RandPeerIDFatal(t)

m := pstore.NewMetrics()
rt, err := NewRoutingTable(1, ConvertPeerID(local), time.Hour, m, NoOpThreshold)
require.NoError(t, err)

// generate peers with cpl 0,1,2 & 3
p1, _ := rt.GenRandPeerID(0)
p2, _ := rt.GenRandPeerID(1)
p3, _ := rt.GenRandPeerID(2)
p4, _ := rt.GenRandPeerID(3)

// remove peer on an empty bucket should not panic.
rt.RemovePeer(p1)

// add peer with cpl 0 and remove it..bucket should still exist as it's the ONLY bucket we have
b, err := rt.TryAddPeer(p1, true)
require.True(t, b)
require.NoError(t, err)
rt.RemovePeer(p1)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 1)
rt.tabLock.Unlock()
require.Empty(t, rt.ListPeers())

// add peer with cpl 0 and cpl 1 and verify we have two buckets.
b, err = rt.TryAddPeer(p1, true)
require.True(t, b)
b, err = rt.TryAddPeer(p2, true)
require.True(t, b)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 2)
rt.tabLock.Unlock()

// removing a peer from the last bucket collapses it.
rt.RemovePeer(p2)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 1)
rt.tabLock.Unlock()
require.Len(t, rt.ListPeers(), 1)
require.Contains(t, rt.ListPeers(), p1)

// add p2 again
b, err = rt.TryAddPeer(p2, true)
require.True(t, b)
require.NoError(t, err)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 2)
rt.tabLock.Unlock()

// now remove a peer from the second-last i.e. first bucket and ensure it collapses
rt.RemovePeer(p1)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 1)
rt.tabLock.Unlock()
require.Len(t, rt.ListPeers(), 1)
require.Contains(t, rt.ListPeers(), p2)

// let's have a total of 4 buckets now
rt.TryAddPeer(p1, true)
rt.TryAddPeer(p2, true)
rt.TryAddPeer(p3, true)
rt.TryAddPeer(p4, true)

rt.tabLock.Lock()
require.Len(t, rt.buckets, 4)
rt.tabLock.Unlock()

// removing from 2,3 and then 4 leaves us with ONLY one bucket
rt.RemovePeer(p2)
rt.RemovePeer(p3)
rt.RemovePeer(p4)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 1)
rt.tabLock.Unlock()

// an empty bucket in the middle DOES NOT collapse buckets
rt.TryAddPeer(p1, true)
rt.TryAddPeer(p2, true)
rt.TryAddPeer(p3, true)
rt.TryAddPeer(p4, true)

rt.tabLock.Lock()
require.Len(t, rt.buckets, 4)
rt.tabLock.Unlock()

rt.RemovePeer(p2)
rt.tabLock.Lock()
require.Len(t, rt.buckets, 4)
rt.tabLock.Unlock()
require.NotContains(t, rt.ListPeers(), p2)
}
aschmahmann marked this conversation as resolved.
Show resolved Hide resolved

func TestRemovePeer(t *testing.T) {
t.Parallel()
local := test.RandPeerIDFatal(t)
Expand Down